var/home/core/zuul-output/0000755000175000017500000000000015070555003014525 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015070570150015471 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000004750673115070570141017712 0ustar rootrootOct 05 20:54:34 crc systemd[1]: Starting Kubernetes Kubelet... Oct 05 20:54:35 crc restorecon[4569]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 05 20:54:35 crc restorecon[4569]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 05 20:54:35 crc restorecon[4569]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Oct 05 20:54:36 crc kubenswrapper[4754]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 05 20:54:36 crc kubenswrapper[4754]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Oct 05 20:54:36 crc kubenswrapper[4754]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 05 20:54:36 crc kubenswrapper[4754]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 05 20:54:36 crc kubenswrapper[4754]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Oct 05 20:54:36 crc kubenswrapper[4754]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.550074 4754 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.555770 4754 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.555793 4754 feature_gate.go:330] unrecognized feature gate: Example Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.555803 4754 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.555811 4754 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.555817 4754 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.555824 4754 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.555830 4754 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.555837 4754 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.555844 4754 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.555852 4754 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.555859 4754 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.555865 4754 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.555871 4754 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.555876 4754 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.555881 4754 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.555886 4754 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.555891 4754 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.555897 4754 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.555902 4754 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.555907 4754 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.555912 4754 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.555918 4754 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.555925 4754 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.555931 4754 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.555936 4754 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.555941 4754 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.555967 4754 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.555973 4754 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.555979 4754 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.555983 4754 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.555988 4754 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.555994 4754 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.555999 4754 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.556004 4754 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.556010 4754 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.556018 4754 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.556024 4754 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.556029 4754 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.556034 4754 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.556039 4754 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.556044 4754 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.556052 4754 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.556057 4754 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.556062 4754 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.556067 4754 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.556071 4754 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.556076 4754 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.556081 4754 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.556086 4754 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.556091 4754 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.556095 4754 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.556101 4754 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.556105 4754 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.556110 4754 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.556115 4754 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.556120 4754 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.556124 4754 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.556130 4754 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.556135 4754 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.556140 4754 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.556145 4754 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.556149 4754 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.556154 4754 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.556159 4754 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.556164 4754 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.556170 4754 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.556175 4754 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.556180 4754 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.556185 4754 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.556189 4754 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.556194 4754 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.556293 4754 flags.go:64] FLAG: --address="0.0.0.0" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.556304 4754 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.556313 4754 flags.go:64] FLAG: --anonymous-auth="true" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.556320 4754 flags.go:64] FLAG: --application-metrics-count-limit="100" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.556327 4754 flags.go:64] FLAG: --authentication-token-webhook="false" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.556333 4754 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.556341 4754 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.556348 4754 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.556354 4754 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.556359 4754 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.556366 4754 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.556372 4754 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.556377 4754 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.556383 4754 flags.go:64] FLAG: --cgroup-root="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.556388 4754 flags.go:64] FLAG: --cgroups-per-qos="true" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.556394 4754 flags.go:64] FLAG: --client-ca-file="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.556400 4754 flags.go:64] FLAG: --cloud-config="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.556405 4754 flags.go:64] FLAG: --cloud-provider="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.556410 4754 flags.go:64] FLAG: --cluster-dns="[]" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.556416 4754 flags.go:64] FLAG: --cluster-domain="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.556422 4754 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.556428 4754 flags.go:64] FLAG: --config-dir="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.556435 4754 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.556441 4754 flags.go:64] FLAG: --container-log-max-files="5" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.556449 4754 flags.go:64] FLAG: --container-log-max-size="10Mi" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.556455 4754 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.556460 4754 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.556470 4754 flags.go:64] FLAG: --containerd-namespace="k8s.io" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.556477 4754 flags.go:64] FLAG: --contention-profiling="false" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.556483 4754 flags.go:64] FLAG: --cpu-cfs-quota="true" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.556507 4754 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.556513 4754 flags.go:64] FLAG: --cpu-manager-policy="none" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.556519 4754 flags.go:64] FLAG: --cpu-manager-policy-options="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.556527 4754 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.556534 4754 flags.go:64] FLAG: --enable-controller-attach-detach="true" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.556540 4754 flags.go:64] FLAG: --enable-debugging-handlers="true" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.556547 4754 flags.go:64] FLAG: --enable-load-reader="false" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.556553 4754 flags.go:64] FLAG: --enable-server="true" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.556560 4754 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.556568 4754 flags.go:64] FLAG: --event-burst="100" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.556574 4754 flags.go:64] FLAG: --event-qps="50" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.556581 4754 flags.go:64] FLAG: --event-storage-age-limit="default=0" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.556586 4754 flags.go:64] FLAG: --event-storage-event-limit="default=0" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.556592 4754 flags.go:64] FLAG: --eviction-hard="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.556598 4754 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.556604 4754 flags.go:64] FLAG: --eviction-minimum-reclaim="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.556610 4754 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.556616 4754 flags.go:64] FLAG: --eviction-soft="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.556622 4754 flags.go:64] FLAG: --eviction-soft-grace-period="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.556627 4754 flags.go:64] FLAG: --exit-on-lock-contention="false" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.556633 4754 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.556638 4754 flags.go:64] FLAG: --experimental-mounter-path="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.556644 4754 flags.go:64] FLAG: --fail-cgroupv1="false" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.556649 4754 flags.go:64] FLAG: --fail-swap-on="true" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.556655 4754 flags.go:64] FLAG: --feature-gates="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.556662 4754 flags.go:64] FLAG: --file-check-frequency="20s" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.556667 4754 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.556673 4754 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.556680 4754 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.556685 4754 flags.go:64] FLAG: --healthz-port="10248" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.556691 4754 flags.go:64] FLAG: --help="false" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.556697 4754 flags.go:64] FLAG: --hostname-override="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.556703 4754 flags.go:64] FLAG: --housekeeping-interval="10s" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.556708 4754 flags.go:64] FLAG: --http-check-frequency="20s" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.556714 4754 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.556719 4754 flags.go:64] FLAG: --image-credential-provider-config="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.556726 4754 flags.go:64] FLAG: --image-gc-high-threshold="85" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.556732 4754 flags.go:64] FLAG: --image-gc-low-threshold="80" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.556738 4754 flags.go:64] FLAG: --image-service-endpoint="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.556744 4754 flags.go:64] FLAG: --kernel-memcg-notification="false" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.556750 4754 flags.go:64] FLAG: --kube-api-burst="100" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.556756 4754 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.556762 4754 flags.go:64] FLAG: --kube-api-qps="50" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.556768 4754 flags.go:64] FLAG: --kube-reserved="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.556806 4754 flags.go:64] FLAG: --kube-reserved-cgroup="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.556812 4754 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.556818 4754 flags.go:64] FLAG: --kubelet-cgroups="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.556824 4754 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.556830 4754 flags.go:64] FLAG: --lock-file="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.556835 4754 flags.go:64] FLAG: --log-cadvisor-usage="false" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.556841 4754 flags.go:64] FLAG: --log-flush-frequency="5s" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.556846 4754 flags.go:64] FLAG: --log-json-info-buffer-size="0" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.556855 4754 flags.go:64] FLAG: --log-json-split-stream="false" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.556861 4754 flags.go:64] FLAG: --log-text-info-buffer-size="0" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.556866 4754 flags.go:64] FLAG: --log-text-split-stream="false" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.556872 4754 flags.go:64] FLAG: --logging-format="text" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.556878 4754 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.556884 4754 flags.go:64] FLAG: --make-iptables-util-chains="true" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.556889 4754 flags.go:64] FLAG: --manifest-url="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.556895 4754 flags.go:64] FLAG: --manifest-url-header="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.556902 4754 flags.go:64] FLAG: --max-housekeeping-interval="15s" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.556908 4754 flags.go:64] FLAG: --max-open-files="1000000" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.556914 4754 flags.go:64] FLAG: --max-pods="110" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.556921 4754 flags.go:64] FLAG: --maximum-dead-containers="-1" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.556928 4754 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.556934 4754 flags.go:64] FLAG: --memory-manager-policy="None" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.556940 4754 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.556946 4754 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.556952 4754 flags.go:64] FLAG: --node-ip="192.168.126.11" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.556958 4754 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.556971 4754 flags.go:64] FLAG: --node-status-max-images="50" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.556980 4754 flags.go:64] FLAG: --node-status-update-frequency="10s" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.556986 4754 flags.go:64] FLAG: --oom-score-adj="-999" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.556992 4754 flags.go:64] FLAG: --pod-cidr="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.556997 4754 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.557006 4754 flags.go:64] FLAG: --pod-manifest-path="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.557011 4754 flags.go:64] FLAG: --pod-max-pids="-1" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.557017 4754 flags.go:64] FLAG: --pods-per-core="0" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.557022 4754 flags.go:64] FLAG: --port="10250" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.557028 4754 flags.go:64] FLAG: --protect-kernel-defaults="false" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.557034 4754 flags.go:64] FLAG: --provider-id="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.557039 4754 flags.go:64] FLAG: --qos-reserved="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.557045 4754 flags.go:64] FLAG: --read-only-port="10255" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.557051 4754 flags.go:64] FLAG: --register-node="true" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.557057 4754 flags.go:64] FLAG: --register-schedulable="true" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.557062 4754 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.557071 4754 flags.go:64] FLAG: --registry-burst="10" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.557077 4754 flags.go:64] FLAG: --registry-qps="5" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.557082 4754 flags.go:64] FLAG: --reserved-cpus="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.557088 4754 flags.go:64] FLAG: --reserved-memory="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.557095 4754 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.557100 4754 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.557106 4754 flags.go:64] FLAG: --rotate-certificates="false" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.557111 4754 flags.go:64] FLAG: --rotate-server-certificates="false" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.557117 4754 flags.go:64] FLAG: --runonce="false" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.557122 4754 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.557128 4754 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.557133 4754 flags.go:64] FLAG: --seccomp-default="false" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.557139 4754 flags.go:64] FLAG: --serialize-image-pulls="true" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.557145 4754 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.557151 4754 flags.go:64] FLAG: --storage-driver-db="cadvisor" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.557158 4754 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.557164 4754 flags.go:64] FLAG: --storage-driver-password="root" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.557171 4754 flags.go:64] FLAG: --storage-driver-secure="false" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.557177 4754 flags.go:64] FLAG: --storage-driver-table="stats" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.557183 4754 flags.go:64] FLAG: --storage-driver-user="root" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.557189 4754 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.557195 4754 flags.go:64] FLAG: --sync-frequency="1m0s" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.557201 4754 flags.go:64] FLAG: --system-cgroups="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.557207 4754 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.557216 4754 flags.go:64] FLAG: --system-reserved-cgroup="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.557222 4754 flags.go:64] FLAG: --tls-cert-file="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.557227 4754 flags.go:64] FLAG: --tls-cipher-suites="[]" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.557233 4754 flags.go:64] FLAG: --tls-min-version="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.557239 4754 flags.go:64] FLAG: --tls-private-key-file="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.557245 4754 flags.go:64] FLAG: --topology-manager-policy="none" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.557250 4754 flags.go:64] FLAG: --topology-manager-policy-options="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.557256 4754 flags.go:64] FLAG: --topology-manager-scope="container" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.557261 4754 flags.go:64] FLAG: --v="2" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.557269 4754 flags.go:64] FLAG: --version="false" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.557276 4754 flags.go:64] FLAG: --vmodule="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.557283 4754 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.557289 4754 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.557448 4754 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.557456 4754 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.557464 4754 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.557469 4754 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.557475 4754 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.557481 4754 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.557486 4754 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.557509 4754 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.557515 4754 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.557520 4754 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.557526 4754 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.557533 4754 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.557541 4754 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.557548 4754 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.557554 4754 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.557560 4754 feature_gate.go:330] unrecognized feature gate: Example Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.557565 4754 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.557571 4754 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.557576 4754 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.557581 4754 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.557586 4754 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.557592 4754 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.557596 4754 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.557602 4754 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.557606 4754 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.557611 4754 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.557618 4754 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.557624 4754 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.557630 4754 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.557635 4754 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.557640 4754 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.557645 4754 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.557650 4754 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.557655 4754 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.557660 4754 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.557665 4754 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.557670 4754 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.557675 4754 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.557680 4754 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.557685 4754 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.557690 4754 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.557695 4754 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.557700 4754 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.557705 4754 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.557713 4754 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.557719 4754 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.557724 4754 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.557730 4754 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.557735 4754 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.557741 4754 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.557745 4754 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.557750 4754 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.557755 4754 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.557760 4754 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.557765 4754 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.557770 4754 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.557775 4754 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.557780 4754 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.557786 4754 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.557792 4754 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.557798 4754 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.557804 4754 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.557810 4754 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.557816 4754 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.557822 4754 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.557829 4754 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.557835 4754 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.557841 4754 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.557847 4754 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.557854 4754 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.557861 4754 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.557871 4754 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.571918 4754 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.571990 4754 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.572142 4754 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.572165 4754 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.572176 4754 feature_gate.go:330] unrecognized feature gate: Example Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.572185 4754 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.572197 4754 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.572206 4754 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.572215 4754 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.572224 4754 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.572232 4754 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.572241 4754 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.572249 4754 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.572260 4754 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.572273 4754 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.572283 4754 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.572292 4754 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.572301 4754 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.572311 4754 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.572321 4754 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.572331 4754 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.572340 4754 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.572348 4754 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.572358 4754 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.572367 4754 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.572481 4754 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.572524 4754 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.572533 4754 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.572542 4754 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.572551 4754 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.572559 4754 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.572568 4754 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.572577 4754 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.572585 4754 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.572593 4754 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.572601 4754 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.572610 4754 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.572619 4754 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.572628 4754 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.572636 4754 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.572645 4754 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.572653 4754 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.572668 4754 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.572682 4754 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.572695 4754 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.572707 4754 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.572718 4754 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.572730 4754 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.572741 4754 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.572752 4754 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.572762 4754 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.572770 4754 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.572779 4754 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.572787 4754 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.572796 4754 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.572805 4754 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.572813 4754 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.572822 4754 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.572831 4754 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.572839 4754 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.572848 4754 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.572856 4754 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.572865 4754 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.572873 4754 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.572885 4754 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.572895 4754 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.572904 4754 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.572914 4754 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.572922 4754 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.572934 4754 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.572944 4754 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.572953 4754 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.572963 4754 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.572978 4754 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.573216 4754 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.573230 4754 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.573242 4754 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.573253 4754 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.573264 4754 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.573275 4754 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.573284 4754 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.573293 4754 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.573302 4754 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.573310 4754 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.573318 4754 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.573327 4754 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.573336 4754 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.573344 4754 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.573355 4754 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.573364 4754 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.573374 4754 feature_gate.go:330] unrecognized feature gate: Example Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.573382 4754 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.573391 4754 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.573400 4754 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.573409 4754 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.573418 4754 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.573431 4754 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.573442 4754 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.573453 4754 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.573463 4754 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.573472 4754 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.573482 4754 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.573522 4754 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.573531 4754 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.573540 4754 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.573548 4754 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.573557 4754 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.573565 4754 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.573574 4754 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.573583 4754 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.573591 4754 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.573599 4754 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.573608 4754 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.573616 4754 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.573626 4754 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.573634 4754 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.573643 4754 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.573651 4754 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.573659 4754 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.573668 4754 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.573676 4754 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.573684 4754 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.573694 4754 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.573702 4754 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.573711 4754 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.573720 4754 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.573728 4754 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.573736 4754 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.573744 4754 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.573757 4754 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.573768 4754 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.573777 4754 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.573786 4754 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.573795 4754 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.573804 4754 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.573813 4754 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.573824 4754 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.573834 4754 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.573842 4754 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.573851 4754 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.573860 4754 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.573868 4754 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.573876 4754 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.573885 4754 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.573893 4754 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.573906 4754 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.574128 4754 server.go:940] "Client rotation is on, will bootstrap in background" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.579987 4754 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.580130 4754 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.582177 4754 server.go:997] "Starting client certificate rotation" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.582226 4754 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.582448 4754 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-12-26 23:41:45.294586004 +0000 UTC Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.582581 4754 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 1970h47m8.712010021s for next certificate rotation Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.614816 4754 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.616964 4754 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.642836 4754 log.go:25] "Validated CRI v1 runtime API" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.686163 4754 log.go:25] "Validated CRI v1 image API" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.689148 4754 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.695688 4754 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-10-05-20-48-53-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.695740 4754 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.719637 4754 manager.go:217] Machine: {Timestamp:2025-10-05 20:54:36.716903009 +0000 UTC m=+0.621021759 CPUVendorID:AuthenticAMD NumCores:8 NumPhysicalCores:1 NumSockets:8 CpuFrequency:2800000 MemoryCapacity:25199480832 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:72d081dd-fe7e-42a6-9cf0-f0ee7f733ac9 BootID:ede2b442-76b0-443a-a450-252ee8711c80 Filesystems:[{Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:3076108 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:12599738368 Type:vfs Inodes:3076108 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:5039898624 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:12599742464 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:2519945216 Type:vfs Inodes:615221 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:429496729600 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:48:04:4d Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:48:04:4d Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:ad:f1:6a Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:ec:29:56 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:63:b1:f6 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:c1:b7:1c Speed:-1 Mtu:1496} {Name:eth10 MacAddress:96:dc:67:67:0e:73 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:f6:ae:7b:eb:cd:cd Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:25199480832 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.720003 4754 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.720206 4754 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.720939 4754 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.721335 4754 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.721404 4754 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.721829 4754 topology_manager.go:138] "Creating topology manager with none policy" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.721856 4754 container_manager_linux.go:303] "Creating device plugin manager" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.722686 4754 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.722758 4754 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.723852 4754 state_mem.go:36] "Initialized new in-memory state store" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.724101 4754 server.go:1245] "Using root directory" path="/var/lib/kubelet" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.729190 4754 kubelet.go:418] "Attempting to sync node with API server" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.729249 4754 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.729308 4754 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.729337 4754 kubelet.go:324] "Adding apiserver pod source" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.729366 4754 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.734778 4754 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.735961 4754 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.736928 4754 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.20:6443: connect: connection refused Oct 05 20:54:36 crc kubenswrapper[4754]: E1005 20:54:36.737172 4754 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.20:6443: connect: connection refused" logger="UnhandledError" Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.737229 4754 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.20:6443: connect: connection refused Oct 05 20:54:36 crc kubenswrapper[4754]: E1005 20:54:36.737331 4754 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.20:6443: connect: connection refused" logger="UnhandledError" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.737834 4754 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.739614 4754 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.739660 4754 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.739674 4754 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.739690 4754 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.739716 4754 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.739729 4754 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.739744 4754 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.739765 4754 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.739782 4754 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.739795 4754 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.739816 4754 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.739829 4754 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.741146 4754 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.741912 4754 server.go:1280] "Started kubelet" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.742086 4754 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.745293 4754 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.746679 4754 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.747488 4754 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.20:6443: connect: connection refused Oct 05 20:54:36 crc systemd[1]: Started Kubernetes Kubelet. Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.752934 4754 server.go:460] "Adding debug handlers to kubelet server" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.753159 4754 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.753232 4754 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.753563 4754 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-10 23:41:15.352649078 +0000 UTC Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.753631 4754 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 866h46m38.599023233s for next certificate rotation Oct 05 20:54:36 crc kubenswrapper[4754]: E1005 20:54:36.753598 4754 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.754946 4754 volume_manager.go:287] "The desired_state_of_world populator starts" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.754988 4754 volume_manager.go:289] "Starting Kubelet Volume Manager" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.755322 4754 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Oct 05 20:54:36 crc kubenswrapper[4754]: E1005 20:54:36.755583 4754 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.20:6443: connect: connection refused" interval="200ms" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.756118 4754 factory.go:55] Registering systemd factory Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.756275 4754 factory.go:221] Registration of the systemd container factory successfully Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.756951 4754 factory.go:153] Registering CRI-O factory Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.757117 4754 factory.go:221] Registration of the crio container factory successfully Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.757317 4754 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.757461 4754 factory.go:103] Registering Raw factory Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.758560 4754 manager.go:1196] Started watching for new ooms in manager Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.763098 4754 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.20:6443: connect: connection refused Oct 05 20:54:36 crc kubenswrapper[4754]: E1005 20:54:36.763340 4754 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.20:6443: connect: connection refused" logger="UnhandledError" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.768741 4754 manager.go:319] Starting recovery of all containers Oct 05 20:54:36 crc kubenswrapper[4754]: E1005 20:54:36.770294 4754 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.20:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.186bb3b6b647647e default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-10-05 20:54:36.741862526 +0000 UTC m=+0.645981276,LastTimestamp:2025-10-05 20:54:36.741862526 +0000 UTC m=+0.645981276,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.786591 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.786659 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.786673 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.786686 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.786700 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.786721 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.786734 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.786748 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.786764 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.786775 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.786788 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.786802 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.786821 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.786841 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.786856 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.786868 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.786880 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.786891 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.786903 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.786915 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.786927 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.786943 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.786955 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.786968 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.786981 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.786995 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.787010 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.787060 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.787075 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.787087 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.787103 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.791149 4754 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.791187 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.791203 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.791217 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.791230 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.791244 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.791257 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.791268 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.791281 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.791295 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.791306 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.791318 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.791331 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.791344 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.791356 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.791368 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.791382 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.791394 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.791405 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.791418 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.791430 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.791441 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.791459 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.791472 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.791487 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.791523 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.791538 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.791551 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.791563 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.791579 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.791610 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.791628 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.791643 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.791656 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.791670 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.791684 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.791696 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.791709 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.791720 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.791761 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.791772 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.791787 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.791800 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.791816 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.791833 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.791885 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.791901 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.791915 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.791928 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.791939 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.791953 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.791969 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.791985 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.792002 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.792017 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.792030 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.792045 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.792056 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.792068 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.792080 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.792093 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.792104 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.792116 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.792128 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.792139 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.792150 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.792161 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.792172 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.792182 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.792193 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.792204 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.792217 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.792228 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.792239 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.792256 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.792268 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.792283 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.792320 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.792333 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.792346 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.792358 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.792371 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.792383 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.792397 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.792408 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.792419 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.792429 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.792440 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.792454 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.792466 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.792479 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.792489 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.792539 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.792551 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.792561 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.792573 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.792584 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.792594 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.792605 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.792618 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.792630 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.792640 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.792651 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.792661 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.792674 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.792688 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.792702 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.792716 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.792732 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.792747 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.792763 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.792778 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.792798 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.792811 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.792840 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.792856 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.792868 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.792881 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.792893 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.792904 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.792916 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.792928 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.792940 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.792953 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.792965 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.792978 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.792990 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.793002 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.793013 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.793026 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.793038 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.793049 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.793060 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.793073 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.793085 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.793096 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.793109 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.793124 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.793137 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.793150 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.793162 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.793185 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.793198 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.793209 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.793220 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.793236 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.793248 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.793260 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.793273 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.793284 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.793296 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.793307 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.793319 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.793330 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.793340 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.793351 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.793363 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.793375 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.793387 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.793398 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.793409 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.793420 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.793431 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.793701 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.793716 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.793727 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.793739 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.793750 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.793761 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.793773 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.793785 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.793799 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.793812 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.793828 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.793845 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.793856 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.793867 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.793878 4754 reconstruct.go:97] "Volume reconstruction finished" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.793887 4754 reconciler.go:26] "Reconciler: start to sync state" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.800640 4754 manager.go:324] Recovery completed Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.818304 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.820699 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.820853 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.820951 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.822036 4754 cpu_manager.go:225] "Starting CPU manager" policy="none" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.822145 4754 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.822373 4754 state_mem.go:36] "Initialized new in-memory state store" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.832411 4754 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.835900 4754 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.835983 4754 status_manager.go:217] "Starting to sync pod status with apiserver" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.836028 4754 kubelet.go:2335] "Starting kubelet main sync loop" Oct 05 20:54:36 crc kubenswrapper[4754]: E1005 20:54:36.836244 4754 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Oct 05 20:54:36 crc kubenswrapper[4754]: W1005 20:54:36.839649 4754 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.20:6443: connect: connection refused Oct 05 20:54:36 crc kubenswrapper[4754]: E1005 20:54:36.839751 4754 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.20:6443: connect: connection refused" logger="UnhandledError" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.841973 4754 policy_none.go:49] "None policy: Start" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.843013 4754 memory_manager.go:170] "Starting memorymanager" policy="None" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.843053 4754 state_mem.go:35] "Initializing new in-memory state store" Oct 05 20:54:36 crc kubenswrapper[4754]: E1005 20:54:36.854328 4754 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.890688 4754 manager.go:334] "Starting Device Plugin manager" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.891028 4754 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.891051 4754 server.go:79] "Starting device plugin registration server" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.891614 4754 eviction_manager.go:189] "Eviction manager: starting control loop" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.891644 4754 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.892122 4754 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.892209 4754 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.892218 4754 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Oct 05 20:54:36 crc kubenswrapper[4754]: E1005 20:54:36.906853 4754 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.937957 4754 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc"] Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.938209 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.940096 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.940149 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.940173 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.940427 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.940586 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.940638 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.941968 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.942007 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.942024 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.942214 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.942969 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.943027 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.943959 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.943992 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.944008 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.944430 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.944457 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.944474 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.944796 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.944817 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.944831 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.944976 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.945638 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.945692 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.947397 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.947430 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.947446 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.949582 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.949616 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.949634 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.949783 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.950400 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.950446 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.952090 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.952137 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.952163 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.952205 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.952288 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.952308 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.952908 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.952997 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.955281 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.955327 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.955349 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:54:36 crc kubenswrapper[4754]: E1005 20:54:36.957914 4754 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.20:6443: connect: connection refused" interval="400ms" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.995959 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.995963 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.996259 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.996298 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.996332 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.996361 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.996390 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.996417 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.996447 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.996473 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.996537 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.996570 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.996601 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.996634 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.996661 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.996688 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.997909 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.997946 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.997966 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:54:36 crc kubenswrapper[4754]: I1005 20:54:36.998002 4754 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 05 20:54:36 crc kubenswrapper[4754]: E1005 20:54:36.998566 4754 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.20:6443: connect: connection refused" node="crc" Oct 05 20:54:37 crc kubenswrapper[4754]: I1005 20:54:37.099053 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 05 20:54:37 crc kubenswrapper[4754]: I1005 20:54:37.099676 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 05 20:54:37 crc kubenswrapper[4754]: I1005 20:54:37.099978 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 05 20:54:37 crc kubenswrapper[4754]: I1005 20:54:37.100231 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 05 20:54:37 crc kubenswrapper[4754]: I1005 20:54:37.100392 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 05 20:54:37 crc kubenswrapper[4754]: I1005 20:54:37.100619 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 05 20:54:37 crc kubenswrapper[4754]: I1005 20:54:37.100778 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 05 20:54:37 crc kubenswrapper[4754]: I1005 20:54:37.100932 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 05 20:54:37 crc kubenswrapper[4754]: I1005 20:54:37.101038 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 05 20:54:37 crc kubenswrapper[4754]: I1005 20:54:37.099393 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 05 20:54:37 crc kubenswrapper[4754]: I1005 20:54:37.100543 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 05 20:54:37 crc kubenswrapper[4754]: I1005 20:54:37.100722 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 05 20:54:37 crc kubenswrapper[4754]: I1005 20:54:37.099744 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 05 20:54:37 crc kubenswrapper[4754]: I1005 20:54:37.100018 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 05 20:54:37 crc kubenswrapper[4754]: I1005 20:54:37.100876 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 05 20:54:37 crc kubenswrapper[4754]: I1005 20:54:37.100318 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 05 20:54:37 crc kubenswrapper[4754]: I1005 20:54:37.101355 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 05 20:54:37 crc kubenswrapper[4754]: I1005 20:54:37.101076 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 05 20:54:37 crc kubenswrapper[4754]: I1005 20:54:37.101867 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 05 20:54:37 crc kubenswrapper[4754]: I1005 20:54:37.102043 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 05 20:54:37 crc kubenswrapper[4754]: I1005 20:54:37.102225 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 05 20:54:37 crc kubenswrapper[4754]: I1005 20:54:37.102385 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 05 20:54:37 crc kubenswrapper[4754]: I1005 20:54:37.103191 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 05 20:54:37 crc kubenswrapper[4754]: I1005 20:54:37.103314 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 05 20:54:37 crc kubenswrapper[4754]: I1005 20:54:37.103382 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 05 20:54:37 crc kubenswrapper[4754]: I1005 20:54:37.103449 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 05 20:54:37 crc kubenswrapper[4754]: I1005 20:54:37.102596 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 05 20:54:37 crc kubenswrapper[4754]: I1005 20:54:37.103580 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 05 20:54:37 crc kubenswrapper[4754]: I1005 20:54:37.103924 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 05 20:54:37 crc kubenswrapper[4754]: I1005 20:54:37.104165 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 05 20:54:37 crc kubenswrapper[4754]: I1005 20:54:37.198981 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 20:54:37 crc kubenswrapper[4754]: I1005 20:54:37.202012 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:54:37 crc kubenswrapper[4754]: I1005 20:54:37.202205 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:54:37 crc kubenswrapper[4754]: I1005 20:54:37.202456 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:54:37 crc kubenswrapper[4754]: I1005 20:54:37.202650 4754 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 05 20:54:37 crc kubenswrapper[4754]: E1005 20:54:37.203449 4754 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.20:6443: connect: connection refused" node="crc" Oct 05 20:54:37 crc kubenswrapper[4754]: I1005 20:54:37.283112 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 05 20:54:37 crc kubenswrapper[4754]: I1005 20:54:37.293453 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Oct 05 20:54:37 crc kubenswrapper[4754]: I1005 20:54:37.317457 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 05 20:54:37 crc kubenswrapper[4754]: I1005 20:54:37.339378 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 05 20:54:37 crc kubenswrapper[4754]: W1005 20:54:37.340564 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-f535a2cbd6bd0efafa8e964d8713689ebffcbcd22b1239b32e376f30fe2a99b7 WatchSource:0}: Error finding container f535a2cbd6bd0efafa8e964d8713689ebffcbcd22b1239b32e376f30fe2a99b7: Status 404 returned error can't find the container with id f535a2cbd6bd0efafa8e964d8713689ebffcbcd22b1239b32e376f30fe2a99b7 Oct 05 20:54:37 crc kubenswrapper[4754]: W1005 20:54:37.343522 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-3eaab3f008ca2ce81c4394760aa36f5870abd255ac59261490364d5a59c408d6 WatchSource:0}: Error finding container 3eaab3f008ca2ce81c4394760aa36f5870abd255ac59261490364d5a59c408d6: Status 404 returned error can't find the container with id 3eaab3f008ca2ce81c4394760aa36f5870abd255ac59261490364d5a59c408d6 Oct 05 20:54:37 crc kubenswrapper[4754]: I1005 20:54:37.348153 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 05 20:54:37 crc kubenswrapper[4754]: W1005 20:54:37.357691 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-a9eb8bcda060a68b1673abde2a48258152dc562a818ebfa9f732ce87cc2c1062 WatchSource:0}: Error finding container a9eb8bcda060a68b1673abde2a48258152dc562a818ebfa9f732ce87cc2c1062: Status 404 returned error can't find the container with id a9eb8bcda060a68b1673abde2a48258152dc562a818ebfa9f732ce87cc2c1062 Oct 05 20:54:37 crc kubenswrapper[4754]: E1005 20:54:37.359527 4754 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.20:6443: connect: connection refused" interval="800ms" Oct 05 20:54:37 crc kubenswrapper[4754]: W1005 20:54:37.364445 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-55731fdd63ab7c4c0341a2f679c3ee22ffdd360f6878b9e443de7193a867369e WatchSource:0}: Error finding container 55731fdd63ab7c4c0341a2f679c3ee22ffdd360f6878b9e443de7193a867369e: Status 404 returned error can't find the container with id 55731fdd63ab7c4c0341a2f679c3ee22ffdd360f6878b9e443de7193a867369e Oct 05 20:54:37 crc kubenswrapper[4754]: W1005 20:54:37.374313 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-78f82598f96d9e7e9c8e6786bad70e8f809cff9963953faba138cf591ea8a508 WatchSource:0}: Error finding container 78f82598f96d9e7e9c8e6786bad70e8f809cff9963953faba138cf591ea8a508: Status 404 returned error can't find the container with id 78f82598f96d9e7e9c8e6786bad70e8f809cff9963953faba138cf591ea8a508 Oct 05 20:54:37 crc kubenswrapper[4754]: W1005 20:54:37.556548 4754 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.20:6443: connect: connection refused Oct 05 20:54:37 crc kubenswrapper[4754]: E1005 20:54:37.556667 4754 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.20:6443: connect: connection refused" logger="UnhandledError" Oct 05 20:54:37 crc kubenswrapper[4754]: I1005 20:54:37.604220 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 20:54:37 crc kubenswrapper[4754]: I1005 20:54:37.607172 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:54:37 crc kubenswrapper[4754]: I1005 20:54:37.607235 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:54:37 crc kubenswrapper[4754]: I1005 20:54:37.607255 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:54:37 crc kubenswrapper[4754]: I1005 20:54:37.607297 4754 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 05 20:54:37 crc kubenswrapper[4754]: E1005 20:54:37.608090 4754 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.20:6443: connect: connection refused" node="crc" Oct 05 20:54:37 crc kubenswrapper[4754]: W1005 20:54:37.679033 4754 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.20:6443: connect: connection refused Oct 05 20:54:37 crc kubenswrapper[4754]: E1005 20:54:37.679171 4754 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.20:6443: connect: connection refused" logger="UnhandledError" Oct 05 20:54:37 crc kubenswrapper[4754]: I1005 20:54:37.749448 4754 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.20:6443: connect: connection refused Oct 05 20:54:37 crc kubenswrapper[4754]: I1005 20:54:37.841831 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"a9eb8bcda060a68b1673abde2a48258152dc562a818ebfa9f732ce87cc2c1062"} Oct 05 20:54:37 crc kubenswrapper[4754]: I1005 20:54:37.844195 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"f535a2cbd6bd0efafa8e964d8713689ebffcbcd22b1239b32e376f30fe2a99b7"} Oct 05 20:54:37 crc kubenswrapper[4754]: I1005 20:54:37.846340 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"3eaab3f008ca2ce81c4394760aa36f5870abd255ac59261490364d5a59c408d6"} Oct 05 20:54:37 crc kubenswrapper[4754]: I1005 20:54:37.847651 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"78f82598f96d9e7e9c8e6786bad70e8f809cff9963953faba138cf591ea8a508"} Oct 05 20:54:37 crc kubenswrapper[4754]: I1005 20:54:37.848927 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"55731fdd63ab7c4c0341a2f679c3ee22ffdd360f6878b9e443de7193a867369e"} Oct 05 20:54:37 crc kubenswrapper[4754]: W1005 20:54:37.935085 4754 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.20:6443: connect: connection refused Oct 05 20:54:37 crc kubenswrapper[4754]: E1005 20:54:37.935341 4754 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.20:6443: connect: connection refused" logger="UnhandledError" Oct 05 20:54:38 crc kubenswrapper[4754]: E1005 20:54:38.161399 4754 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.20:6443: connect: connection refused" interval="1.6s" Oct 05 20:54:38 crc kubenswrapper[4754]: W1005 20:54:38.236918 4754 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.20:6443: connect: connection refused Oct 05 20:54:38 crc kubenswrapper[4754]: E1005 20:54:38.237063 4754 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.20:6443: connect: connection refused" logger="UnhandledError" Oct 05 20:54:38 crc kubenswrapper[4754]: I1005 20:54:38.409031 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 20:54:38 crc kubenswrapper[4754]: I1005 20:54:38.410283 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:54:38 crc kubenswrapper[4754]: I1005 20:54:38.410322 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:54:38 crc kubenswrapper[4754]: I1005 20:54:38.410336 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:54:38 crc kubenswrapper[4754]: I1005 20:54:38.410370 4754 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 05 20:54:38 crc kubenswrapper[4754]: E1005 20:54:38.410928 4754 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.20:6443: connect: connection refused" node="crc" Oct 05 20:54:38 crc kubenswrapper[4754]: I1005 20:54:38.749247 4754 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.20:6443: connect: connection refused Oct 05 20:54:38 crc kubenswrapper[4754]: I1005 20:54:38.858970 4754 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="c187ac7d3514f4e68984ee98c274c53418df1818159cc9d6c16626d162fa3f97" exitCode=0 Oct 05 20:54:38 crc kubenswrapper[4754]: I1005 20:54:38.859172 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 20:54:38 crc kubenswrapper[4754]: I1005 20:54:38.859031 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"c187ac7d3514f4e68984ee98c274c53418df1818159cc9d6c16626d162fa3f97"} Oct 05 20:54:38 crc kubenswrapper[4754]: I1005 20:54:38.861170 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:54:38 crc kubenswrapper[4754]: I1005 20:54:38.861217 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:54:38 crc kubenswrapper[4754]: I1005 20:54:38.861237 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:54:38 crc kubenswrapper[4754]: I1005 20:54:38.863124 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"600bc3bb551b747cf2f18e26745883232d7a37ac5877fad0c67100e52d7a43aa"} Oct 05 20:54:38 crc kubenswrapper[4754]: I1005 20:54:38.863167 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"9042df2f43243d3d79ba9a3940f8dd4f6c7a89ef9e3510e81fe60b1944d7e83b"} Oct 05 20:54:38 crc kubenswrapper[4754]: I1005 20:54:38.863179 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"5366bbf46a91667355f43f5df7335a536b395cda788b90b35ef18e2b53bc4c02"} Oct 05 20:54:38 crc kubenswrapper[4754]: I1005 20:54:38.866182 4754 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="54c568b637f3737667433047b956c7eb22fdf0f9301acdfd3b49fb84c39de6ed" exitCode=0 Oct 05 20:54:38 crc kubenswrapper[4754]: I1005 20:54:38.866246 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"54c568b637f3737667433047b956c7eb22fdf0f9301acdfd3b49fb84c39de6ed"} Oct 05 20:54:38 crc kubenswrapper[4754]: I1005 20:54:38.866350 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 20:54:38 crc kubenswrapper[4754]: I1005 20:54:38.867299 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:54:38 crc kubenswrapper[4754]: I1005 20:54:38.867318 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:54:38 crc kubenswrapper[4754]: I1005 20:54:38.867327 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:54:38 crc kubenswrapper[4754]: I1005 20:54:38.868813 4754 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="9ea160a5b37aea7f2970ae4dd2614b09d3dcf5dcbd1fc103b9997f8ecc3a804c" exitCode=0 Oct 05 20:54:38 crc kubenswrapper[4754]: I1005 20:54:38.868883 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"9ea160a5b37aea7f2970ae4dd2614b09d3dcf5dcbd1fc103b9997f8ecc3a804c"} Oct 05 20:54:38 crc kubenswrapper[4754]: I1005 20:54:38.868959 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 20:54:38 crc kubenswrapper[4754]: I1005 20:54:38.868993 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 20:54:38 crc kubenswrapper[4754]: I1005 20:54:38.870286 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:54:38 crc kubenswrapper[4754]: I1005 20:54:38.870321 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:54:38 crc kubenswrapper[4754]: I1005 20:54:38.870334 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:54:38 crc kubenswrapper[4754]: I1005 20:54:38.871323 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:54:38 crc kubenswrapper[4754]: I1005 20:54:38.871355 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:54:38 crc kubenswrapper[4754]: I1005 20:54:38.871369 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:54:38 crc kubenswrapper[4754]: I1005 20:54:38.877078 4754 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="a6482afbcf8935223f52760fa7bf9ab57e151e8fe9999f0ca43fc6db9f9f7b6f" exitCode=0 Oct 05 20:54:38 crc kubenswrapper[4754]: I1005 20:54:38.877109 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"a6482afbcf8935223f52760fa7bf9ab57e151e8fe9999f0ca43fc6db9f9f7b6f"} Oct 05 20:54:38 crc kubenswrapper[4754]: I1005 20:54:38.877172 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 20:54:38 crc kubenswrapper[4754]: I1005 20:54:38.877987 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:54:38 crc kubenswrapper[4754]: I1005 20:54:38.878056 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:54:38 crc kubenswrapper[4754]: I1005 20:54:38.878078 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:54:39 crc kubenswrapper[4754]: I1005 20:54:39.748383 4754 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.20:6443: connect: connection refused Oct 05 20:54:39 crc kubenswrapper[4754]: E1005 20:54:39.762814 4754 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.20:6443: connect: connection refused" interval="3.2s" Oct 05 20:54:39 crc kubenswrapper[4754]: I1005 20:54:39.882964 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"a5d1820b7c2ba7c5262f52297111f58e8f05771f3d7e34a59f8f3a05a7fb0bfe"} Oct 05 20:54:39 crc kubenswrapper[4754]: I1005 20:54:39.883027 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"1a0c1914423693dd5ca96eddab941162a0f1e9efaf4bd02c47514f092e39e9e6"} Oct 05 20:54:39 crc kubenswrapper[4754]: I1005 20:54:39.883048 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"02dec08702aeaa03bc26151781909a4c6226fdaf37b921feb2a53ad6fa3c94c3"} Oct 05 20:54:39 crc kubenswrapper[4754]: I1005 20:54:39.883203 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 20:54:39 crc kubenswrapper[4754]: I1005 20:54:39.884132 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:54:39 crc kubenswrapper[4754]: I1005 20:54:39.884164 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:54:39 crc kubenswrapper[4754]: I1005 20:54:39.884179 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:54:39 crc kubenswrapper[4754]: I1005 20:54:39.890966 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"78f74982180048eb50b9be79fa72c04f62f925b54fe511920f37e6407f864631"} Oct 05 20:54:39 crc kubenswrapper[4754]: I1005 20:54:39.891105 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 20:54:39 crc kubenswrapper[4754]: I1005 20:54:39.897955 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:54:39 crc kubenswrapper[4754]: I1005 20:54:39.897992 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:54:39 crc kubenswrapper[4754]: I1005 20:54:39.898005 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:54:39 crc kubenswrapper[4754]: I1005 20:54:39.902712 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"4efb34f3493cc8854bc4cff6b29f63973c52b754404d7a21daec4e1f612d327d"} Oct 05 20:54:39 crc kubenswrapper[4754]: I1005 20:54:39.902780 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"df6d700a947d19c84b4f65fd2dd2f8d1f4f1f80fa9b4800c2acfa8dae3912639"} Oct 05 20:54:39 crc kubenswrapper[4754]: I1005 20:54:39.902801 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"35ca8ee77a0e58cf969bd67b277a3392614c4e73b963e6f6002097709370715d"} Oct 05 20:54:39 crc kubenswrapper[4754]: I1005 20:54:39.902815 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"33da5a6bb3cd3b042c193bc571e6622f2f87f27e560cdd0bbb9be31e2010d707"} Oct 05 20:54:39 crc kubenswrapper[4754]: I1005 20:54:39.904866 4754 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="17ae6aa6c53e7dd061f70b41428ef50bbc7ffcc65352fad1a4a9d3c69dbd2ba9" exitCode=0 Oct 05 20:54:39 crc kubenswrapper[4754]: I1005 20:54:39.904943 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"17ae6aa6c53e7dd061f70b41428ef50bbc7ffcc65352fad1a4a9d3c69dbd2ba9"} Oct 05 20:54:39 crc kubenswrapper[4754]: I1005 20:54:39.905131 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 20:54:39 crc kubenswrapper[4754]: I1005 20:54:39.906075 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:54:39 crc kubenswrapper[4754]: I1005 20:54:39.906109 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:54:39 crc kubenswrapper[4754]: I1005 20:54:39.906126 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:54:39 crc kubenswrapper[4754]: I1005 20:54:39.906882 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"be822fa37e7f0aabf2715000b5a12edf3fe5c3df2999739551d2a89242c18129"} Oct 05 20:54:39 crc kubenswrapper[4754]: I1005 20:54:39.906933 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 20:54:39 crc kubenswrapper[4754]: I1005 20:54:39.907790 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:54:39 crc kubenswrapper[4754]: I1005 20:54:39.907817 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:54:39 crc kubenswrapper[4754]: I1005 20:54:39.907830 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:54:39 crc kubenswrapper[4754]: I1005 20:54:39.939303 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 05 20:54:40 crc kubenswrapper[4754]: I1005 20:54:40.011572 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 20:54:40 crc kubenswrapper[4754]: I1005 20:54:40.026784 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:54:40 crc kubenswrapper[4754]: I1005 20:54:40.026844 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:54:40 crc kubenswrapper[4754]: I1005 20:54:40.026856 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:54:40 crc kubenswrapper[4754]: I1005 20:54:40.026887 4754 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 05 20:54:40 crc kubenswrapper[4754]: E1005 20:54:40.027519 4754 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.20:6443: connect: connection refused" node="crc" Oct 05 20:54:40 crc kubenswrapper[4754]: W1005 20:54:40.155250 4754 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.20:6443: connect: connection refused Oct 05 20:54:40 crc kubenswrapper[4754]: E1005 20:54:40.155364 4754 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.20:6443: connect: connection refused" logger="UnhandledError" Oct 05 20:54:40 crc kubenswrapper[4754]: W1005 20:54:40.396034 4754 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.20:6443: connect: connection refused Oct 05 20:54:40 crc kubenswrapper[4754]: E1005 20:54:40.396168 4754 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.20:6443: connect: connection refused" logger="UnhandledError" Oct 05 20:54:40 crc kubenswrapper[4754]: I1005 20:54:40.535324 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 05 20:54:40 crc kubenswrapper[4754]: I1005 20:54:40.748832 4754 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.20:6443: connect: connection refused Oct 05 20:54:40 crc kubenswrapper[4754]: W1005 20:54:40.752988 4754 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.20:6443: connect: connection refused Oct 05 20:54:40 crc kubenswrapper[4754]: E1005 20:54:40.753243 4754 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.20:6443: connect: connection refused" logger="UnhandledError" Oct 05 20:54:40 crc kubenswrapper[4754]: I1005 20:54:40.921791 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"9d9b0b384ffcc4040a3cc8aea4be89ded347240018f198471da1922681c8efec"} Oct 05 20:54:40 crc kubenswrapper[4754]: I1005 20:54:40.922065 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 20:54:40 crc kubenswrapper[4754]: I1005 20:54:40.923334 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:54:40 crc kubenswrapper[4754]: I1005 20:54:40.923380 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:54:40 crc kubenswrapper[4754]: I1005 20:54:40.923394 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:54:40 crc kubenswrapper[4754]: I1005 20:54:40.926178 4754 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="ad4cd69df7608c7d87d839fe09d4557c674c7efbc4730f4e4d3daaf9a26f5aee" exitCode=0 Oct 05 20:54:40 crc kubenswrapper[4754]: I1005 20:54:40.926273 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"ad4cd69df7608c7d87d839fe09d4557c674c7efbc4730f4e4d3daaf9a26f5aee"} Oct 05 20:54:40 crc kubenswrapper[4754]: I1005 20:54:40.926304 4754 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 05 20:54:40 crc kubenswrapper[4754]: I1005 20:54:40.926411 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 20:54:40 crc kubenswrapper[4754]: I1005 20:54:40.926414 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 20:54:40 crc kubenswrapper[4754]: I1005 20:54:40.926550 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 20:54:40 crc kubenswrapper[4754]: I1005 20:54:40.926589 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 20:54:40 crc kubenswrapper[4754]: I1005 20:54:40.927926 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:54:40 crc kubenswrapper[4754]: I1005 20:54:40.927970 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:54:40 crc kubenswrapper[4754]: I1005 20:54:40.927989 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:54:40 crc kubenswrapper[4754]: I1005 20:54:40.928748 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:54:40 crc kubenswrapper[4754]: I1005 20:54:40.928782 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:54:40 crc kubenswrapper[4754]: I1005 20:54:40.928797 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:54:40 crc kubenswrapper[4754]: I1005 20:54:40.929041 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:54:40 crc kubenswrapper[4754]: I1005 20:54:40.929067 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:54:40 crc kubenswrapper[4754]: I1005 20:54:40.929085 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:54:40 crc kubenswrapper[4754]: I1005 20:54:40.929722 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:54:40 crc kubenswrapper[4754]: I1005 20:54:40.929978 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:54:40 crc kubenswrapper[4754]: I1005 20:54:40.930162 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:54:41 crc kubenswrapper[4754]: W1005 20:54:41.345556 4754 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.20:6443: connect: connection refused Oct 05 20:54:41 crc kubenswrapper[4754]: E1005 20:54:41.345687 4754 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.20:6443: connect: connection refused" logger="UnhandledError" Oct 05 20:54:41 crc kubenswrapper[4754]: I1005 20:54:41.934476 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"d969d3d50d6ad280fc6654df9441bab238bb538f277ed6d1273779813ffbaaed"} Oct 05 20:54:41 crc kubenswrapper[4754]: I1005 20:54:41.934581 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"dde1519111fc3af5d14463fdb58d4f5c62a7045e7de98c60981c7e4da37b0e7c"} Oct 05 20:54:41 crc kubenswrapper[4754]: I1005 20:54:41.937011 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Oct 05 20:54:41 crc kubenswrapper[4754]: I1005 20:54:41.940613 4754 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="9d9b0b384ffcc4040a3cc8aea4be89ded347240018f198471da1922681c8efec" exitCode=255 Oct 05 20:54:41 crc kubenswrapper[4754]: I1005 20:54:41.940818 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 20:54:41 crc kubenswrapper[4754]: I1005 20:54:41.941714 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 20:54:41 crc kubenswrapper[4754]: I1005 20:54:41.942251 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"9d9b0b384ffcc4040a3cc8aea4be89ded347240018f198471da1922681c8efec"} Oct 05 20:54:41 crc kubenswrapper[4754]: I1005 20:54:41.942959 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:54:41 crc kubenswrapper[4754]: I1005 20:54:41.943019 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:54:41 crc kubenswrapper[4754]: I1005 20:54:41.943034 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:54:41 crc kubenswrapper[4754]: I1005 20:54:41.943956 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:54:41 crc kubenswrapper[4754]: I1005 20:54:41.944056 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:54:41 crc kubenswrapper[4754]: I1005 20:54:41.944080 4754 scope.go:117] "RemoveContainer" containerID="9d9b0b384ffcc4040a3cc8aea4be89ded347240018f198471da1922681c8efec" Oct 05 20:54:41 crc kubenswrapper[4754]: I1005 20:54:41.944090 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:54:42 crc kubenswrapper[4754]: I1005 20:54:42.940437 4754 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 05 20:54:42 crc kubenswrapper[4754]: I1005 20:54:42.940678 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 05 20:54:42 crc kubenswrapper[4754]: I1005 20:54:42.949183 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Oct 05 20:54:42 crc kubenswrapper[4754]: I1005 20:54:42.952799 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"8907b9b69b249c34ec6db1e460d641b31711b75d6d31d4caafd867339e6571d0"} Oct 05 20:54:42 crc kubenswrapper[4754]: I1005 20:54:42.953032 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 20:54:42 crc kubenswrapper[4754]: I1005 20:54:42.953279 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 05 20:54:42 crc kubenswrapper[4754]: I1005 20:54:42.954918 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:54:42 crc kubenswrapper[4754]: I1005 20:54:42.955128 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:54:42 crc kubenswrapper[4754]: I1005 20:54:42.955153 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:54:42 crc kubenswrapper[4754]: I1005 20:54:42.962162 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"93b729f3f7f993db0d6bc3ba60c04930aada3cef8272ffcc3c6a8a242a9b1475"} Oct 05 20:54:42 crc kubenswrapper[4754]: I1005 20:54:42.962243 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"8cd9b08e83dc174413cc37891c14562816787028471a4201680c86dd620b7a21"} Oct 05 20:54:42 crc kubenswrapper[4754]: I1005 20:54:42.962269 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"800a731ed8ef57714b2318af14054ef23f0368936630887a093fc77167b33d12"} Oct 05 20:54:42 crc kubenswrapper[4754]: I1005 20:54:42.962353 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 20:54:42 crc kubenswrapper[4754]: I1005 20:54:42.964007 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:54:42 crc kubenswrapper[4754]: I1005 20:54:42.964072 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:54:42 crc kubenswrapper[4754]: I1005 20:54:42.964093 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:54:43 crc kubenswrapper[4754]: I1005 20:54:43.227990 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 20:54:43 crc kubenswrapper[4754]: I1005 20:54:43.230244 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:54:43 crc kubenswrapper[4754]: I1005 20:54:43.230313 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:54:43 crc kubenswrapper[4754]: I1005 20:54:43.230343 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:54:43 crc kubenswrapper[4754]: I1005 20:54:43.230393 4754 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 05 20:54:43 crc kubenswrapper[4754]: I1005 20:54:43.514895 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 05 20:54:43 crc kubenswrapper[4754]: I1005 20:54:43.515191 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 20:54:43 crc kubenswrapper[4754]: I1005 20:54:43.517294 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:54:43 crc kubenswrapper[4754]: I1005 20:54:43.517367 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:54:43 crc kubenswrapper[4754]: I1005 20:54:43.517386 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:54:43 crc kubenswrapper[4754]: I1005 20:54:43.966094 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 20:54:43 crc kubenswrapper[4754]: I1005 20:54:43.966203 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 20:54:43 crc kubenswrapper[4754]: I1005 20:54:43.967222 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 05 20:54:43 crc kubenswrapper[4754]: I1005 20:54:43.967951 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:54:43 crc kubenswrapper[4754]: I1005 20:54:43.968006 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:54:43 crc kubenswrapper[4754]: I1005 20:54:43.968024 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:54:43 crc kubenswrapper[4754]: I1005 20:54:43.969139 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:54:43 crc kubenswrapper[4754]: I1005 20:54:43.969198 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:54:43 crc kubenswrapper[4754]: I1005 20:54:43.969217 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:54:44 crc kubenswrapper[4754]: I1005 20:54:44.643897 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 05 20:54:44 crc kubenswrapper[4754]: I1005 20:54:44.644106 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 20:54:44 crc kubenswrapper[4754]: I1005 20:54:44.645931 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:54:44 crc kubenswrapper[4754]: I1005 20:54:44.645984 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:54:44 crc kubenswrapper[4754]: I1005 20:54:44.646003 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:54:44 crc kubenswrapper[4754]: I1005 20:54:44.651523 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 05 20:54:44 crc kubenswrapper[4754]: I1005 20:54:44.969585 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 20:54:44 crc kubenswrapper[4754]: I1005 20:54:44.969598 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 20:54:44 crc kubenswrapper[4754]: I1005 20:54:44.971291 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:54:44 crc kubenswrapper[4754]: I1005 20:54:44.971356 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:54:44 crc kubenswrapper[4754]: I1005 20:54:44.971378 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:54:44 crc kubenswrapper[4754]: I1005 20:54:44.971719 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:54:44 crc kubenswrapper[4754]: I1005 20:54:44.971768 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:54:44 crc kubenswrapper[4754]: I1005 20:54:44.971789 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:54:45 crc kubenswrapper[4754]: I1005 20:54:45.240048 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 05 20:54:45 crc kubenswrapper[4754]: I1005 20:54:45.286638 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Oct 05 20:54:45 crc kubenswrapper[4754]: I1005 20:54:45.286908 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 20:54:45 crc kubenswrapper[4754]: I1005 20:54:45.288680 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:54:45 crc kubenswrapper[4754]: I1005 20:54:45.288729 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:54:45 crc kubenswrapper[4754]: I1005 20:54:45.288746 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:54:45 crc kubenswrapper[4754]: I1005 20:54:45.763256 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 05 20:54:45 crc kubenswrapper[4754]: I1005 20:54:45.862392 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 05 20:54:45 crc kubenswrapper[4754]: I1005 20:54:45.972487 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 20:54:45 crc kubenswrapper[4754]: I1005 20:54:45.972582 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 20:54:45 crc kubenswrapper[4754]: I1005 20:54:45.974688 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:54:45 crc kubenswrapper[4754]: I1005 20:54:45.974760 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:54:45 crc kubenswrapper[4754]: I1005 20:54:45.974780 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:54:45 crc kubenswrapper[4754]: I1005 20:54:45.975420 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:54:45 crc kubenswrapper[4754]: I1005 20:54:45.975486 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:54:45 crc kubenswrapper[4754]: I1005 20:54:45.975566 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:54:46 crc kubenswrapper[4754]: E1005 20:54:46.907209 4754 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Oct 05 20:54:46 crc kubenswrapper[4754]: I1005 20:54:46.975297 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 20:54:46 crc kubenswrapper[4754]: I1005 20:54:46.976892 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:54:46 crc kubenswrapper[4754]: I1005 20:54:46.976960 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:54:46 crc kubenswrapper[4754]: I1005 20:54:46.976983 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:54:50 crc kubenswrapper[4754]: I1005 20:54:50.543186 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 05 20:54:50 crc kubenswrapper[4754]: I1005 20:54:50.543448 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 20:54:50 crc kubenswrapper[4754]: I1005 20:54:50.545658 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:54:50 crc kubenswrapper[4754]: I1005 20:54:50.545726 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:54:50 crc kubenswrapper[4754]: I1005 20:54:50.545745 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:54:50 crc kubenswrapper[4754]: I1005 20:54:50.717038 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Oct 05 20:54:50 crc kubenswrapper[4754]: I1005 20:54:50.717399 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 20:54:50 crc kubenswrapper[4754]: I1005 20:54:50.719080 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:54:50 crc kubenswrapper[4754]: I1005 20:54:50.719142 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:54:50 crc kubenswrapper[4754]: I1005 20:54:50.719161 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:54:51 crc kubenswrapper[4754]: I1005 20:54:51.750662 4754 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Oct 05 20:54:52 crc kubenswrapper[4754]: I1005 20:54:52.821192 4754 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Oct 05 20:54:52 crc kubenswrapper[4754]: I1005 20:54:52.821276 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Oct 05 20:54:52 crc kubenswrapper[4754]: I1005 20:54:52.827011 4754 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\": RBAC: [clusterrole.rbac.authorization.k8s.io \"system:openshift:public-info-viewer\" not found, clusterrole.rbac.authorization.k8s.io \"system:public-info-viewer\" not found]","reason":"Forbidden","details":{},"code":403} Oct 05 20:54:52 crc kubenswrapper[4754]: I1005 20:54:52.827104 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Oct 05 20:54:52 crc kubenswrapper[4754]: I1005 20:54:52.940754 4754 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 05 20:54:52 crc kubenswrapper[4754]: I1005 20:54:52.940848 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 05 20:54:55 crc kubenswrapper[4754]: I1005 20:54:55.869283 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 05 20:54:55 crc kubenswrapper[4754]: I1005 20:54:55.869586 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 20:54:55 crc kubenswrapper[4754]: I1005 20:54:55.871896 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:54:55 crc kubenswrapper[4754]: I1005 20:54:55.872374 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:54:55 crc kubenswrapper[4754]: I1005 20:54:55.872399 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:54:55 crc kubenswrapper[4754]: I1005 20:54:55.876243 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 05 20:54:56 crc kubenswrapper[4754]: I1005 20:54:56.000827 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 20:54:56 crc kubenswrapper[4754]: I1005 20:54:56.001702 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:54:56 crc kubenswrapper[4754]: I1005 20:54:56.001756 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:54:56 crc kubenswrapper[4754]: I1005 20:54:56.001770 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:54:56 crc kubenswrapper[4754]: E1005 20:54:56.907339 4754 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Oct 05 20:54:57 crc kubenswrapper[4754]: E1005 20:54:57.788076 4754 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="6.4s" Oct 05 20:54:57 crc kubenswrapper[4754]: I1005 20:54:57.791123 4754 trace.go:236] Trace[658101130]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (05-Oct-2025 20:54:47.555) (total time: 10235ms): Oct 05 20:54:57 crc kubenswrapper[4754]: Trace[658101130]: ---"Objects listed" error: 10235ms (20:54:57.790) Oct 05 20:54:57 crc kubenswrapper[4754]: Trace[658101130]: [10.235831985s] [10.235831985s] END Oct 05 20:54:57 crc kubenswrapper[4754]: I1005 20:54:57.791167 4754 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Oct 05 20:54:57 crc kubenswrapper[4754]: E1005 20:54:57.794177 4754 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Oct 05 20:54:57 crc kubenswrapper[4754]: I1005 20:54:57.795147 4754 trace.go:236] Trace[909690425]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (05-Oct-2025 20:54:44.151) (total time: 13643ms): Oct 05 20:54:57 crc kubenswrapper[4754]: Trace[909690425]: ---"Objects listed" error: 13643ms (20:54:57.795) Oct 05 20:54:57 crc kubenswrapper[4754]: Trace[909690425]: [13.643313836s] [13.643313836s] END Oct 05 20:54:57 crc kubenswrapper[4754]: I1005 20:54:57.795198 4754 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Oct 05 20:54:57 crc kubenswrapper[4754]: I1005 20:54:57.797010 4754 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Oct 05 20:54:57 crc kubenswrapper[4754]: I1005 20:54:57.797080 4754 trace.go:236] Trace[135455052]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (05-Oct-2025 20:54:45.140) (total time: 12656ms): Oct 05 20:54:57 crc kubenswrapper[4754]: Trace[135455052]: ---"Objects listed" error: 12656ms (20:54:57.796) Oct 05 20:54:57 crc kubenswrapper[4754]: Trace[135455052]: [12.656743494s] [12.656743494s] END Oct 05 20:54:57 crc kubenswrapper[4754]: I1005 20:54:57.797104 4754 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Oct 05 20:54:57 crc kubenswrapper[4754]: I1005 20:54:57.797370 4754 trace.go:236] Trace[623202390]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (05-Oct-2025 20:54:45.635) (total time: 12162ms): Oct 05 20:54:57 crc kubenswrapper[4754]: Trace[623202390]: ---"Objects listed" error: 12162ms (20:54:57.797) Oct 05 20:54:57 crc kubenswrapper[4754]: Trace[623202390]: [12.162262351s] [12.162262351s] END Oct 05 20:54:57 crc kubenswrapper[4754]: I1005 20:54:57.797391 4754 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.155396 4754 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:38660->192.168.126.11:17697: read: connection reset by peer" start-of-body= Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.155464 4754 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:38660->192.168.126.11:17697: read: connection reset by peer" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.155876 4754 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.155943 4754 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.156251 4754 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.156289 4754 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.741767 4754 apiserver.go:52] "Watching apiserver" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.747784 4754 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.748450 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf"] Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.749842 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.751007 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 20:54:58 crc kubenswrapper[4754]: E1005 20:54:58.751140 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.751597 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 20:54:58 crc kubenswrapper[4754]: E1005 20:54:58.751651 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.751761 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.752981 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.753896 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 20:54:58 crc kubenswrapper[4754]: E1005 20:54:58.756204 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.756716 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.756761 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.756975 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.757010 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.757190 4754 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.757353 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.758156 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.758369 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.761618 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.767093 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.799456 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.803007 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.803062 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.803089 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.803113 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.803142 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.803169 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.803188 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.803207 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.803225 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.803244 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.803263 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.803290 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.803310 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.803332 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.803351 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.803417 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.803437 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.803463 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.803465 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.803483 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.803566 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.803589 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.803608 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.803626 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.803644 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.803661 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.803679 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.803700 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.803716 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.803734 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.803751 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.803768 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.803784 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.803806 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.803824 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.803842 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.803869 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.803901 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.803922 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.803940 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.803962 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.803986 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.804004 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.804022 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.804039 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.804059 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.804077 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.804095 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.804113 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.804132 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.804150 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.804189 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.804208 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.804231 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.804255 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.804278 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.804297 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.804316 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.804334 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.804347 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.804352 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.804436 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.804457 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.804474 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.804513 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.804533 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.804539 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.804549 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.804568 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.804589 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.804613 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.804631 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.804647 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.804664 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.804684 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.804699 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.804714 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.804731 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.804756 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.804773 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.804788 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.804804 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.804819 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.804798 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.804837 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.805004 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.805039 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.805062 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.805125 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.805151 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.805279 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.805307 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.805328 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.805354 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.805373 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.805406 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.805436 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.805459 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.805466 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.805478 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.805517 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.805543 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.805551 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.805564 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.805587 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.805607 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.805624 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.805646 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.805664 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.805682 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.805700 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.805716 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.805735 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.805753 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.805770 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.805789 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.805805 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.805861 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.805879 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.805881 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.805931 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.805956 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.805984 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.806013 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.806033 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.806054 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.806076 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.806096 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.806121 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.806123 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.806144 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.806169 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.806189 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.806208 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.806227 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.806245 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.806270 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.806297 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.806309 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.806324 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.806351 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.806372 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.806393 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.806416 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.806434 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.806438 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.806453 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.806475 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.806515 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.806537 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.806556 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.806577 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.806596 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.806618 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.806636 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.806654 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.806676 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.806710 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.806733 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.806754 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.806775 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.806798 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.806817 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.806837 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.806864 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.806883 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.806902 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.806920 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.806938 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.806957 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.806978 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.806999 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.807019 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.807043 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.807061 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.807081 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.807102 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.807122 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.807140 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.807158 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.807180 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.807200 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.807220 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.807238 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.807258 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.807277 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.807296 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.807315 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.807335 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.807353 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.807372 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.807390 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.807408 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.807426 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.807446 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.807465 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.807484 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.807517 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.807538 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.807558 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.807578 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.807602 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.807624 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.807643 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.807662 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.807682 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.807733 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.807757 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.807784 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.807806 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.807829 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.807853 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.807877 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.807896 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.807923 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.807943 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.807977 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.807999 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.808023 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.808043 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.808122 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.808135 4754 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.808147 4754 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.808158 4754 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.808169 4754 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.808178 4754 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.808188 4754 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.808198 4754 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.808222 4754 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.808233 4754 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.819762 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.821795 4754 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.824366 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.838333 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.842607 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.852118 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.806470 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.806655 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.806659 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.806814 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.806831 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.806997 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.807162 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.807199 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.856422 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.807453 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.807574 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.807718 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.807867 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.808010 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.808168 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.808276 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: E1005 20:54:58.808327 4754 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.808330 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.808397 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.808565 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: E1005 20:54:58.808589 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 20:54:59.308481572 +0000 UTC m=+23.212600482 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.808885 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.808934 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.808997 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.809370 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.809863 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.810047 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.811673 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.811731 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.811818 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.811889 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.812026 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.812059 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.812092 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.812209 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.812340 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.812350 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.812602 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.812611 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.812838 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.812844 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.812864 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.813034 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.813067 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.813093 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.813284 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.813390 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.813530 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.813674 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.813720 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.813952 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.814060 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.814241 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.814435 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.814329 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.814557 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.816591 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.816754 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.817006 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.817167 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.817486 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.818575 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.818596 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.818643 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.818794 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.819009 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.819045 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.819203 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.819310 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.819485 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.819642 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.819678 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.819878 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.819941 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.820106 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.820120 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.820309 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.820319 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.820455 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.820542 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.820772 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.820960 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.821062 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.821165 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.821518 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: E1005 20:54:58.824631 4754 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.829737 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.829917 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.832513 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.832786 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.833020 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.833699 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.834070 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.834122 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.834455 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.834697 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.834922 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.835378 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.837268 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.837884 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.838054 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.838242 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.838687 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.839343 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.839481 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.839833 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.839996 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.840049 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.840306 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.840559 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.840939 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.841669 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.841796 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.842674 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.843240 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.845647 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.846371 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.847739 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.847902 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.848039 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.848114 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.848439 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.849252 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.849607 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.851743 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.852338 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: E1005 20:54:58.852888 4754 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.854099 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.854727 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.854808 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.854920 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.854990 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.855127 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: E1005 20:54:58.855598 4754 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.855765 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.855940 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.856016 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.856105 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.856296 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.857468 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.861767 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.862752 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.863132 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.863385 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.863455 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.863534 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.863776 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.864218 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.864305 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: E1005 20:54:58.864546 4754 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 05 20:54:58 crc kubenswrapper[4754]: E1005 20:54:58.864567 4754 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 05 20:54:58 crc kubenswrapper[4754]: E1005 20:54:58.864573 4754 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 05 20:54:58 crc kubenswrapper[4754]: E1005 20:54:58.864592 4754 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.865210 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.865364 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.866332 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.866572 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.866686 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.866738 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.867209 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.867956 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: E1005 20:54:58.869670 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-05 20:54:59.369619997 +0000 UTC m=+23.273738707 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 05 20:54:58 crc kubenswrapper[4754]: E1005 20:54:58.870283 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-05 20:54:59.370244323 +0000 UTC m=+23.274363033 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 05 20:54:58 crc kubenswrapper[4754]: E1005 20:54:58.870443 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-05 20:54:59.370427138 +0000 UTC m=+23.274545838 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 05 20:54:58 crc kubenswrapper[4754]: E1005 20:54:58.870553 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-05 20:54:59.370533331 +0000 UTC m=+23.274652041 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.871461 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.875231 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.875516 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.877849 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.877844 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.878181 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.878457 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.880936 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.883353 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.883390 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.883868 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.884031 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.884185 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.885718 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.886346 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.886872 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.887122 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.887663 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.888665 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.892714 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.893097 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.893219 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.893653 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.893668 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.893921 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.893950 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.894050 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.894221 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.894312 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.894949 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.895129 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.895418 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.895950 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.895960 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.896078 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.897211 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.902925 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.907844 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.908711 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.908817 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.908930 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.908993 4754 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.909051 4754 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.909106 4754 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.909157 4754 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.909214 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.909270 4754 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.909321 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.909371 4754 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.909447 4754 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.909523 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.909590 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.909642 4754 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.909695 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.909758 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.909810 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.909864 4754 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.909913 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.909962 4754 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.910019 4754 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.910072 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.910122 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.910178 4754 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.910227 4754 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.910283 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.910333 4754 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.910383 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.910435 4754 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.910507 4754 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.910563 4754 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.910625 4754 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.910677 4754 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.910727 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.910783 4754 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.910838 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.910891 4754 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.910944 4754 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.910999 4754 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.911054 4754 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.911110 4754 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.911164 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.911213 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.911269 4754 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.911323 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.911373 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.911426 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.911477 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.911542 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.911605 4754 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.911655 4754 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.911713 4754 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.911767 4754 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.911818 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.911871 4754 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.911921 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.911970 4754 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.912024 4754 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.912078 4754 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.912128 4754 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.912179 4754 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.912237 4754 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.912298 4754 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.912348 4754 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.912400 4754 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.912454 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.912529 4754 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.912597 4754 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.912724 4754 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.912777 4754 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.912836 4754 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.912888 4754 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.912941 4754 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.912994 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.913049 4754 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.913103 4754 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.913154 4754 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.913208 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.913259 4754 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.913322 4754 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.913395 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.913454 4754 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.913524 4754 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.913585 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.913638 4754 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.913695 4754 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.913746 4754 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.913799 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.913852 4754 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.913905 4754 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.913963 4754 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.914017 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.914068 4754 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.914117 4754 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.914179 4754 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.914233 4754 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.914295 4754 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.914347 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.914396 4754 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.914446 4754 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.914528 4754 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.914583 4754 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.914646 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.914702 4754 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.914757 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.914810 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.914863 4754 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.914916 4754 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.924122 4754 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.924187 4754 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.924247 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.924316 4754 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.924369 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.924437 4754 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.924502 4754 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.924562 4754 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.924614 4754 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.924672 4754 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.924800 4754 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.924868 4754 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.924927 4754 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.924979 4754 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.925044 4754 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.925105 4754 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.925158 4754 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.925219 4754 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.925282 4754 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.925334 4754 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.925404 4754 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.925462 4754 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.925551 4754 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.925614 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.925671 4754 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.925726 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.927534 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.927662 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.927767 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.929509 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.925788 4754 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.933701 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.933725 4754 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.933739 4754 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.933768 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.933780 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.933790 4754 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.933807 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.933825 4754 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.933838 4754 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.933848 4754 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.933858 4754 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.933872 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.933883 4754 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.933894 4754 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.933910 4754 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.933920 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.933931 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.933941 4754 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.933956 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.933967 4754 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.934284 4754 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.934301 4754 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.934318 4754 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.934332 4754 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.934342 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.934352 4754 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.934364 4754 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.934374 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.934385 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.934397 4754 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.934407 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.934416 4754 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.934426 4754 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.934438 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.934447 4754 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.934457 4754 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.934467 4754 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.934483 4754 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.934510 4754 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.934520 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.934534 4754 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.934552 4754 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.934565 4754 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.934577 4754 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.934590 4754 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.934600 4754 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.935907 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.952974 4754 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.953106 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.958895 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.959383 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.959845 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.961486 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.967790 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.968253 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.968784 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.969421 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.971641 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.972086 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.973028 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.974841 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.975419 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.977591 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.978118 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.978985 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.979700 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.983537 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.985656 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.986433 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.987155 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.988348 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.989084 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.990130 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.990720 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-57sr4"] Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.991142 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-57sr4" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.994441 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.994831 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Oct 05 20:54:58 crc kubenswrapper[4754]: I1005 20:54:58.995053 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.018558 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.019687 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.022128 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.026814 4754 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="8907b9b69b249c34ec6db1e460d641b31711b75d6d31d4caafd867339e6571d0" exitCode=255 Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.026887 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"8907b9b69b249c34ec6db1e460d641b31711b75d6d31d4caafd867339e6571d0"} Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.026984 4754 scope.go:117] "RemoveContainer" containerID="9d9b0b384ffcc4040a3cc8aea4be89ded347240018f198471da1922681c8efec" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.035875 4754 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.035899 4754 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.035910 4754 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.035922 4754 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.051936 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.068754 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.070955 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.071373 4754 scope.go:117] "RemoveContainer" containerID="8907b9b69b249c34ec6db1e460d641b31711b75d6d31d4caafd867339e6571d0" Oct 05 20:54:59 crc kubenswrapper[4754]: E1005 20:54:59.071550 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.081417 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.088908 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.106396 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 20:54:59 crc kubenswrapper[4754]: W1005 20:54:59.120380 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-c4e51bd877a767cbcffdfd35e9de2de7690c934b7216a8578e9d1916fc03fb51 WatchSource:0}: Error finding container c4e51bd877a767cbcffdfd35e9de2de7690c934b7216a8578e9d1916fc03fb51: Status 404 returned error can't find the container with id c4e51bd877a767cbcffdfd35e9de2de7690c934b7216a8578e9d1916fc03fb51 Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.131559 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.137835 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vw2ps\" (UniqueName: \"kubernetes.io/projected/e854c7b9-1229-4da4-80e9-fea6479459b8-kube-api-access-vw2ps\") pod \"node-resolver-57sr4\" (UID: \"e854c7b9-1229-4da4-80e9-fea6479459b8\") " pod="openshift-dns/node-resolver-57sr4" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.137877 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/e854c7b9-1229-4da4-80e9-fea6479459b8-hosts-file\") pod \"node-resolver-57sr4\" (UID: \"e854c7b9-1229-4da4-80e9-fea6479459b8\") " pod="openshift-dns/node-resolver-57sr4" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.151870 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.177832 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.187815 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-57sr4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e854c7b9-1229-4da4-80e9-fea6479459b8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vw2ps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:58Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-57sr4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.202197 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.218774 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.238276 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vw2ps\" (UniqueName: \"kubernetes.io/projected/e854c7b9-1229-4da4-80e9-fea6479459b8-kube-api-access-vw2ps\") pod \"node-resolver-57sr4\" (UID: \"e854c7b9-1229-4da4-80e9-fea6479459b8\") " pod="openshift-dns/node-resolver-57sr4" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.238328 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/e854c7b9-1229-4da4-80e9-fea6479459b8-hosts-file\") pod \"node-resolver-57sr4\" (UID: \"e854c7b9-1229-4da4-80e9-fea6479459b8\") " pod="openshift-dns/node-resolver-57sr4" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.238422 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/e854c7b9-1229-4da4-80e9-fea6479459b8-hosts-file\") pod \"node-resolver-57sr4\" (UID: \"e854c7b9-1229-4da4-80e9-fea6479459b8\") " pod="openshift-dns/node-resolver-57sr4" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.258389 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vw2ps\" (UniqueName: \"kubernetes.io/projected/e854c7b9-1229-4da4-80e9-fea6479459b8-kube-api-access-vw2ps\") pod \"node-resolver-57sr4\" (UID: \"e854c7b9-1229-4da4-80e9-fea6479459b8\") " pod="openshift-dns/node-resolver-57sr4" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.312103 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-57sr4" Oct 05 20:54:59 crc kubenswrapper[4754]: W1005 20:54:59.333133 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode854c7b9_1229_4da4_80e9_fea6479459b8.slice/crio-26d9e615d742750e95d13d37f7dbc3da5b7bfc5a005c4466ffbce353232924ca WatchSource:0}: Error finding container 26d9e615d742750e95d13d37f7dbc3da5b7bfc5a005c4466ffbce353232924ca: Status 404 returned error can't find the container with id 26d9e615d742750e95d13d37f7dbc3da5b7bfc5a005c4466ffbce353232924ca Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.339585 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 20:54:59 crc kubenswrapper[4754]: E1005 20:54:59.339860 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 20:55:00.339839569 +0000 UTC m=+24.243958279 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.379790 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-b2h9k"] Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.380148 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-tskt7"] Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.380710 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-tskt7" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.381058 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.382300 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-cn76s"] Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.382452 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-cn76s" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.383436 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.383947 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.384255 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.384533 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.385038 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.385325 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.385591 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.385884 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.386173 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.386446 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.390267 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.393235 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.409256 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.421961 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.441378 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.441439 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.441474 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.441531 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 20:54:59 crc kubenswrapper[4754]: E1005 20:54:59.441748 4754 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 05 20:54:59 crc kubenswrapper[4754]: E1005 20:54:59.441790 4754 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 05 20:54:59 crc kubenswrapper[4754]: E1005 20:54:59.441853 4754 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 05 20:54:59 crc kubenswrapper[4754]: E1005 20:54:59.441847 4754 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 05 20:54:59 crc kubenswrapper[4754]: E1005 20:54:59.441872 4754 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 05 20:54:59 crc kubenswrapper[4754]: E1005 20:54:59.441857 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-05 20:55:00.441830584 +0000 UTC m=+24.345949294 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 05 20:54:59 crc kubenswrapper[4754]: E1005 20:54:59.441963 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-05 20:55:00.441934047 +0000 UTC m=+24.346052757 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 05 20:54:59 crc kubenswrapper[4754]: E1005 20:54:59.441978 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-05 20:55:00.441971698 +0000 UTC m=+24.346090408 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 05 20:54:59 crc kubenswrapper[4754]: E1005 20:54:59.442117 4754 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 05 20:54:59 crc kubenswrapper[4754]: E1005 20:54:59.442174 4754 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 05 20:54:59 crc kubenswrapper[4754]: E1005 20:54:59.442306 4754 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 05 20:54:59 crc kubenswrapper[4754]: E1005 20:54:59.442374 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-05 20:55:00.442347968 +0000 UTC m=+24.346466668 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.446276 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.456172 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-57sr4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e854c7b9-1229-4da4-80e9-fea6479459b8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vw2ps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:58Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-57sr4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.465950 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7506d496-65ff-41dc-8230-24c8ee653a2a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33da5a6bb3cd3b042c193bc571e6622f2f87f27e560cdd0bbb9be31e2010d707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df6d700a947d19c84b4f65fd2dd2f8d1f4f1f80fa9b4800c2acfa8dae3912639\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35ca8ee77a0e58cf969bd67b277a3392614c4e73b963e6f6002097709370715d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8907b9b69b249c34ec6db1e460d641b31711b75d6d31d4caafd867339e6571d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9d9b0b384ffcc4040a3cc8aea4be89ded347240018f198471da1922681c8efec\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T20:54:40Z\\\",\\\"message\\\":\\\"W1005 20:54:40.114418 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1005 20:54:40.115271 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759697680 cert, and key in /tmp/serving-cert-3287406251/serving-signer.crt, /tmp/serving-cert-3287406251/serving-signer.key\\\\nI1005 20:54:40.625469 1 observer_polling.go:159] Starting file observer\\\\nW1005 20:54:40.628724 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1005 20:54:40.628993 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 20:54:40.631041 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3287406251/tls.crt::/tmp/serving-cert-3287406251/tls.key\\\\\\\"\\\\nF1005 20:54:40.864813 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:40Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8907b9b69b249c34ec6db1e460d641b31711b75d6d31d4caafd867339e6571d0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1005 20:54:52.589115 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 20:54:52.590181 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1304624737/tls.crt::/tmp/serving-cert-1304624737/tls.key\\\\\\\"\\\\nI1005 20:54:58.070544 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1005 20:54:58.090352 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1005 20:54:58.090382 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1005 20:54:58.090412 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1005 20:54:58.090418 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1005 20:54:58.144656 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1005 20:54:58.144688 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 20:54:58.144695 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 20:54:58.144699 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1005 20:54:58.144702 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1005 20:54:58.144705 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1005 20:54:58.144708 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1005 20:54:58.145024 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1005 20:54:58.148112 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4efb34f3493cc8854bc4cff6b29f63973c52b754404d7a21daec4e1f612d327d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54c568b637f3737667433047b956c7eb22fdf0f9301acdfd3b49fb84c39de6ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54c568b637f3737667433047b956c7eb22fdf0f9301acdfd3b49fb84c39de6ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.478760 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.495673 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tskt7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a68e9c89-465b-4c01-a470-e21a480ee465\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tskt7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.505146 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be95a413-6a59-45b4-84b7-b43cae694a26\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2ggg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2ggg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-b2h9k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.515792 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.525252 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.536596 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.542565 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/be95a413-6a59-45b4-84b7-b43cae694a26-mcd-auth-proxy-config\") pod \"machine-config-daemon-b2h9k\" (UID: \"be95a413-6a59-45b4-84b7-b43cae694a26\") " pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.542615 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-24dst\" (UniqueName: \"kubernetes.io/projected/a68e9c89-465b-4c01-a470-e21a480ee465-kube-api-access-24dst\") pod \"multus-additional-cni-plugins-tskt7\" (UID: \"a68e9c89-465b-4c01-a470-e21a480ee465\") " pod="openshift-multus/multus-additional-cni-plugins-tskt7" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.542673 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/a68e9c89-465b-4c01-a470-e21a480ee465-os-release\") pod \"multus-additional-cni-plugins-tskt7\" (UID: \"a68e9c89-465b-4c01-a470-e21a480ee465\") " pod="openshift-multus/multus-additional-cni-plugins-tskt7" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.542714 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/02d5e3f9-73c2-4496-9aca-0787184aef19-system-cni-dir\") pod \"multus-cn76s\" (UID: \"02d5e3f9-73c2-4496-9aca-0787184aef19\") " pod="openshift-multus/multus-cn76s" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.542743 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/02d5e3f9-73c2-4496-9aca-0787184aef19-host-run-multus-certs\") pod \"multus-cn76s\" (UID: \"02d5e3f9-73c2-4496-9aca-0787184aef19\") " pod="openshift-multus/multus-cn76s" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.542762 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/be95a413-6a59-45b4-84b7-b43cae694a26-proxy-tls\") pod \"machine-config-daemon-b2h9k\" (UID: \"be95a413-6a59-45b4-84b7-b43cae694a26\") " pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.542780 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/02d5e3f9-73c2-4496-9aca-0787184aef19-etc-kubernetes\") pod \"multus-cn76s\" (UID: \"02d5e3f9-73c2-4496-9aca-0787184aef19\") " pod="openshift-multus/multus-cn76s" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.542798 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/a68e9c89-465b-4c01-a470-e21a480ee465-system-cni-dir\") pod \"multus-additional-cni-plugins-tskt7\" (UID: \"a68e9c89-465b-4c01-a470-e21a480ee465\") " pod="openshift-multus/multus-additional-cni-plugins-tskt7" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.542823 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/a68e9c89-465b-4c01-a470-e21a480ee465-cni-binary-copy\") pod \"multus-additional-cni-plugins-tskt7\" (UID: \"a68e9c89-465b-4c01-a470-e21a480ee465\") " pod="openshift-multus/multus-additional-cni-plugins-tskt7" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.542842 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/02d5e3f9-73c2-4496-9aca-0787184aef19-cni-binary-copy\") pod \"multus-cn76s\" (UID: \"02d5e3f9-73c2-4496-9aca-0787184aef19\") " pod="openshift-multus/multus-cn76s" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.542866 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m2ggg\" (UniqueName: \"kubernetes.io/projected/be95a413-6a59-45b4-84b7-b43cae694a26-kube-api-access-m2ggg\") pod \"machine-config-daemon-b2h9k\" (UID: \"be95a413-6a59-45b4-84b7-b43cae694a26\") " pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.542998 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/02d5e3f9-73c2-4496-9aca-0787184aef19-host-run-netns\") pod \"multus-cn76s\" (UID: \"02d5e3f9-73c2-4496-9aca-0787184aef19\") " pod="openshift-multus/multus-cn76s" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.543020 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/02d5e3f9-73c2-4496-9aca-0787184aef19-multus-conf-dir\") pod \"multus-cn76s\" (UID: \"02d5e3f9-73c2-4496-9aca-0787184aef19\") " pod="openshift-multus/multus-cn76s" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.543069 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bpx5r\" (UniqueName: \"kubernetes.io/projected/02d5e3f9-73c2-4496-9aca-0787184aef19-kube-api-access-bpx5r\") pod \"multus-cn76s\" (UID: \"02d5e3f9-73c2-4496-9aca-0787184aef19\") " pod="openshift-multus/multus-cn76s" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.543088 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/02d5e3f9-73c2-4496-9aca-0787184aef19-host-var-lib-kubelet\") pod \"multus-cn76s\" (UID: \"02d5e3f9-73c2-4496-9aca-0787184aef19\") " pod="openshift-multus/multus-cn76s" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.543152 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/02d5e3f9-73c2-4496-9aca-0787184aef19-multus-cni-dir\") pod \"multus-cn76s\" (UID: \"02d5e3f9-73c2-4496-9aca-0787184aef19\") " pod="openshift-multus/multus-cn76s" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.543180 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/02d5e3f9-73c2-4496-9aca-0787184aef19-cnibin\") pod \"multus-cn76s\" (UID: \"02d5e3f9-73c2-4496-9aca-0787184aef19\") " pod="openshift-multus/multus-cn76s" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.543202 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/02d5e3f9-73c2-4496-9aca-0787184aef19-hostroot\") pod \"multus-cn76s\" (UID: \"02d5e3f9-73c2-4496-9aca-0787184aef19\") " pod="openshift-multus/multus-cn76s" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.543225 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/02d5e3f9-73c2-4496-9aca-0787184aef19-multus-daemon-config\") pod \"multus-cn76s\" (UID: \"02d5e3f9-73c2-4496-9aca-0787184aef19\") " pod="openshift-multus/multus-cn76s" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.543246 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/a68e9c89-465b-4c01-a470-e21a480ee465-tuning-conf-dir\") pod \"multus-additional-cni-plugins-tskt7\" (UID: \"a68e9c89-465b-4c01-a470-e21a480ee465\") " pod="openshift-multus/multus-additional-cni-plugins-tskt7" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.543267 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/be95a413-6a59-45b4-84b7-b43cae694a26-rootfs\") pod \"machine-config-daemon-b2h9k\" (UID: \"be95a413-6a59-45b4-84b7-b43cae694a26\") " pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.543285 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/02d5e3f9-73c2-4496-9aca-0787184aef19-multus-socket-dir-parent\") pod \"multus-cn76s\" (UID: \"02d5e3f9-73c2-4496-9aca-0787184aef19\") " pod="openshift-multus/multus-cn76s" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.543304 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/02d5e3f9-73c2-4496-9aca-0787184aef19-host-run-k8s-cni-cncf-io\") pod \"multus-cn76s\" (UID: \"02d5e3f9-73c2-4496-9aca-0787184aef19\") " pod="openshift-multus/multus-cn76s" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.543319 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/a68e9c89-465b-4c01-a470-e21a480ee465-cnibin\") pod \"multus-additional-cni-plugins-tskt7\" (UID: \"a68e9c89-465b-4c01-a470-e21a480ee465\") " pod="openshift-multus/multus-additional-cni-plugins-tskt7" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.543338 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/a68e9c89-465b-4c01-a470-e21a480ee465-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-tskt7\" (UID: \"a68e9c89-465b-4c01-a470-e21a480ee465\") " pod="openshift-multus/multus-additional-cni-plugins-tskt7" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.543358 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/02d5e3f9-73c2-4496-9aca-0787184aef19-host-var-lib-cni-bin\") pod \"multus-cn76s\" (UID: \"02d5e3f9-73c2-4496-9aca-0787184aef19\") " pod="openshift-multus/multus-cn76s" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.543388 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/02d5e3f9-73c2-4496-9aca-0787184aef19-os-release\") pod \"multus-cn76s\" (UID: \"02d5e3f9-73c2-4496-9aca-0787184aef19\") " pod="openshift-multus/multus-cn76s" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.543409 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/02d5e3f9-73c2-4496-9aca-0787184aef19-host-var-lib-cni-multus\") pod \"multus-cn76s\" (UID: \"02d5e3f9-73c2-4496-9aca-0787184aef19\") " pod="openshift-multus/multus-cn76s" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.546954 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.557134 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.570279 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.580860 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.590867 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-57sr4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e854c7b9-1229-4da4-80e9-fea6479459b8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vw2ps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:58Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-57sr4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.606157 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cn76s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02d5e3f9-73c2-4496-9aca-0787184aef19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpx5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cn76s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.621927 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7506d496-65ff-41dc-8230-24c8ee653a2a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33da5a6bb3cd3b042c193bc571e6622f2f87f27e560cdd0bbb9be31e2010d707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df6d700a947d19c84b4f65fd2dd2f8d1f4f1f80fa9b4800c2acfa8dae3912639\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35ca8ee77a0e58cf969bd67b277a3392614c4e73b963e6f6002097709370715d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8907b9b69b249c34ec6db1e460d641b31711b75d6d31d4caafd867339e6571d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9d9b0b384ffcc4040a3cc8aea4be89ded347240018f198471da1922681c8efec\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T20:54:40Z\\\",\\\"message\\\":\\\"W1005 20:54:40.114418 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1005 20:54:40.115271 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759697680 cert, and key in /tmp/serving-cert-3287406251/serving-signer.crt, /tmp/serving-cert-3287406251/serving-signer.key\\\\nI1005 20:54:40.625469 1 observer_polling.go:159] Starting file observer\\\\nW1005 20:54:40.628724 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1005 20:54:40.628993 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 20:54:40.631041 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3287406251/tls.crt::/tmp/serving-cert-3287406251/tls.key\\\\\\\"\\\\nF1005 20:54:40.864813 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:40Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8907b9b69b249c34ec6db1e460d641b31711b75d6d31d4caafd867339e6571d0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1005 20:54:52.589115 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 20:54:52.590181 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1304624737/tls.crt::/tmp/serving-cert-1304624737/tls.key\\\\\\\"\\\\nI1005 20:54:58.070544 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1005 20:54:58.090352 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1005 20:54:58.090382 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1005 20:54:58.090412 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1005 20:54:58.090418 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1005 20:54:58.144656 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1005 20:54:58.144688 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 20:54:58.144695 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 20:54:58.144699 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1005 20:54:58.144702 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1005 20:54:58.144705 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1005 20:54:58.144708 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1005 20:54:58.145024 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1005 20:54:58.148112 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4efb34f3493cc8854bc4cff6b29f63973c52b754404d7a21daec4e1f612d327d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54c568b637f3737667433047b956c7eb22fdf0f9301acdfd3b49fb84c39de6ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54c568b637f3737667433047b956c7eb22fdf0f9301acdfd3b49fb84c39de6ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.635517 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.644561 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/02d5e3f9-73c2-4496-9aca-0787184aef19-cni-binary-copy\") pod \"multus-cn76s\" (UID: \"02d5e3f9-73c2-4496-9aca-0787184aef19\") " pod="openshift-multus/multus-cn76s" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.644641 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m2ggg\" (UniqueName: \"kubernetes.io/projected/be95a413-6a59-45b4-84b7-b43cae694a26-kube-api-access-m2ggg\") pod \"machine-config-daemon-b2h9k\" (UID: \"be95a413-6a59-45b4-84b7-b43cae694a26\") " pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.644671 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/02d5e3f9-73c2-4496-9aca-0787184aef19-host-run-netns\") pod \"multus-cn76s\" (UID: \"02d5e3f9-73c2-4496-9aca-0787184aef19\") " pod="openshift-multus/multus-cn76s" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.644698 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/02d5e3f9-73c2-4496-9aca-0787184aef19-multus-conf-dir\") pod \"multus-cn76s\" (UID: \"02d5e3f9-73c2-4496-9aca-0787184aef19\") " pod="openshift-multus/multus-cn76s" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.644725 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bpx5r\" (UniqueName: \"kubernetes.io/projected/02d5e3f9-73c2-4496-9aca-0787184aef19-kube-api-access-bpx5r\") pod \"multus-cn76s\" (UID: \"02d5e3f9-73c2-4496-9aca-0787184aef19\") " pod="openshift-multus/multus-cn76s" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.644758 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/02d5e3f9-73c2-4496-9aca-0787184aef19-host-var-lib-kubelet\") pod \"multus-cn76s\" (UID: \"02d5e3f9-73c2-4496-9aca-0787184aef19\") " pod="openshift-multus/multus-cn76s" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.644789 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/02d5e3f9-73c2-4496-9aca-0787184aef19-hostroot\") pod \"multus-cn76s\" (UID: \"02d5e3f9-73c2-4496-9aca-0787184aef19\") " pod="openshift-multus/multus-cn76s" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.644814 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/02d5e3f9-73c2-4496-9aca-0787184aef19-multus-daemon-config\") pod \"multus-cn76s\" (UID: \"02d5e3f9-73c2-4496-9aca-0787184aef19\") " pod="openshift-multus/multus-cn76s" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.644840 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/02d5e3f9-73c2-4496-9aca-0787184aef19-multus-cni-dir\") pod \"multus-cn76s\" (UID: \"02d5e3f9-73c2-4496-9aca-0787184aef19\") " pod="openshift-multus/multus-cn76s" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.644868 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/02d5e3f9-73c2-4496-9aca-0787184aef19-cnibin\") pod \"multus-cn76s\" (UID: \"02d5e3f9-73c2-4496-9aca-0787184aef19\") " pod="openshift-multus/multus-cn76s" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.644895 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/a68e9c89-465b-4c01-a470-e21a480ee465-tuning-conf-dir\") pod \"multus-additional-cni-plugins-tskt7\" (UID: \"a68e9c89-465b-4c01-a470-e21a480ee465\") " pod="openshift-multus/multus-additional-cni-plugins-tskt7" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.644932 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/02d5e3f9-73c2-4496-9aca-0787184aef19-multus-socket-dir-parent\") pod \"multus-cn76s\" (UID: \"02d5e3f9-73c2-4496-9aca-0787184aef19\") " pod="openshift-multus/multus-cn76s" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.644964 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/02d5e3f9-73c2-4496-9aca-0787184aef19-host-run-k8s-cni-cncf-io\") pod \"multus-cn76s\" (UID: \"02d5e3f9-73c2-4496-9aca-0787184aef19\") " pod="openshift-multus/multus-cn76s" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.644999 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/a68e9c89-465b-4c01-a470-e21a480ee465-cnibin\") pod \"multus-additional-cni-plugins-tskt7\" (UID: \"a68e9c89-465b-4c01-a470-e21a480ee465\") " pod="openshift-multus/multus-additional-cni-plugins-tskt7" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.645063 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/a68e9c89-465b-4c01-a470-e21a480ee465-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-tskt7\" (UID: \"a68e9c89-465b-4c01-a470-e21a480ee465\") " pod="openshift-multus/multus-additional-cni-plugins-tskt7" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.645101 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/be95a413-6a59-45b4-84b7-b43cae694a26-rootfs\") pod \"machine-config-daemon-b2h9k\" (UID: \"be95a413-6a59-45b4-84b7-b43cae694a26\") " pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.645145 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/02d5e3f9-73c2-4496-9aca-0787184aef19-host-var-lib-cni-bin\") pod \"multus-cn76s\" (UID: \"02d5e3f9-73c2-4496-9aca-0787184aef19\") " pod="openshift-multus/multus-cn76s" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.645198 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/02d5e3f9-73c2-4496-9aca-0787184aef19-os-release\") pod \"multus-cn76s\" (UID: \"02d5e3f9-73c2-4496-9aca-0787184aef19\") " pod="openshift-multus/multus-cn76s" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.645223 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/02d5e3f9-73c2-4496-9aca-0787184aef19-host-var-lib-cni-multus\") pod \"multus-cn76s\" (UID: \"02d5e3f9-73c2-4496-9aca-0787184aef19\") " pod="openshift-multus/multus-cn76s" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.645249 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/be95a413-6a59-45b4-84b7-b43cae694a26-mcd-auth-proxy-config\") pod \"machine-config-daemon-b2h9k\" (UID: \"be95a413-6a59-45b4-84b7-b43cae694a26\") " pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.645275 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-24dst\" (UniqueName: \"kubernetes.io/projected/a68e9c89-465b-4c01-a470-e21a480ee465-kube-api-access-24dst\") pod \"multus-additional-cni-plugins-tskt7\" (UID: \"a68e9c89-465b-4c01-a470-e21a480ee465\") " pod="openshift-multus/multus-additional-cni-plugins-tskt7" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.645326 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/a68e9c89-465b-4c01-a470-e21a480ee465-os-release\") pod \"multus-additional-cni-plugins-tskt7\" (UID: \"a68e9c89-465b-4c01-a470-e21a480ee465\") " pod="openshift-multus/multus-additional-cni-plugins-tskt7" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.645359 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/02d5e3f9-73c2-4496-9aca-0787184aef19-system-cni-dir\") pod \"multus-cn76s\" (UID: \"02d5e3f9-73c2-4496-9aca-0787184aef19\") " pod="openshift-multus/multus-cn76s" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.645382 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/02d5e3f9-73c2-4496-9aca-0787184aef19-host-run-multus-certs\") pod \"multus-cn76s\" (UID: \"02d5e3f9-73c2-4496-9aca-0787184aef19\") " pod="openshift-multus/multus-cn76s" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.645427 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/be95a413-6a59-45b4-84b7-b43cae694a26-proxy-tls\") pod \"machine-config-daemon-b2h9k\" (UID: \"be95a413-6a59-45b4-84b7-b43cae694a26\") " pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.645451 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/02d5e3f9-73c2-4496-9aca-0787184aef19-etc-kubernetes\") pod \"multus-cn76s\" (UID: \"02d5e3f9-73c2-4496-9aca-0787184aef19\") " pod="openshift-multus/multus-cn76s" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.645477 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/a68e9c89-465b-4c01-a470-e21a480ee465-system-cni-dir\") pod \"multus-additional-cni-plugins-tskt7\" (UID: \"a68e9c89-465b-4c01-a470-e21a480ee465\") " pod="openshift-multus/multus-additional-cni-plugins-tskt7" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.645584 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/a68e9c89-465b-4c01-a470-e21a480ee465-system-cni-dir\") pod \"multus-additional-cni-plugins-tskt7\" (UID: \"a68e9c89-465b-4c01-a470-e21a480ee465\") " pod="openshift-multus/multus-additional-cni-plugins-tskt7" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.646155 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/02d5e3f9-73c2-4496-9aca-0787184aef19-host-run-netns\") pod \"multus-cn76s\" (UID: \"02d5e3f9-73c2-4496-9aca-0787184aef19\") " pod="openshift-multus/multus-cn76s" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.646214 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/02d5e3f9-73c2-4496-9aca-0787184aef19-multus-conf-dir\") pod \"multus-cn76s\" (UID: \"02d5e3f9-73c2-4496-9aca-0787184aef19\") " pod="openshift-multus/multus-cn76s" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.646372 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/02d5e3f9-73c2-4496-9aca-0787184aef19-host-var-lib-kubelet\") pod \"multus-cn76s\" (UID: \"02d5e3f9-73c2-4496-9aca-0787184aef19\") " pod="openshift-multus/multus-cn76s" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.646409 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/02d5e3f9-73c2-4496-9aca-0787184aef19-hostroot\") pod \"multus-cn76s\" (UID: \"02d5e3f9-73c2-4496-9aca-0787184aef19\") " pod="openshift-multus/multus-cn76s" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.646582 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/02d5e3f9-73c2-4496-9aca-0787184aef19-cni-binary-copy\") pod \"multus-cn76s\" (UID: \"02d5e3f9-73c2-4496-9aca-0787184aef19\") " pod="openshift-multus/multus-cn76s" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.646647 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/a68e9c89-465b-4c01-a470-e21a480ee465-cni-binary-copy\") pod \"multus-additional-cni-plugins-tskt7\" (UID: \"a68e9c89-465b-4c01-a470-e21a480ee465\") " pod="openshift-multus/multus-additional-cni-plugins-tskt7" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.646958 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/02d5e3f9-73c2-4496-9aca-0787184aef19-multus-daemon-config\") pod \"multus-cn76s\" (UID: \"02d5e3f9-73c2-4496-9aca-0787184aef19\") " pod="openshift-multus/multus-cn76s" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.647034 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/02d5e3f9-73c2-4496-9aca-0787184aef19-multus-cni-dir\") pod \"multus-cn76s\" (UID: \"02d5e3f9-73c2-4496-9aca-0787184aef19\") " pod="openshift-multus/multus-cn76s" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.647080 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/02d5e3f9-73c2-4496-9aca-0787184aef19-cnibin\") pod \"multus-cn76s\" (UID: \"02d5e3f9-73c2-4496-9aca-0787184aef19\") " pod="openshift-multus/multus-cn76s" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.647367 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/a68e9c89-465b-4c01-a470-e21a480ee465-cni-binary-copy\") pod \"multus-additional-cni-plugins-tskt7\" (UID: \"a68e9c89-465b-4c01-a470-e21a480ee465\") " pod="openshift-multus/multus-additional-cni-plugins-tskt7" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.647452 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/02d5e3f9-73c2-4496-9aca-0787184aef19-host-var-lib-cni-multus\") pod \"multus-cn76s\" (UID: \"02d5e3f9-73c2-4496-9aca-0787184aef19\") " pod="openshift-multus/multus-cn76s" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.647525 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/a68e9c89-465b-4c01-a470-e21a480ee465-tuning-conf-dir\") pod \"multus-additional-cni-plugins-tskt7\" (UID: \"a68e9c89-465b-4c01-a470-e21a480ee465\") " pod="openshift-multus/multus-additional-cni-plugins-tskt7" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.647465 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/02d5e3f9-73c2-4496-9aca-0787184aef19-multus-socket-dir-parent\") pod \"multus-cn76s\" (UID: \"02d5e3f9-73c2-4496-9aca-0787184aef19\") " pod="openshift-multus/multus-cn76s" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.647608 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/a68e9c89-465b-4c01-a470-e21a480ee465-cnibin\") pod \"multus-additional-cni-plugins-tskt7\" (UID: \"a68e9c89-465b-4c01-a470-e21a480ee465\") " pod="openshift-multus/multus-additional-cni-plugins-tskt7" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.647829 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/02d5e3f9-73c2-4496-9aca-0787184aef19-system-cni-dir\") pod \"multus-cn76s\" (UID: \"02d5e3f9-73c2-4496-9aca-0787184aef19\") " pod="openshift-multus/multus-cn76s" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.647846 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/be95a413-6a59-45b4-84b7-b43cae694a26-rootfs\") pod \"machine-config-daemon-b2h9k\" (UID: \"be95a413-6a59-45b4-84b7-b43cae694a26\") " pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.647878 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/02d5e3f9-73c2-4496-9aca-0787184aef19-host-var-lib-cni-bin\") pod \"multus-cn76s\" (UID: \"02d5e3f9-73c2-4496-9aca-0787184aef19\") " pod="openshift-multus/multus-cn76s" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.647878 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/a68e9c89-465b-4c01-a470-e21a480ee465-os-release\") pod \"multus-additional-cni-plugins-tskt7\" (UID: \"a68e9c89-465b-4c01-a470-e21a480ee465\") " pod="openshift-multus/multus-additional-cni-plugins-tskt7" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.647896 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/02d5e3f9-73c2-4496-9aca-0787184aef19-host-run-multus-certs\") pod \"multus-cn76s\" (UID: \"02d5e3f9-73c2-4496-9aca-0787184aef19\") " pod="openshift-multus/multus-cn76s" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.647917 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/02d5e3f9-73c2-4496-9aca-0787184aef19-etc-kubernetes\") pod \"multus-cn76s\" (UID: \"02d5e3f9-73c2-4496-9aca-0787184aef19\") " pod="openshift-multus/multus-cn76s" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.647918 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/02d5e3f9-73c2-4496-9aca-0787184aef19-os-release\") pod \"multus-cn76s\" (UID: \"02d5e3f9-73c2-4496-9aca-0787184aef19\") " pod="openshift-multus/multus-cn76s" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.647957 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/02d5e3f9-73c2-4496-9aca-0787184aef19-host-run-k8s-cni-cncf-io\") pod \"multus-cn76s\" (UID: \"02d5e3f9-73c2-4496-9aca-0787184aef19\") " pod="openshift-multus/multus-cn76s" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.648466 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/be95a413-6a59-45b4-84b7-b43cae694a26-mcd-auth-proxy-config\") pod \"machine-config-daemon-b2h9k\" (UID: \"be95a413-6a59-45b4-84b7-b43cae694a26\") " pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.648725 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/a68e9c89-465b-4c01-a470-e21a480ee465-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-tskt7\" (UID: \"a68e9c89-465b-4c01-a470-e21a480ee465\") " pod="openshift-multus/multus-additional-cni-plugins-tskt7" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.653121 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/be95a413-6a59-45b4-84b7-b43cae694a26-proxy-tls\") pod \"machine-config-daemon-b2h9k\" (UID: \"be95a413-6a59-45b4-84b7-b43cae694a26\") " pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.666574 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m2ggg\" (UniqueName: \"kubernetes.io/projected/be95a413-6a59-45b4-84b7-b43cae694a26-kube-api-access-m2ggg\") pod \"machine-config-daemon-b2h9k\" (UID: \"be95a413-6a59-45b4-84b7-b43cae694a26\") " pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.669951 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bpx5r\" (UniqueName: \"kubernetes.io/projected/02d5e3f9-73c2-4496-9aca-0787184aef19-kube-api-access-bpx5r\") pod \"multus-cn76s\" (UID: \"02d5e3f9-73c2-4496-9aca-0787184aef19\") " pod="openshift-multus/multus-cn76s" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.672349 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tskt7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a68e9c89-465b-4c01-a470-e21a480ee465\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tskt7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.674057 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-24dst\" (UniqueName: \"kubernetes.io/projected/a68e9c89-465b-4c01-a470-e21a480ee465-kube-api-access-24dst\") pod \"multus-additional-cni-plugins-tskt7\" (UID: \"a68e9c89-465b-4c01-a470-e21a480ee465\") " pod="openshift-multus/multus-additional-cni-plugins-tskt7" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.682214 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be95a413-6a59-45b4-84b7-b43cae694a26\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2ggg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2ggg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-b2h9k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.712324 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-tskt7" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.722251 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" Oct 05 20:54:59 crc kubenswrapper[4754]: W1005 20:54:59.726190 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda68e9c89_465b_4c01_a470_e21a480ee465.slice/crio-6b6b7f27c10d9b52260897b10d5fd0c3584eeea9e77f6765470a6dd8c813fe81 WatchSource:0}: Error finding container 6b6b7f27c10d9b52260897b10d5fd0c3584eeea9e77f6765470a6dd8c813fe81: Status 404 returned error can't find the container with id 6b6b7f27c10d9b52260897b10d5fd0c3584eeea9e77f6765470a6dd8c813fe81 Oct 05 20:54:59 crc kubenswrapper[4754]: W1005 20:54:59.736527 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbe95a413_6a59_45b4_84b7_b43cae694a26.slice/crio-623438fe73a4f207ac2907dda4b138a7c0a60c98f2e75f3f9fa56fee7d70f94b WatchSource:0}: Error finding container 623438fe73a4f207ac2907dda4b138a7c0a60c98f2e75f3f9fa56fee7d70f94b: Status 404 returned error can't find the container with id 623438fe73a4f207ac2907dda4b138a7c0a60c98f2e75f3f9fa56fee7d70f94b Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.737779 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-cn76s" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.753618 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-7rhps"] Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.754536 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.756836 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.757403 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.757749 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.758256 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.758556 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Oct 05 20:54:59 crc kubenswrapper[4754]: W1005 20:54:59.761786 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod02d5e3f9_73c2_4496_9aca_0787184aef19.slice/crio-9654324a0fffc12f38089e6e26ec327a082e5f6d76725dedf9f264d3e4ffbd3e WatchSource:0}: Error finding container 9654324a0fffc12f38089e6e26ec327a082e5f6d76725dedf9f264d3e4ffbd3e: Status 404 returned error can't find the container with id 9654324a0fffc12f38089e6e26ec327a082e5f6d76725dedf9f264d3e4ffbd3e Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.762611 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.762695 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.771064 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.804754 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.819000 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:54:59Z is after 2025-08-24T17:21:41Z" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.846303 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b44790fb-fecb-4ec3-9816-8e23dd72ab33\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7rhps\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:54:59Z is after 2025-08-24T17:21:41Z" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.857337 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b44790fb-fecb-4ec3-9816-8e23dd72ab33-etc-openvswitch\") pod \"ovnkube-node-7rhps\" (UID: \"b44790fb-fecb-4ec3-9816-8e23dd72ab33\") " pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.857375 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b44790fb-fecb-4ec3-9816-8e23dd72ab33-host-run-ovn-kubernetes\") pod \"ovnkube-node-7rhps\" (UID: \"b44790fb-fecb-4ec3-9816-8e23dd72ab33\") " pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.857400 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/b44790fb-fecb-4ec3-9816-8e23dd72ab33-run-ovn\") pod \"ovnkube-node-7rhps\" (UID: \"b44790fb-fecb-4ec3-9816-8e23dd72ab33\") " pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.857418 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b44790fb-fecb-4ec3-9816-8e23dd72ab33-var-lib-openvswitch\") pod \"ovnkube-node-7rhps\" (UID: \"b44790fb-fecb-4ec3-9816-8e23dd72ab33\") " pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.857439 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/b44790fb-fecb-4ec3-9816-8e23dd72ab33-ovnkube-script-lib\") pod \"ovnkube-node-7rhps\" (UID: \"b44790fb-fecb-4ec3-9816-8e23dd72ab33\") " pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.857454 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gnzsc\" (UniqueName: \"kubernetes.io/projected/b44790fb-fecb-4ec3-9816-8e23dd72ab33-kube-api-access-gnzsc\") pod \"ovnkube-node-7rhps\" (UID: \"b44790fb-fecb-4ec3-9816-8e23dd72ab33\") " pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.857468 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b44790fb-fecb-4ec3-9816-8e23dd72ab33-run-openvswitch\") pod \"ovnkube-node-7rhps\" (UID: \"b44790fb-fecb-4ec3-9816-8e23dd72ab33\") " pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.857528 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/b44790fb-fecb-4ec3-9816-8e23dd72ab33-node-log\") pod \"ovnkube-node-7rhps\" (UID: \"b44790fb-fecb-4ec3-9816-8e23dd72ab33\") " pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.857545 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/b44790fb-fecb-4ec3-9816-8e23dd72ab33-log-socket\") pod \"ovnkube-node-7rhps\" (UID: \"b44790fb-fecb-4ec3-9816-8e23dd72ab33\") " pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.857562 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/b44790fb-fecb-4ec3-9816-8e23dd72ab33-env-overrides\") pod \"ovnkube-node-7rhps\" (UID: \"b44790fb-fecb-4ec3-9816-8e23dd72ab33\") " pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.857589 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/b44790fb-fecb-4ec3-9816-8e23dd72ab33-run-systemd\") pod \"ovnkube-node-7rhps\" (UID: \"b44790fb-fecb-4ec3-9816-8e23dd72ab33\") " pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.857604 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/b44790fb-fecb-4ec3-9816-8e23dd72ab33-host-cni-bin\") pod \"ovnkube-node-7rhps\" (UID: \"b44790fb-fecb-4ec3-9816-8e23dd72ab33\") " pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.857617 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/b44790fb-fecb-4ec3-9816-8e23dd72ab33-ovnkube-config\") pod \"ovnkube-node-7rhps\" (UID: \"b44790fb-fecb-4ec3-9816-8e23dd72ab33\") " pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.857631 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/b44790fb-fecb-4ec3-9816-8e23dd72ab33-ovn-node-metrics-cert\") pod \"ovnkube-node-7rhps\" (UID: \"b44790fb-fecb-4ec3-9816-8e23dd72ab33\") " pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.857653 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/b44790fb-fecb-4ec3-9816-8e23dd72ab33-systemd-units\") pod \"ovnkube-node-7rhps\" (UID: \"b44790fb-fecb-4ec3-9816-8e23dd72ab33\") " pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.857674 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b44790fb-fecb-4ec3-9816-8e23dd72ab33-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-7rhps\" (UID: \"b44790fb-fecb-4ec3-9816-8e23dd72ab33\") " pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.857698 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/b44790fb-fecb-4ec3-9816-8e23dd72ab33-host-kubelet\") pod \"ovnkube-node-7rhps\" (UID: \"b44790fb-fecb-4ec3-9816-8e23dd72ab33\") " pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.857712 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/b44790fb-fecb-4ec3-9816-8e23dd72ab33-host-run-netns\") pod \"ovnkube-node-7rhps\" (UID: \"b44790fb-fecb-4ec3-9816-8e23dd72ab33\") " pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.857727 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/b44790fb-fecb-4ec3-9816-8e23dd72ab33-host-cni-netd\") pod \"ovnkube-node-7rhps\" (UID: \"b44790fb-fecb-4ec3-9816-8e23dd72ab33\") " pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.857741 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/b44790fb-fecb-4ec3-9816-8e23dd72ab33-host-slash\") pod \"ovnkube-node-7rhps\" (UID: \"b44790fb-fecb-4ec3-9816-8e23dd72ab33\") " pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.865934 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:54:59Z is after 2025-08-24T17:21:41Z" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.880968 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:54:59Z is after 2025-08-24T17:21:41Z" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.900278 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-57sr4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e854c7b9-1229-4da4-80e9-fea6479459b8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vw2ps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:58Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-57sr4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:54:59Z is after 2025-08-24T17:21:41Z" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.912402 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cn76s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02d5e3f9-73c2-4496-9aca-0787184aef19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpx5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cn76s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:54:59Z is after 2025-08-24T17:21:41Z" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.930048 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7506d496-65ff-41dc-8230-24c8ee653a2a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33da5a6bb3cd3b042c193bc571e6622f2f87f27e560cdd0bbb9be31e2010d707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df6d700a947d19c84b4f65fd2dd2f8d1f4f1f80fa9b4800c2acfa8dae3912639\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35ca8ee77a0e58cf969bd67b277a3392614c4e73b963e6f6002097709370715d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8907b9b69b249c34ec6db1e460d641b31711b75d6d31d4caafd867339e6571d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9d9b0b384ffcc4040a3cc8aea4be89ded347240018f198471da1922681c8efec\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T20:54:40Z\\\",\\\"message\\\":\\\"W1005 20:54:40.114418 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1005 20:54:40.115271 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759697680 cert, and key in /tmp/serving-cert-3287406251/serving-signer.crt, /tmp/serving-cert-3287406251/serving-signer.key\\\\nI1005 20:54:40.625469 1 observer_polling.go:159] Starting file observer\\\\nW1005 20:54:40.628724 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1005 20:54:40.628993 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 20:54:40.631041 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3287406251/tls.crt::/tmp/serving-cert-3287406251/tls.key\\\\\\\"\\\\nF1005 20:54:40.864813 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:40Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8907b9b69b249c34ec6db1e460d641b31711b75d6d31d4caafd867339e6571d0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1005 20:54:52.589115 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 20:54:52.590181 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1304624737/tls.crt::/tmp/serving-cert-1304624737/tls.key\\\\\\\"\\\\nI1005 20:54:58.070544 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1005 20:54:58.090352 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1005 20:54:58.090382 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1005 20:54:58.090412 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1005 20:54:58.090418 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1005 20:54:58.144656 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1005 20:54:58.144688 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 20:54:58.144695 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 20:54:58.144699 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1005 20:54:58.144702 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1005 20:54:58.144705 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1005 20:54:58.144708 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1005 20:54:58.145024 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1005 20:54:58.148112 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4efb34f3493cc8854bc4cff6b29f63973c52b754404d7a21daec4e1f612d327d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54c568b637f3737667433047b956c7eb22fdf0f9301acdfd3b49fb84c39de6ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54c568b637f3737667433047b956c7eb22fdf0f9301acdfd3b49fb84c39de6ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:54:59Z is after 2025-08-24T17:21:41Z" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.944564 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.946730 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:54:59Z is after 2025-08-24T17:21:41Z" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.948831 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.952316 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.958313 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/b44790fb-fecb-4ec3-9816-8e23dd72ab33-host-kubelet\") pod \"ovnkube-node-7rhps\" (UID: \"b44790fb-fecb-4ec3-9816-8e23dd72ab33\") " pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.959142 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/b44790fb-fecb-4ec3-9816-8e23dd72ab33-host-run-netns\") pod \"ovnkube-node-7rhps\" (UID: \"b44790fb-fecb-4ec3-9816-8e23dd72ab33\") " pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.958675 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/b44790fb-fecb-4ec3-9816-8e23dd72ab33-host-kubelet\") pod \"ovnkube-node-7rhps\" (UID: \"b44790fb-fecb-4ec3-9816-8e23dd72ab33\") " pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.959173 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/b44790fb-fecb-4ec3-9816-8e23dd72ab33-host-cni-netd\") pod \"ovnkube-node-7rhps\" (UID: \"b44790fb-fecb-4ec3-9816-8e23dd72ab33\") " pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.959195 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/b44790fb-fecb-4ec3-9816-8e23dd72ab33-host-slash\") pod \"ovnkube-node-7rhps\" (UID: \"b44790fb-fecb-4ec3-9816-8e23dd72ab33\") " pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.959218 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b44790fb-fecb-4ec3-9816-8e23dd72ab33-host-run-ovn-kubernetes\") pod \"ovnkube-node-7rhps\" (UID: \"b44790fb-fecb-4ec3-9816-8e23dd72ab33\") " pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.959236 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b44790fb-fecb-4ec3-9816-8e23dd72ab33-etc-openvswitch\") pod \"ovnkube-node-7rhps\" (UID: \"b44790fb-fecb-4ec3-9816-8e23dd72ab33\") " pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.959238 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/b44790fb-fecb-4ec3-9816-8e23dd72ab33-host-run-netns\") pod \"ovnkube-node-7rhps\" (UID: \"b44790fb-fecb-4ec3-9816-8e23dd72ab33\") " pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.959251 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/b44790fb-fecb-4ec3-9816-8e23dd72ab33-run-ovn\") pod \"ovnkube-node-7rhps\" (UID: \"b44790fb-fecb-4ec3-9816-8e23dd72ab33\") " pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.959270 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b44790fb-fecb-4ec3-9816-8e23dd72ab33-var-lib-openvswitch\") pod \"ovnkube-node-7rhps\" (UID: \"b44790fb-fecb-4ec3-9816-8e23dd72ab33\") " pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.959277 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/b44790fb-fecb-4ec3-9816-8e23dd72ab33-host-slash\") pod \"ovnkube-node-7rhps\" (UID: \"b44790fb-fecb-4ec3-9816-8e23dd72ab33\") " pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.959291 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/b44790fb-fecb-4ec3-9816-8e23dd72ab33-ovnkube-script-lib\") pod \"ovnkube-node-7rhps\" (UID: \"b44790fb-fecb-4ec3-9816-8e23dd72ab33\") " pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.959311 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gnzsc\" (UniqueName: \"kubernetes.io/projected/b44790fb-fecb-4ec3-9816-8e23dd72ab33-kube-api-access-gnzsc\") pod \"ovnkube-node-7rhps\" (UID: \"b44790fb-fecb-4ec3-9816-8e23dd72ab33\") " pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.959345 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/b44790fb-fecb-4ec3-9816-8e23dd72ab33-run-ovn\") pod \"ovnkube-node-7rhps\" (UID: \"b44790fb-fecb-4ec3-9816-8e23dd72ab33\") " pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.959347 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b44790fb-fecb-4ec3-9816-8e23dd72ab33-run-openvswitch\") pod \"ovnkube-node-7rhps\" (UID: \"b44790fb-fecb-4ec3-9816-8e23dd72ab33\") " pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.959389 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b44790fb-fecb-4ec3-9816-8e23dd72ab33-run-openvswitch\") pod \"ovnkube-node-7rhps\" (UID: \"b44790fb-fecb-4ec3-9816-8e23dd72ab33\") " pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.959409 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/b44790fb-fecb-4ec3-9816-8e23dd72ab33-node-log\") pod \"ovnkube-node-7rhps\" (UID: \"b44790fb-fecb-4ec3-9816-8e23dd72ab33\") " pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.959429 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b44790fb-fecb-4ec3-9816-8e23dd72ab33-var-lib-openvswitch\") pod \"ovnkube-node-7rhps\" (UID: \"b44790fb-fecb-4ec3-9816-8e23dd72ab33\") " pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.959441 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/b44790fb-fecb-4ec3-9816-8e23dd72ab33-log-socket\") pod \"ovnkube-node-7rhps\" (UID: \"b44790fb-fecb-4ec3-9816-8e23dd72ab33\") " pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.959472 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/b44790fb-fecb-4ec3-9816-8e23dd72ab33-env-overrides\") pod \"ovnkube-node-7rhps\" (UID: \"b44790fb-fecb-4ec3-9816-8e23dd72ab33\") " pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.959511 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/b44790fb-fecb-4ec3-9816-8e23dd72ab33-run-systemd\") pod \"ovnkube-node-7rhps\" (UID: \"b44790fb-fecb-4ec3-9816-8e23dd72ab33\") " pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.959532 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/b44790fb-fecb-4ec3-9816-8e23dd72ab33-host-cni-bin\") pod \"ovnkube-node-7rhps\" (UID: \"b44790fb-fecb-4ec3-9816-8e23dd72ab33\") " pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.959555 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/b44790fb-fecb-4ec3-9816-8e23dd72ab33-ovnkube-config\") pod \"ovnkube-node-7rhps\" (UID: \"b44790fb-fecb-4ec3-9816-8e23dd72ab33\") " pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.959617 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/b44790fb-fecb-4ec3-9816-8e23dd72ab33-systemd-units\") pod \"ovnkube-node-7rhps\" (UID: \"b44790fb-fecb-4ec3-9816-8e23dd72ab33\") " pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.959639 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b44790fb-fecb-4ec3-9816-8e23dd72ab33-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-7rhps\" (UID: \"b44790fb-fecb-4ec3-9816-8e23dd72ab33\") " pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.959662 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/b44790fb-fecb-4ec3-9816-8e23dd72ab33-ovn-node-metrics-cert\") pod \"ovnkube-node-7rhps\" (UID: \"b44790fb-fecb-4ec3-9816-8e23dd72ab33\") " pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.961335 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/b44790fb-fecb-4ec3-9816-8e23dd72ab33-ovnkube-script-lib\") pod \"ovnkube-node-7rhps\" (UID: \"b44790fb-fecb-4ec3-9816-8e23dd72ab33\") " pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.961365 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/b44790fb-fecb-4ec3-9816-8e23dd72ab33-env-overrides\") pod \"ovnkube-node-7rhps\" (UID: \"b44790fb-fecb-4ec3-9816-8e23dd72ab33\") " pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.961417 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b44790fb-fecb-4ec3-9816-8e23dd72ab33-etc-openvswitch\") pod \"ovnkube-node-7rhps\" (UID: \"b44790fb-fecb-4ec3-9816-8e23dd72ab33\") " pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.961445 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/b44790fb-fecb-4ec3-9816-8e23dd72ab33-node-log\") pod \"ovnkube-node-7rhps\" (UID: \"b44790fb-fecb-4ec3-9816-8e23dd72ab33\") " pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.961440 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b44790fb-fecb-4ec3-9816-8e23dd72ab33-host-run-ovn-kubernetes\") pod \"ovnkube-node-7rhps\" (UID: \"b44790fb-fecb-4ec3-9816-8e23dd72ab33\") " pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.961471 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/b44790fb-fecb-4ec3-9816-8e23dd72ab33-log-socket\") pod \"ovnkube-node-7rhps\" (UID: \"b44790fb-fecb-4ec3-9816-8e23dd72ab33\") " pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.959315 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/b44790fb-fecb-4ec3-9816-8e23dd72ab33-host-cni-netd\") pod \"ovnkube-node-7rhps\" (UID: \"b44790fb-fecb-4ec3-9816-8e23dd72ab33\") " pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.961538 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/b44790fb-fecb-4ec3-9816-8e23dd72ab33-systemd-units\") pod \"ovnkube-node-7rhps\" (UID: \"b44790fb-fecb-4ec3-9816-8e23dd72ab33\") " pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.961563 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/b44790fb-fecb-4ec3-9816-8e23dd72ab33-run-systemd\") pod \"ovnkube-node-7rhps\" (UID: \"b44790fb-fecb-4ec3-9816-8e23dd72ab33\") " pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.961595 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/b44790fb-fecb-4ec3-9816-8e23dd72ab33-host-cni-bin\") pod \"ovnkube-node-7rhps\" (UID: \"b44790fb-fecb-4ec3-9816-8e23dd72ab33\") " pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.961621 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b44790fb-fecb-4ec3-9816-8e23dd72ab33-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-7rhps\" (UID: \"b44790fb-fecb-4ec3-9816-8e23dd72ab33\") " pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.962118 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/b44790fb-fecb-4ec3-9816-8e23dd72ab33-ovnkube-config\") pod \"ovnkube-node-7rhps\" (UID: \"b44790fb-fecb-4ec3-9816-8e23dd72ab33\") " pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.964028 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/b44790fb-fecb-4ec3-9816-8e23dd72ab33-ovn-node-metrics-cert\") pod \"ovnkube-node-7rhps\" (UID: \"b44790fb-fecb-4ec3-9816-8e23dd72ab33\") " pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.968781 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tskt7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a68e9c89-465b-4c01-a470-e21a480ee465\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tskt7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:54:59Z is after 2025-08-24T17:21:41Z" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.981425 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be95a413-6a59-45b4-84b7-b43cae694a26\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2ggg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2ggg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-b2h9k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:54:59Z is after 2025-08-24T17:21:41Z" Oct 05 20:54:59 crc kubenswrapper[4754]: I1005 20:54:59.988228 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gnzsc\" (UniqueName: \"kubernetes.io/projected/b44790fb-fecb-4ec3-9816-8e23dd72ab33-kube-api-access-gnzsc\") pod \"ovnkube-node-7rhps\" (UID: \"b44790fb-fecb-4ec3-9816-8e23dd72ab33\") " pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" Oct 05 20:55:00 crc kubenswrapper[4754]: I1005 20:55:00.002045 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7506d496-65ff-41dc-8230-24c8ee653a2a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33da5a6bb3cd3b042c193bc571e6622f2f87f27e560cdd0bbb9be31e2010d707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df6d700a947d19c84b4f65fd2dd2f8d1f4f1f80fa9b4800c2acfa8dae3912639\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35ca8ee77a0e58cf969bd67b277a3392614c4e73b963e6f6002097709370715d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8907b9b69b249c34ec6db1e460d641b31711b75d6d31d4caafd867339e6571d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9d9b0b384ffcc4040a3cc8aea4be89ded347240018f198471da1922681c8efec\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T20:54:40Z\\\",\\\"message\\\":\\\"W1005 20:54:40.114418 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1005 20:54:40.115271 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759697680 cert, and key in /tmp/serving-cert-3287406251/serving-signer.crt, /tmp/serving-cert-3287406251/serving-signer.key\\\\nI1005 20:54:40.625469 1 observer_polling.go:159] Starting file observer\\\\nW1005 20:54:40.628724 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1005 20:54:40.628993 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 20:54:40.631041 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3287406251/tls.crt::/tmp/serving-cert-3287406251/tls.key\\\\\\\"\\\\nF1005 20:54:40.864813 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:40Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8907b9b69b249c34ec6db1e460d641b31711b75d6d31d4caafd867339e6571d0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1005 20:54:52.589115 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 20:54:52.590181 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1304624737/tls.crt::/tmp/serving-cert-1304624737/tls.key\\\\\\\"\\\\nI1005 20:54:58.070544 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1005 20:54:58.090352 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1005 20:54:58.090382 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1005 20:54:58.090412 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1005 20:54:58.090418 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1005 20:54:58.144656 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1005 20:54:58.144688 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 20:54:58.144695 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 20:54:58.144699 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1005 20:54:58.144702 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1005 20:54:58.144705 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1005 20:54:58.144708 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1005 20:54:58.145024 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1005 20:54:58.148112 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4efb34f3493cc8854bc4cff6b29f63973c52b754404d7a21daec4e1f612d327d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54c568b637f3737667433047b956c7eb22fdf0f9301acdfd3b49fb84c39de6ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54c568b637f3737667433047b956c7eb22fdf0f9301acdfd3b49fb84c39de6ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:54:59Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:00 crc kubenswrapper[4754]: I1005 20:55:00.023939 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:00Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:00 crc kubenswrapper[4754]: I1005 20:55:00.031397 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-tskt7" event={"ID":"a68e9c89-465b-4c01-a470-e21a480ee465","Type":"ContainerStarted","Data":"5d14ec99610d2167b484bd7712af7bb6d993d26c74b2828da1c89314f16fda22"} Oct 05 20:55:00 crc kubenswrapper[4754]: I1005 20:55:00.031458 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-tskt7" event={"ID":"a68e9c89-465b-4c01-a470-e21a480ee465","Type":"ContainerStarted","Data":"6b6b7f27c10d9b52260897b10d5fd0c3584eeea9e77f6765470a6dd8c813fe81"} Oct 05 20:55:00 crc kubenswrapper[4754]: I1005 20:55:00.033868 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-57sr4" event={"ID":"e854c7b9-1229-4da4-80e9-fea6479459b8","Type":"ContainerStarted","Data":"ceb577e1055d84007936a50d0b3d54b9ae0299a1fc1a6618b37163a6d3fbb9cf"} Oct 05 20:55:00 crc kubenswrapper[4754]: I1005 20:55:00.033900 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-57sr4" event={"ID":"e854c7b9-1229-4da4-80e9-fea6479459b8","Type":"ContainerStarted","Data":"26d9e615d742750e95d13d37f7dbc3da5b7bfc5a005c4466ffbce353232924ca"} Oct 05 20:55:00 crc kubenswrapper[4754]: I1005 20:55:00.036732 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"c76e6e4740035a6ca54191a770c4516e28101aae69afbb89739174c2f076ef5e"} Oct 05 20:55:00 crc kubenswrapper[4754]: I1005 20:55:00.036770 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"f9eededc0cb1af94fb7b5daba08785f6b077a85e9cb5ddd8b9f397b658194f61"} Oct 05 20:55:00 crc kubenswrapper[4754]: I1005 20:55:00.036786 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"236c0608440f80e146deef69071e1f27712870db81ce7125536a9744bb1886da"} Oct 05 20:55:00 crc kubenswrapper[4754]: I1005 20:55:00.038801 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"e380f9b2a0e319dc325c54d1247a6db98346292548fa688cb41a2ca1b3048419"} Oct 05 20:55:00 crc kubenswrapper[4754]: I1005 20:55:00.038836 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"0aace4887b483170c69617bdf8f33ec669e2159bd7e9751ee5db71b416ab346c"} Oct 05 20:55:00 crc kubenswrapper[4754]: I1005 20:55:00.046209 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-cn76s" event={"ID":"02d5e3f9-73c2-4496-9aca-0787184aef19","Type":"ContainerStarted","Data":"f7ba2b076839f4a5f586d2e6c4f4d7188254128b782bfd2180e22bf3ea5ffdd7"} Oct 05 20:55:00 crc kubenswrapper[4754]: I1005 20:55:00.046246 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-cn76s" event={"ID":"02d5e3f9-73c2-4496-9aca-0787184aef19","Type":"ContainerStarted","Data":"9654324a0fffc12f38089e6e26ec327a082e5f6d76725dedf9f264d3e4ffbd3e"} Oct 05 20:55:00 crc kubenswrapper[4754]: I1005 20:55:00.049053 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" event={"ID":"be95a413-6a59-45b4-84b7-b43cae694a26","Type":"ContainerStarted","Data":"10b211cf6e47dbd1ee42c82e49de0a13a45581b92a0c4f6bbba0326c7149e41d"} Oct 05 20:55:00 crc kubenswrapper[4754]: I1005 20:55:00.049181 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" event={"ID":"be95a413-6a59-45b4-84b7-b43cae694a26","Type":"ContainerStarted","Data":"4bb8c5903e8b70dd19eebb111097d96f725e17de44e06c3e74c067376f6c7360"} Oct 05 20:55:00 crc kubenswrapper[4754]: I1005 20:55:00.049283 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" event={"ID":"be95a413-6a59-45b4-84b7-b43cae694a26","Type":"ContainerStarted","Data":"623438fe73a4f207ac2907dda4b138a7c0a60c98f2e75f3f9fa56fee7d70f94b"} Oct 05 20:55:00 crc kubenswrapper[4754]: I1005 20:55:00.050453 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"c4e51bd877a767cbcffdfd35e9de2de7690c934b7216a8578e9d1916fc03fb51"} Oct 05 20:55:00 crc kubenswrapper[4754]: I1005 20:55:00.052239 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Oct 05 20:55:00 crc kubenswrapper[4754]: I1005 20:55:00.055237 4754 scope.go:117] "RemoveContainer" containerID="8907b9b69b249c34ec6db1e460d641b31711b75d6d31d4caafd867339e6571d0" Oct 05 20:55:00 crc kubenswrapper[4754]: E1005 20:55:00.055401 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Oct 05 20:55:00 crc kubenswrapper[4754]: I1005 20:55:00.062773 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tskt7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a68e9c89-465b-4c01-a470-e21a480ee465\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tskt7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:00Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:00 crc kubenswrapper[4754]: I1005 20:55:00.072409 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" Oct 05 20:55:00 crc kubenswrapper[4754]: I1005 20:55:00.076213 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be95a413-6a59-45b4-84b7-b43cae694a26\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2ggg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2ggg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-b2h9k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:00Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:00 crc kubenswrapper[4754]: W1005 20:55:00.092583 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb44790fb_fecb_4ec3_9816_8e23dd72ab33.slice/crio-fd436f3a6ea66b669ec762475a28d06906ed8dbaaff8e697b5cf00cf82b92886 WatchSource:0}: Error finding container fd436f3a6ea66b669ec762475a28d06906ed8dbaaff8e697b5cf00cf82b92886: Status 404 returned error can't find the container with id fd436f3a6ea66b669ec762475a28d06906ed8dbaaff8e697b5cf00cf82b92886 Oct 05 20:55:00 crc kubenswrapper[4754]: I1005 20:55:00.098447 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:00Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:00 crc kubenswrapper[4754]: I1005 20:55:00.114672 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:00Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:00 crc kubenswrapper[4754]: I1005 20:55:00.128289 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:00Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:00 crc kubenswrapper[4754]: I1005 20:55:00.143462 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b44790fb-fecb-4ec3-9816-8e23dd72ab33\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7rhps\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:00Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:00 crc kubenswrapper[4754]: I1005 20:55:00.155792 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cn76s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02d5e3f9-73c2-4496-9aca-0787184aef19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpx5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cn76s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:00Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:00 crc kubenswrapper[4754]: I1005 20:55:00.173050 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"934bcd66-9c03-48b9-82dd-5093338f9c8f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9042df2f43243d3d79ba9a3940f8dd4f6c7a89ef9e3510e81fe60b1944d7e83b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5366bbf46a91667355f43f5df7335a536b395cda788b90b35ef18e2b53bc4c02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://600bc3bb551b747cf2f18e26745883232d7a37ac5877fad0c67100e52d7a43aa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78f74982180048eb50b9be79fa72c04f62f925b54fe511920f37e6407f864631\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:00Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:00 crc kubenswrapper[4754]: I1005 20:55:00.198660 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:00Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:00 crc kubenswrapper[4754]: I1005 20:55:00.212925 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:00Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:00 crc kubenswrapper[4754]: I1005 20:55:00.224056 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-57sr4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e854c7b9-1229-4da4-80e9-fea6479459b8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vw2ps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:58Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-57sr4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:00Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:00 crc kubenswrapper[4754]: I1005 20:55:00.242749 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7506d496-65ff-41dc-8230-24c8ee653a2a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33da5a6bb3cd3b042c193bc571e6622f2f87f27e560cdd0bbb9be31e2010d707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df6d700a947d19c84b4f65fd2dd2f8d1f4f1f80fa9b4800c2acfa8dae3912639\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35ca8ee77a0e58cf969bd67b277a3392614c4e73b963e6f6002097709370715d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8907b9b69b249c34ec6db1e460d641b31711b75d6d31d4caafd867339e6571d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8907b9b69b249c34ec6db1e460d641b31711b75d6d31d4caafd867339e6571d0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1005 20:54:52.589115 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 20:54:52.590181 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1304624737/tls.crt::/tmp/serving-cert-1304624737/tls.key\\\\\\\"\\\\nI1005 20:54:58.070544 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1005 20:54:58.090352 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1005 20:54:58.090382 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1005 20:54:58.090412 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1005 20:54:58.090418 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1005 20:54:58.144656 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1005 20:54:58.144688 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 20:54:58.144695 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 20:54:58.144699 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1005 20:54:58.144702 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1005 20:54:58.144705 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1005 20:54:58.144708 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1005 20:54:58.145024 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1005 20:54:58.148112 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4efb34f3493cc8854bc4cff6b29f63973c52b754404d7a21daec4e1f612d327d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54c568b637f3737667433047b956c7eb22fdf0f9301acdfd3b49fb84c39de6ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54c568b637f3737667433047b956c7eb22fdf0f9301acdfd3b49fb84c39de6ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:00Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:00 crc kubenswrapper[4754]: I1005 20:55:00.256651 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c76e6e4740035a6ca54191a770c4516e28101aae69afbb89739174c2f076ef5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9eededc0cb1af94fb7b5daba08785f6b077a85e9cb5ddd8b9f397b658194f61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:00Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:00 crc kubenswrapper[4754]: I1005 20:55:00.272587 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tskt7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a68e9c89-465b-4c01-a470-e21a480ee465\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d14ec99610d2167b484bd7712af7bb6d993d26c74b2828da1c89314f16fda22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tskt7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:00Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:00 crc kubenswrapper[4754]: I1005 20:55:00.287105 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be95a413-6a59-45b4-84b7-b43cae694a26\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10b211cf6e47dbd1ee42c82e49de0a13a45581b92a0c4f6bbba0326c7149e41d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2ggg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4bb8c5903e8b70dd19eebb111097d96f725e17de44e06c3e74c067376f6c7360\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2ggg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-b2h9k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:00Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:00 crc kubenswrapper[4754]: I1005 20:55:00.300192 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e380f9b2a0e319dc325c54d1247a6db98346292548fa688cb41a2ca1b3048419\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:00Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:00 crc kubenswrapper[4754]: I1005 20:55:00.310882 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:00Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:00 crc kubenswrapper[4754]: I1005 20:55:00.324057 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:00Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:00 crc kubenswrapper[4754]: I1005 20:55:00.356417 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b44790fb-fecb-4ec3-9816-8e23dd72ab33\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7rhps\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:00Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:00 crc kubenswrapper[4754]: I1005 20:55:00.369866 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 20:55:00 crc kubenswrapper[4754]: E1005 20:55:00.370146 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 20:55:02.370101976 +0000 UTC m=+26.274220696 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:55:00 crc kubenswrapper[4754]: I1005 20:55:00.397453 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cn76s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02d5e3f9-73c2-4496-9aca-0787184aef19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7ba2b076839f4a5f586d2e6c4f4d7188254128b782bfd2180e22bf3ea5ffdd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpx5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cn76s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:00Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:00 crc kubenswrapper[4754]: I1005 20:55:00.433525 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"934bcd66-9c03-48b9-82dd-5093338f9c8f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9042df2f43243d3d79ba9a3940f8dd4f6c7a89ef9e3510e81fe60b1944d7e83b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5366bbf46a91667355f43f5df7335a536b395cda788b90b35ef18e2b53bc4c02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://600bc3bb551b747cf2f18e26745883232d7a37ac5877fad0c67100e52d7a43aa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78f74982180048eb50b9be79fa72c04f62f925b54fe511920f37e6407f864631\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:00Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:00 crc kubenswrapper[4754]: I1005 20:55:00.471179 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 20:55:00 crc kubenswrapper[4754]: I1005 20:55:00.471232 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 20:55:00 crc kubenswrapper[4754]: I1005 20:55:00.471261 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 20:55:00 crc kubenswrapper[4754]: I1005 20:55:00.471288 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 20:55:00 crc kubenswrapper[4754]: E1005 20:55:00.471358 4754 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 05 20:55:00 crc kubenswrapper[4754]: E1005 20:55:00.471439 4754 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 05 20:55:00 crc kubenswrapper[4754]: E1005 20:55:00.471449 4754 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 05 20:55:00 crc kubenswrapper[4754]: E1005 20:55:00.471472 4754 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 05 20:55:00 crc kubenswrapper[4754]: E1005 20:55:00.471486 4754 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 05 20:55:00 crc kubenswrapper[4754]: E1005 20:55:00.471513 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-05 20:55:02.471478555 +0000 UTC m=+26.375597265 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 05 20:55:00 crc kubenswrapper[4754]: E1005 20:55:00.471531 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-05 20:55:02.471524886 +0000 UTC m=+26.375643596 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 05 20:55:00 crc kubenswrapper[4754]: E1005 20:55:00.471570 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-05 20:55:02.471550737 +0000 UTC m=+26.375669457 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 05 20:55:00 crc kubenswrapper[4754]: E1005 20:55:00.471599 4754 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 05 20:55:00 crc kubenswrapper[4754]: E1005 20:55:00.471651 4754 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 05 20:55:00 crc kubenswrapper[4754]: E1005 20:55:00.471665 4754 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 05 20:55:00 crc kubenswrapper[4754]: E1005 20:55:00.471746 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-05 20:55:02.471722622 +0000 UTC m=+26.375841332 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 05 20:55:00 crc kubenswrapper[4754]: I1005 20:55:00.483898 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:00Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:00 crc kubenswrapper[4754]: I1005 20:55:00.517057 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:00Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:00 crc kubenswrapper[4754]: I1005 20:55:00.518040 4754 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 05 20:55:00 crc kubenswrapper[4754]: I1005 20:55:00.556258 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-57sr4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e854c7b9-1229-4da4-80e9-fea6479459b8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceb577e1055d84007936a50d0b3d54b9ae0299a1fc1a6618b37163a6d3fbb9cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vw2ps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:58Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-57sr4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:00Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:00 crc kubenswrapper[4754]: I1005 20:55:00.742918 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Oct 05 20:55:00 crc kubenswrapper[4754]: I1005 20:55:00.753573 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Oct 05 20:55:00 crc kubenswrapper[4754]: I1005 20:55:00.758220 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e380f9b2a0e319dc325c54d1247a6db98346292548fa688cb41a2ca1b3048419\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:00Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:00 crc kubenswrapper[4754]: I1005 20:55:00.759993 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Oct 05 20:55:00 crc kubenswrapper[4754]: I1005 20:55:00.772571 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:00Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:00 crc kubenswrapper[4754]: I1005 20:55:00.787910 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:00Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:00 crc kubenswrapper[4754]: I1005 20:55:00.809367 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b44790fb-fecb-4ec3-9816-8e23dd72ab33\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7rhps\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:00Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:00 crc kubenswrapper[4754]: I1005 20:55:00.824420 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:00Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:00 crc kubenswrapper[4754]: I1005 20:55:00.837102 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 20:55:00 crc kubenswrapper[4754]: I1005 20:55:00.837120 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 20:55:00 crc kubenswrapper[4754]: I1005 20:55:00.837095 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 20:55:00 crc kubenswrapper[4754]: E1005 20:55:00.837248 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 20:55:00 crc kubenswrapper[4754]: E1005 20:55:00.837360 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 20:55:00 crc kubenswrapper[4754]: E1005 20:55:00.837442 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 20:55:00 crc kubenswrapper[4754]: I1005 20:55:00.838271 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:00Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:00 crc kubenswrapper[4754]: I1005 20:55:00.841871 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Oct 05 20:55:00 crc kubenswrapper[4754]: I1005 20:55:00.842373 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Oct 05 20:55:00 crc kubenswrapper[4754]: I1005 20:55:00.843640 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Oct 05 20:55:00 crc kubenswrapper[4754]: I1005 20:55:00.844251 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Oct 05 20:55:00 crc kubenswrapper[4754]: I1005 20:55:00.845258 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Oct 05 20:55:00 crc kubenswrapper[4754]: I1005 20:55:00.845800 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Oct 05 20:55:00 crc kubenswrapper[4754]: I1005 20:55:00.846371 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Oct 05 20:55:00 crc kubenswrapper[4754]: I1005 20:55:00.847334 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Oct 05 20:55:00 crc kubenswrapper[4754]: I1005 20:55:00.847994 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Oct 05 20:55:00 crc kubenswrapper[4754]: I1005 20:55:00.848899 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Oct 05 20:55:00 crc kubenswrapper[4754]: I1005 20:55:00.849407 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Oct 05 20:55:00 crc kubenswrapper[4754]: I1005 20:55:00.850434 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Oct 05 20:55:00 crc kubenswrapper[4754]: I1005 20:55:00.850970 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Oct 05 20:55:00 crc kubenswrapper[4754]: I1005 20:55:00.851453 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Oct 05 20:55:00 crc kubenswrapper[4754]: I1005 20:55:00.852324 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Oct 05 20:55:00 crc kubenswrapper[4754]: I1005 20:55:00.852873 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Oct 05 20:55:00 crc kubenswrapper[4754]: I1005 20:55:00.853428 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-57sr4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e854c7b9-1229-4da4-80e9-fea6479459b8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceb577e1055d84007936a50d0b3d54b9ae0299a1fc1a6618b37163a6d3fbb9cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vw2ps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:58Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-57sr4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:00Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:00 crc kubenswrapper[4754]: I1005 20:55:00.854095 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Oct 05 20:55:00 crc kubenswrapper[4754]: I1005 20:55:00.854661 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Oct 05 20:55:00 crc kubenswrapper[4754]: I1005 20:55:00.855388 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Oct 05 20:55:00 crc kubenswrapper[4754]: I1005 20:55:00.858905 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Oct 05 20:55:00 crc kubenswrapper[4754]: I1005 20:55:00.859439 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Oct 05 20:55:00 crc kubenswrapper[4754]: I1005 20:55:00.860044 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Oct 05 20:55:00 crc kubenswrapper[4754]: I1005 20:55:00.860892 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Oct 05 20:55:00 crc kubenswrapper[4754]: I1005 20:55:00.861591 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Oct 05 20:55:00 crc kubenswrapper[4754]: I1005 20:55:00.862974 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Oct 05 20:55:00 crc kubenswrapper[4754]: I1005 20:55:00.893663 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cn76s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02d5e3f9-73c2-4496-9aca-0787184aef19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7ba2b076839f4a5f586d2e6c4f4d7188254128b782bfd2180e22bf3ea5ffdd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpx5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cn76s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:00Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:00 crc kubenswrapper[4754]: I1005 20:55:00.931474 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"934bcd66-9c03-48b9-82dd-5093338f9c8f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9042df2f43243d3d79ba9a3940f8dd4f6c7a89ef9e3510e81fe60b1944d7e83b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5366bbf46a91667355f43f5df7335a536b395cda788b90b35ef18e2b53bc4c02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://600bc3bb551b747cf2f18e26745883232d7a37ac5877fad0c67100e52d7a43aa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78f74982180048eb50b9be79fa72c04f62f925b54fe511920f37e6407f864631\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:00Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:01 crc kubenswrapper[4754]: I1005 20:55:01.009887 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c76e6e4740035a6ca54191a770c4516e28101aae69afbb89739174c2f076ef5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9eededc0cb1af94fb7b5daba08785f6b077a85e9cb5ddd8b9f397b658194f61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:01Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:01 crc kubenswrapper[4754]: I1005 20:55:01.030372 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tskt7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a68e9c89-465b-4c01-a470-e21a480ee465\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d14ec99610d2167b484bd7712af7bb6d993d26c74b2828da1c89314f16fda22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tskt7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:01Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:01 crc kubenswrapper[4754]: I1005 20:55:01.053453 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be95a413-6a59-45b4-84b7-b43cae694a26\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10b211cf6e47dbd1ee42c82e49de0a13a45581b92a0c4f6bbba0326c7149e41d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2ggg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4bb8c5903e8b70dd19eebb111097d96f725e17de44e06c3e74c067376f6c7360\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2ggg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-b2h9k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:01Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:01 crc kubenswrapper[4754]: I1005 20:55:01.058765 4754 generic.go:334] "Generic (PLEG): container finished" podID="b44790fb-fecb-4ec3-9816-8e23dd72ab33" containerID="35ebeb7a87d5d47b6567917d38eca38a3e622c497fba58971296c4609c287423" exitCode=0 Oct 05 20:55:01 crc kubenswrapper[4754]: I1005 20:55:01.058871 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" event={"ID":"b44790fb-fecb-4ec3-9816-8e23dd72ab33","Type":"ContainerDied","Data":"35ebeb7a87d5d47b6567917d38eca38a3e622c497fba58971296c4609c287423"} Oct 05 20:55:01 crc kubenswrapper[4754]: I1005 20:55:01.058939 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" event={"ID":"b44790fb-fecb-4ec3-9816-8e23dd72ab33","Type":"ContainerStarted","Data":"fd436f3a6ea66b669ec762475a28d06906ed8dbaaff8e697b5cf00cf82b92886"} Oct 05 20:55:01 crc kubenswrapper[4754]: I1005 20:55:01.060573 4754 generic.go:334] "Generic (PLEG): container finished" podID="a68e9c89-465b-4c01-a470-e21a480ee465" containerID="5d14ec99610d2167b484bd7712af7bb6d993d26c74b2828da1c89314f16fda22" exitCode=0 Oct 05 20:55:01 crc kubenswrapper[4754]: I1005 20:55:01.060751 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-tskt7" event={"ID":"a68e9c89-465b-4c01-a470-e21a480ee465","Type":"ContainerDied","Data":"5d14ec99610d2167b484bd7712af7bb6d993d26c74b2828da1c89314f16fda22"} Oct 05 20:55:01 crc kubenswrapper[4754]: I1005 20:55:01.061782 4754 scope.go:117] "RemoveContainer" containerID="8907b9b69b249c34ec6db1e460d641b31711b75d6d31d4caafd867339e6571d0" Oct 05 20:55:01 crc kubenswrapper[4754]: E1005 20:55:01.062049 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Oct 05 20:55:01 crc kubenswrapper[4754]: I1005 20:55:01.098633 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7506d496-65ff-41dc-8230-24c8ee653a2a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33da5a6bb3cd3b042c193bc571e6622f2f87f27e560cdd0bbb9be31e2010d707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df6d700a947d19c84b4f65fd2dd2f8d1f4f1f80fa9b4800c2acfa8dae3912639\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35ca8ee77a0e58cf969bd67b277a3392614c4e73b963e6f6002097709370715d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8907b9b69b249c34ec6db1e460d641b31711b75d6d31d4caafd867339e6571d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8907b9b69b249c34ec6db1e460d641b31711b75d6d31d4caafd867339e6571d0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1005 20:54:52.589115 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 20:54:52.590181 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1304624737/tls.crt::/tmp/serving-cert-1304624737/tls.key\\\\\\\"\\\\nI1005 20:54:58.070544 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1005 20:54:58.090352 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1005 20:54:58.090382 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1005 20:54:58.090412 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1005 20:54:58.090418 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1005 20:54:58.144656 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1005 20:54:58.144688 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 20:54:58.144695 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 20:54:58.144699 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1005 20:54:58.144702 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1005 20:54:58.144705 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1005 20:54:58.144708 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1005 20:54:58.145024 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1005 20:54:58.148112 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4efb34f3493cc8854bc4cff6b29f63973c52b754404d7a21daec4e1f612d327d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54c568b637f3737667433047b956c7eb22fdf0f9301acdfd3b49fb84c39de6ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54c568b637f3737667433047b956c7eb22fdf0f9301acdfd3b49fb84c39de6ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:01Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:01 crc kubenswrapper[4754]: I1005 20:55:01.136619 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e380f9b2a0e319dc325c54d1247a6db98346292548fa688cb41a2ca1b3048419\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:01Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:01 crc kubenswrapper[4754]: I1005 20:55:01.175240 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:01Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:01 crc kubenswrapper[4754]: I1005 20:55:01.226476 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9309179-b92a-4fa4-9fba-747b8f911936\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d969d3d50d6ad280fc6654df9441bab238bb538f277ed6d1273779813ffbaaed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://800a731ed8ef57714b2318af14054ef23f0368936630887a093fc77167b33d12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8cd9b08e83dc174413cc37891c14562816787028471a4201680c86dd620b7a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93b729f3f7f993db0d6bc3ba60c04930aada3cef8272ffcc3c6a8a242a9b1475\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dde1519111fc3af5d14463fdb58d4f5c62a7045e7de98c60981c7e4da37b0e7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ea160a5b37aea7f2970ae4dd2614b09d3dcf5dcbd1fc103b9997f8ecc3a804c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ea160a5b37aea7f2970ae4dd2614b09d3dcf5dcbd1fc103b9997f8ecc3a804c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://17ae6aa6c53e7dd061f70b41428ef50bbc7ffcc65352fad1a4a9d3c69dbd2ba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17ae6aa6c53e7dd061f70b41428ef50bbc7ffcc65352fad1a4a9d3c69dbd2ba9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://ad4cd69df7608c7d87d839fe09d4557c674c7efbc4730f4e4d3daaf9a26f5aee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad4cd69df7608c7d87d839fe09d4557c674c7efbc4730f4e4d3daaf9a26f5aee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:36Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:01Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:01 crc kubenswrapper[4754]: I1005 20:55:01.255595 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:01Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:01 crc kubenswrapper[4754]: I1005 20:55:01.298478 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b44790fb-fecb-4ec3-9816-8e23dd72ab33\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35ebeb7a87d5d47b6567917d38eca38a3e622c497fba58971296c4609c287423\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35ebeb7a87d5d47b6567917d38eca38a3e622c497fba58971296c4609c287423\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7rhps\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:01Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:01 crc kubenswrapper[4754]: I1005 20:55:01.333211 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cn76s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02d5e3f9-73c2-4496-9aca-0787184aef19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7ba2b076839f4a5f586d2e6c4f4d7188254128b782bfd2180e22bf3ea5ffdd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpx5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cn76s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:01Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:01 crc kubenswrapper[4754]: I1005 20:55:01.375020 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"934bcd66-9c03-48b9-82dd-5093338f9c8f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9042df2f43243d3d79ba9a3940f8dd4f6c7a89ef9e3510e81fe60b1944d7e83b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5366bbf46a91667355f43f5df7335a536b395cda788b90b35ef18e2b53bc4c02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://600bc3bb551b747cf2f18e26745883232d7a37ac5877fad0c67100e52d7a43aa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78f74982180048eb50b9be79fa72c04f62f925b54fe511920f37e6407f864631\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:01Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:01 crc kubenswrapper[4754]: I1005 20:55:01.418423 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:01Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:01 crc kubenswrapper[4754]: I1005 20:55:01.460271 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:01Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:01 crc kubenswrapper[4754]: I1005 20:55:01.497588 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-57sr4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e854c7b9-1229-4da4-80e9-fea6479459b8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceb577e1055d84007936a50d0b3d54b9ae0299a1fc1a6618b37163a6d3fbb9cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vw2ps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:58Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-57sr4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:01Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:01 crc kubenswrapper[4754]: I1005 20:55:01.538069 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7506d496-65ff-41dc-8230-24c8ee653a2a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33da5a6bb3cd3b042c193bc571e6622f2f87f27e560cdd0bbb9be31e2010d707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df6d700a947d19c84b4f65fd2dd2f8d1f4f1f80fa9b4800c2acfa8dae3912639\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35ca8ee77a0e58cf969bd67b277a3392614c4e73b963e6f6002097709370715d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8907b9b69b249c34ec6db1e460d641b31711b75d6d31d4caafd867339e6571d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8907b9b69b249c34ec6db1e460d641b31711b75d6d31d4caafd867339e6571d0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1005 20:54:52.589115 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 20:54:52.590181 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1304624737/tls.crt::/tmp/serving-cert-1304624737/tls.key\\\\\\\"\\\\nI1005 20:54:58.070544 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1005 20:54:58.090352 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1005 20:54:58.090382 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1005 20:54:58.090412 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1005 20:54:58.090418 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1005 20:54:58.144656 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1005 20:54:58.144688 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 20:54:58.144695 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 20:54:58.144699 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1005 20:54:58.144702 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1005 20:54:58.144705 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1005 20:54:58.144708 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1005 20:54:58.145024 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1005 20:54:58.148112 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4efb34f3493cc8854bc4cff6b29f63973c52b754404d7a21daec4e1f612d327d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54c568b637f3737667433047b956c7eb22fdf0f9301acdfd3b49fb84c39de6ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54c568b637f3737667433047b956c7eb22fdf0f9301acdfd3b49fb84c39de6ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:01Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:01 crc kubenswrapper[4754]: I1005 20:55:01.577069 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c76e6e4740035a6ca54191a770c4516e28101aae69afbb89739174c2f076ef5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9eededc0cb1af94fb7b5daba08785f6b077a85e9cb5ddd8b9f397b658194f61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:01Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:01 crc kubenswrapper[4754]: I1005 20:55:01.632701 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tskt7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a68e9c89-465b-4c01-a470-e21a480ee465\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d14ec99610d2167b484bd7712af7bb6d993d26c74b2828da1c89314f16fda22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d14ec99610d2167b484bd7712af7bb6d993d26c74b2828da1c89314f16fda22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tskt7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:01Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:01 crc kubenswrapper[4754]: I1005 20:55:01.660236 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be95a413-6a59-45b4-84b7-b43cae694a26\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10b211cf6e47dbd1ee42c82e49de0a13a45581b92a0c4f6bbba0326c7149e41d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2ggg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4bb8c5903e8b70dd19eebb111097d96f725e17de44e06c3e74c067376f6c7360\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2ggg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-b2h9k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:01Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:02 crc kubenswrapper[4754]: I1005 20:55:02.071613 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-tskt7" event={"ID":"a68e9c89-465b-4c01-a470-e21a480ee465","Type":"ContainerStarted","Data":"6f6f08dd853b25881e2f3c66206aa80d0c493518284217780aa40ac4b69f0c0b"} Oct 05 20:55:02 crc kubenswrapper[4754]: I1005 20:55:02.074601 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"d9ad070bb699b643d29b82bcc1126ce3b585f7492f5ae2817fa491b309b1691d"} Oct 05 20:55:02 crc kubenswrapper[4754]: I1005 20:55:02.084311 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" event={"ID":"b44790fb-fecb-4ec3-9816-8e23dd72ab33","Type":"ContainerStarted","Data":"c3a63451ffaa14cacc3b78d11340f4ea81445fa3af96178e9b172d063d9697d5"} Oct 05 20:55:02 crc kubenswrapper[4754]: I1005 20:55:02.084367 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" event={"ID":"b44790fb-fecb-4ec3-9816-8e23dd72ab33","Type":"ContainerStarted","Data":"f3cfd3f2274a8a46fa36660e94ac6c9c298c791283f12d52330c5594ac131998"} Oct 05 20:55:02 crc kubenswrapper[4754]: I1005 20:55:02.084380 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" event={"ID":"b44790fb-fecb-4ec3-9816-8e23dd72ab33","Type":"ContainerStarted","Data":"dfdd8b1e542ab676576c311e2923fb02861f92de91fe1f4b3e4b968c549c375d"} Oct 05 20:55:02 crc kubenswrapper[4754]: I1005 20:55:02.084394 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" event={"ID":"b44790fb-fecb-4ec3-9816-8e23dd72ab33","Type":"ContainerStarted","Data":"5a7dc53e7f96e48213d724a2cb4859514df925005644108ae3a0c33b14dcf52d"} Oct 05 20:55:02 crc kubenswrapper[4754]: I1005 20:55:02.084404 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" event={"ID":"b44790fb-fecb-4ec3-9816-8e23dd72ab33","Type":"ContainerStarted","Data":"6f626cb108716da0b595844ea0251b7a04d49b5908a1d6dd0aef5cb2d402c4c9"} Oct 05 20:55:02 crc kubenswrapper[4754]: I1005 20:55:02.084415 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" event={"ID":"b44790fb-fecb-4ec3-9816-8e23dd72ab33","Type":"ContainerStarted","Data":"822223f515b5add6d71d193fd5bc77df029b8d37a1cc07be4145b53c7226bf61"} Oct 05 20:55:02 crc kubenswrapper[4754]: I1005 20:55:02.094792 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9309179-b92a-4fa4-9fba-747b8f911936\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d969d3d50d6ad280fc6654df9441bab238bb538f277ed6d1273779813ffbaaed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://800a731ed8ef57714b2318af14054ef23f0368936630887a093fc77167b33d12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8cd9b08e83dc174413cc37891c14562816787028471a4201680c86dd620b7a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93b729f3f7f993db0d6bc3ba60c04930aada3cef8272ffcc3c6a8a242a9b1475\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dde1519111fc3af5d14463fdb58d4f5c62a7045e7de98c60981c7e4da37b0e7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ea160a5b37aea7f2970ae4dd2614b09d3dcf5dcbd1fc103b9997f8ecc3a804c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ea160a5b37aea7f2970ae4dd2614b09d3dcf5dcbd1fc103b9997f8ecc3a804c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://17ae6aa6c53e7dd061f70b41428ef50bbc7ffcc65352fad1a4a9d3c69dbd2ba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17ae6aa6c53e7dd061f70b41428ef50bbc7ffcc65352fad1a4a9d3c69dbd2ba9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://ad4cd69df7608c7d87d839fe09d4557c674c7efbc4730f4e4d3daaf9a26f5aee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad4cd69df7608c7d87d839fe09d4557c674c7efbc4730f4e4d3daaf9a26f5aee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:36Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:02Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:02 crc kubenswrapper[4754]: I1005 20:55:02.106921 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:02Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:02 crc kubenswrapper[4754]: I1005 20:55:02.128219 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b44790fb-fecb-4ec3-9816-8e23dd72ab33\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35ebeb7a87d5d47b6567917d38eca38a3e622c497fba58971296c4609c287423\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35ebeb7a87d5d47b6567917d38eca38a3e622c497fba58971296c4609c287423\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7rhps\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:02Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:02 crc kubenswrapper[4754]: I1005 20:55:02.145109 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"934bcd66-9c03-48b9-82dd-5093338f9c8f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9042df2f43243d3d79ba9a3940f8dd4f6c7a89ef9e3510e81fe60b1944d7e83b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5366bbf46a91667355f43f5df7335a536b395cda788b90b35ef18e2b53bc4c02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://600bc3bb551b747cf2f18e26745883232d7a37ac5877fad0c67100e52d7a43aa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78f74982180048eb50b9be79fa72c04f62f925b54fe511920f37e6407f864631\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:02Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:02 crc kubenswrapper[4754]: I1005 20:55:02.157559 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:02Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:02 crc kubenswrapper[4754]: I1005 20:55:02.167134 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:02Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:02 crc kubenswrapper[4754]: I1005 20:55:02.177258 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-57sr4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e854c7b9-1229-4da4-80e9-fea6479459b8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceb577e1055d84007936a50d0b3d54b9ae0299a1fc1a6618b37163a6d3fbb9cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vw2ps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:58Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-57sr4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:02Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:02 crc kubenswrapper[4754]: I1005 20:55:02.193283 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cn76s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02d5e3f9-73c2-4496-9aca-0787184aef19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7ba2b076839f4a5f586d2e6c4f4d7188254128b782bfd2180e22bf3ea5ffdd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpx5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cn76s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:02Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:02 crc kubenswrapper[4754]: I1005 20:55:02.208649 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7506d496-65ff-41dc-8230-24c8ee653a2a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33da5a6bb3cd3b042c193bc571e6622f2f87f27e560cdd0bbb9be31e2010d707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df6d700a947d19c84b4f65fd2dd2f8d1f4f1f80fa9b4800c2acfa8dae3912639\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35ca8ee77a0e58cf969bd67b277a3392614c4e73b963e6f6002097709370715d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8907b9b69b249c34ec6db1e460d641b31711b75d6d31d4caafd867339e6571d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8907b9b69b249c34ec6db1e460d641b31711b75d6d31d4caafd867339e6571d0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1005 20:54:52.589115 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 20:54:52.590181 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1304624737/tls.crt::/tmp/serving-cert-1304624737/tls.key\\\\\\\"\\\\nI1005 20:54:58.070544 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1005 20:54:58.090352 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1005 20:54:58.090382 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1005 20:54:58.090412 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1005 20:54:58.090418 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1005 20:54:58.144656 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1005 20:54:58.144688 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 20:54:58.144695 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 20:54:58.144699 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1005 20:54:58.144702 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1005 20:54:58.144705 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1005 20:54:58.144708 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1005 20:54:58.145024 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1005 20:54:58.148112 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4efb34f3493cc8854bc4cff6b29f63973c52b754404d7a21daec4e1f612d327d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54c568b637f3737667433047b956c7eb22fdf0f9301acdfd3b49fb84c39de6ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54c568b637f3737667433047b956c7eb22fdf0f9301acdfd3b49fb84c39de6ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:02Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:02 crc kubenswrapper[4754]: I1005 20:55:02.226800 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c76e6e4740035a6ca54191a770c4516e28101aae69afbb89739174c2f076ef5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9eededc0cb1af94fb7b5daba08785f6b077a85e9cb5ddd8b9f397b658194f61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:02Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:02 crc kubenswrapper[4754]: I1005 20:55:02.247370 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tskt7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a68e9c89-465b-4c01-a470-e21a480ee465\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d14ec99610d2167b484bd7712af7bb6d993d26c74b2828da1c89314f16fda22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d14ec99610d2167b484bd7712af7bb6d993d26c74b2828da1c89314f16fda22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f6f08dd853b25881e2f3c66206aa80d0c493518284217780aa40ac4b69f0c0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tskt7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:02Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:02 crc kubenswrapper[4754]: I1005 20:55:02.260694 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be95a413-6a59-45b4-84b7-b43cae694a26\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10b211cf6e47dbd1ee42c82e49de0a13a45581b92a0c4f6bbba0326c7149e41d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2ggg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4bb8c5903e8b70dd19eebb111097d96f725e17de44e06c3e74c067376f6c7360\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2ggg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-b2h9k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:02Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:02 crc kubenswrapper[4754]: I1005 20:55:02.274759 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e380f9b2a0e319dc325c54d1247a6db98346292548fa688cb41a2ca1b3048419\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:02Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:02 crc kubenswrapper[4754]: I1005 20:55:02.284920 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:02Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:02 crc kubenswrapper[4754]: I1005 20:55:02.294586 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9ad070bb699b643d29b82bcc1126ce3b585f7492f5ae2817fa491b309b1691d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:02Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:02 crc kubenswrapper[4754]: I1005 20:55:02.303106 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-57sr4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e854c7b9-1229-4da4-80e9-fea6479459b8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceb577e1055d84007936a50d0b3d54b9ae0299a1fc1a6618b37163a6d3fbb9cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vw2ps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:58Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-57sr4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:02Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:02 crc kubenswrapper[4754]: I1005 20:55:02.335149 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cn76s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02d5e3f9-73c2-4496-9aca-0787184aef19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7ba2b076839f4a5f586d2e6c4f4d7188254128b782bfd2180e22bf3ea5ffdd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpx5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cn76s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:02Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:02 crc kubenswrapper[4754]: I1005 20:55:02.374388 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"934bcd66-9c03-48b9-82dd-5093338f9c8f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9042df2f43243d3d79ba9a3940f8dd4f6c7a89ef9e3510e81fe60b1944d7e83b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5366bbf46a91667355f43f5df7335a536b395cda788b90b35ef18e2b53bc4c02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://600bc3bb551b747cf2f18e26745883232d7a37ac5877fad0c67100e52d7a43aa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78f74982180048eb50b9be79fa72c04f62f925b54fe511920f37e6407f864631\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:02Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:02 crc kubenswrapper[4754]: I1005 20:55:02.413166 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:02Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:02 crc kubenswrapper[4754]: I1005 20:55:02.419597 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 20:55:02 crc kubenswrapper[4754]: E1005 20:55:02.419929 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 20:55:06.41989266 +0000 UTC m=+30.324011370 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:55:02 crc kubenswrapper[4754]: I1005 20:55:02.445151 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-ss7lm"] Oct 05 20:55:02 crc kubenswrapper[4754]: I1005 20:55:02.446015 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-ss7lm" Oct 05 20:55:02 crc kubenswrapper[4754]: I1005 20:55:02.464014 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tskt7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a68e9c89-465b-4c01-a470-e21a480ee465\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d14ec99610d2167b484bd7712af7bb6d993d26c74b2828da1c89314f16fda22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d14ec99610d2167b484bd7712af7bb6d993d26c74b2828da1c89314f16fda22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f6f08dd853b25881e2f3c66206aa80d0c493518284217780aa40ac4b69f0c0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tskt7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:02Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:02 crc kubenswrapper[4754]: I1005 20:55:02.465695 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Oct 05 20:55:02 crc kubenswrapper[4754]: I1005 20:55:02.484075 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Oct 05 20:55:02 crc kubenswrapper[4754]: I1005 20:55:02.503736 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Oct 05 20:55:02 crc kubenswrapper[4754]: I1005 20:55:02.520594 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 20:55:02 crc kubenswrapper[4754]: I1005 20:55:02.520755 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 20:55:02 crc kubenswrapper[4754]: I1005 20:55:02.520829 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 20:55:02 crc kubenswrapper[4754]: I1005 20:55:02.520890 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4wpwc\" (UniqueName: \"kubernetes.io/projected/4f9ef817-0622-4185-a7fb-57221690aa75-kube-api-access-4wpwc\") pod \"node-ca-ss7lm\" (UID: \"4f9ef817-0622-4185-a7fb-57221690aa75\") " pod="openshift-image-registry/node-ca-ss7lm" Oct 05 20:55:02 crc kubenswrapper[4754]: I1005 20:55:02.520967 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/4f9ef817-0622-4185-a7fb-57221690aa75-serviceca\") pod \"node-ca-ss7lm\" (UID: \"4f9ef817-0622-4185-a7fb-57221690aa75\") " pod="openshift-image-registry/node-ca-ss7lm" Oct 05 20:55:02 crc kubenswrapper[4754]: I1005 20:55:02.521026 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4f9ef817-0622-4185-a7fb-57221690aa75-host\") pod \"node-ca-ss7lm\" (UID: \"4f9ef817-0622-4185-a7fb-57221690aa75\") " pod="openshift-image-registry/node-ca-ss7lm" Oct 05 20:55:02 crc kubenswrapper[4754]: E1005 20:55:02.521065 4754 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 05 20:55:02 crc kubenswrapper[4754]: E1005 20:55:02.521182 4754 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 05 20:55:02 crc kubenswrapper[4754]: E1005 20:55:02.521214 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-05 20:55:06.521166866 +0000 UTC m=+30.425285606 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 05 20:55:02 crc kubenswrapper[4754]: I1005 20:55:02.521084 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 20:55:02 crc kubenswrapper[4754]: E1005 20:55:02.521279 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-05 20:55:06.521253458 +0000 UTC m=+30.425372198 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 05 20:55:02 crc kubenswrapper[4754]: E1005 20:55:02.521403 4754 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 05 20:55:02 crc kubenswrapper[4754]: E1005 20:55:02.521479 4754 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 05 20:55:02 crc kubenswrapper[4754]: E1005 20:55:02.521561 4754 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 05 20:55:02 crc kubenswrapper[4754]: E1005 20:55:02.521660 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-05 20:55:06.521642858 +0000 UTC m=+30.425761568 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 05 20:55:02 crc kubenswrapper[4754]: E1005 20:55:02.521774 4754 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 05 20:55:02 crc kubenswrapper[4754]: E1005 20:55:02.521856 4754 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 05 20:55:02 crc kubenswrapper[4754]: E1005 20:55:02.521926 4754 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 05 20:55:02 crc kubenswrapper[4754]: E1005 20:55:02.522008 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-05 20:55:06.521997268 +0000 UTC m=+30.426115978 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 05 20:55:02 crc kubenswrapper[4754]: I1005 20:55:02.523262 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Oct 05 20:55:02 crc kubenswrapper[4754]: I1005 20:55:02.575906 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be95a413-6a59-45b4-84b7-b43cae694a26\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10b211cf6e47dbd1ee42c82e49de0a13a45581b92a0c4f6bbba0326c7149e41d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2ggg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4bb8c5903e8b70dd19eebb111097d96f725e17de44e06c3e74c067376f6c7360\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2ggg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-b2h9k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:02Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:02 crc kubenswrapper[4754]: I1005 20:55:02.622132 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4wpwc\" (UniqueName: \"kubernetes.io/projected/4f9ef817-0622-4185-a7fb-57221690aa75-kube-api-access-4wpwc\") pod \"node-ca-ss7lm\" (UID: \"4f9ef817-0622-4185-a7fb-57221690aa75\") " pod="openshift-image-registry/node-ca-ss7lm" Oct 05 20:55:02 crc kubenswrapper[4754]: I1005 20:55:02.622194 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/4f9ef817-0622-4185-a7fb-57221690aa75-serviceca\") pod \"node-ca-ss7lm\" (UID: \"4f9ef817-0622-4185-a7fb-57221690aa75\") " pod="openshift-image-registry/node-ca-ss7lm" Oct 05 20:55:02 crc kubenswrapper[4754]: I1005 20:55:02.622084 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7506d496-65ff-41dc-8230-24c8ee653a2a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33da5a6bb3cd3b042c193bc571e6622f2f87f27e560cdd0bbb9be31e2010d707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df6d700a947d19c84b4f65fd2dd2f8d1f4f1f80fa9b4800c2acfa8dae3912639\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35ca8ee77a0e58cf969bd67b277a3392614c4e73b963e6f6002097709370715d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8907b9b69b249c34ec6db1e460d641b31711b75d6d31d4caafd867339e6571d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8907b9b69b249c34ec6db1e460d641b31711b75d6d31d4caafd867339e6571d0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1005 20:54:52.589115 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 20:54:52.590181 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1304624737/tls.crt::/tmp/serving-cert-1304624737/tls.key\\\\\\\"\\\\nI1005 20:54:58.070544 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1005 20:54:58.090352 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1005 20:54:58.090382 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1005 20:54:58.090412 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1005 20:54:58.090418 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1005 20:54:58.144656 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1005 20:54:58.144688 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 20:54:58.144695 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 20:54:58.144699 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1005 20:54:58.144702 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1005 20:54:58.144705 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1005 20:54:58.144708 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1005 20:54:58.145024 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1005 20:54:58.148112 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4efb34f3493cc8854bc4cff6b29f63973c52b754404d7a21daec4e1f612d327d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54c568b637f3737667433047b956c7eb22fdf0f9301acdfd3b49fb84c39de6ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54c568b637f3737667433047b956c7eb22fdf0f9301acdfd3b49fb84c39de6ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:02Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:02 crc kubenswrapper[4754]: I1005 20:55:02.622303 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4f9ef817-0622-4185-a7fb-57221690aa75-host\") pod \"node-ca-ss7lm\" (UID: \"4f9ef817-0622-4185-a7fb-57221690aa75\") " pod="openshift-image-registry/node-ca-ss7lm" Oct 05 20:55:02 crc kubenswrapper[4754]: I1005 20:55:02.622239 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4f9ef817-0622-4185-a7fb-57221690aa75-host\") pod \"node-ca-ss7lm\" (UID: \"4f9ef817-0622-4185-a7fb-57221690aa75\") " pod="openshift-image-registry/node-ca-ss7lm" Oct 05 20:55:02 crc kubenswrapper[4754]: I1005 20:55:02.624176 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/4f9ef817-0622-4185-a7fb-57221690aa75-serviceca\") pod \"node-ca-ss7lm\" (UID: \"4f9ef817-0622-4185-a7fb-57221690aa75\") " pod="openshift-image-registry/node-ca-ss7lm" Oct 05 20:55:02 crc kubenswrapper[4754]: I1005 20:55:02.669468 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4wpwc\" (UniqueName: \"kubernetes.io/projected/4f9ef817-0622-4185-a7fb-57221690aa75-kube-api-access-4wpwc\") pod \"node-ca-ss7lm\" (UID: \"4f9ef817-0622-4185-a7fb-57221690aa75\") " pod="openshift-image-registry/node-ca-ss7lm" Oct 05 20:55:02 crc kubenswrapper[4754]: I1005 20:55:02.683075 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c76e6e4740035a6ca54191a770c4516e28101aae69afbb89739174c2f076ef5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9eededc0cb1af94fb7b5daba08785f6b077a85e9cb5ddd8b9f397b658194f61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:02Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:02 crc kubenswrapper[4754]: I1005 20:55:02.716552 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e380f9b2a0e319dc325c54d1247a6db98346292548fa688cb41a2ca1b3048419\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:02Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:02 crc kubenswrapper[4754]: I1005 20:55:02.754286 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:02Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:02 crc kubenswrapper[4754]: I1005 20:55:02.766291 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-ss7lm" Oct 05 20:55:02 crc kubenswrapper[4754]: I1005 20:55:02.814275 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b44790fb-fecb-4ec3-9816-8e23dd72ab33\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35ebeb7a87d5d47b6567917d38eca38a3e622c497fba58971296c4609c287423\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35ebeb7a87d5d47b6567917d38eca38a3e622c497fba58971296c4609c287423\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7rhps\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:02Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:02 crc kubenswrapper[4754]: I1005 20:55:02.838704 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 20:55:02 crc kubenswrapper[4754]: I1005 20:55:02.838705 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 20:55:02 crc kubenswrapper[4754]: I1005 20:55:02.838845 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 20:55:02 crc kubenswrapper[4754]: E1005 20:55:02.838923 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 20:55:02 crc kubenswrapper[4754]: E1005 20:55:02.839106 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 20:55:02 crc kubenswrapper[4754]: E1005 20:55:02.839299 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 20:55:02 crc kubenswrapper[4754]: I1005 20:55:02.852557 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9309179-b92a-4fa4-9fba-747b8f911936\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d969d3d50d6ad280fc6654df9441bab238bb538f277ed6d1273779813ffbaaed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://800a731ed8ef57714b2318af14054ef23f0368936630887a093fc77167b33d12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8cd9b08e83dc174413cc37891c14562816787028471a4201680c86dd620b7a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93b729f3f7f993db0d6bc3ba60c04930aada3cef8272ffcc3c6a8a242a9b1475\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dde1519111fc3af5d14463fdb58d4f5c62a7045e7de98c60981c7e4da37b0e7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ea160a5b37aea7f2970ae4dd2614b09d3dcf5dcbd1fc103b9997f8ecc3a804c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ea160a5b37aea7f2970ae4dd2614b09d3dcf5dcbd1fc103b9997f8ecc3a804c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://17ae6aa6c53e7dd061f70b41428ef50bbc7ffcc65352fad1a4a9d3c69dbd2ba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17ae6aa6c53e7dd061f70b41428ef50bbc7ffcc65352fad1a4a9d3c69dbd2ba9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://ad4cd69df7608c7d87d839fe09d4557c674c7efbc4730f4e4d3daaf9a26f5aee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad4cd69df7608c7d87d839fe09d4557c674c7efbc4730f4e4d3daaf9a26f5aee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:36Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:02Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:02 crc kubenswrapper[4754]: I1005 20:55:02.881662 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:02Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:02 crc kubenswrapper[4754]: I1005 20:55:02.917728 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ss7lm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f9ef817-0622-4185-a7fb-57221690aa75\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wpwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:55:02Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ss7lm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:02Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:02 crc kubenswrapper[4754]: I1005 20:55:02.956688 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"934bcd66-9c03-48b9-82dd-5093338f9c8f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9042df2f43243d3d79ba9a3940f8dd4f6c7a89ef9e3510e81fe60b1944d7e83b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5366bbf46a91667355f43f5df7335a536b395cda788b90b35ef18e2b53bc4c02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://600bc3bb551b747cf2f18e26745883232d7a37ac5877fad0c67100e52d7a43aa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78f74982180048eb50b9be79fa72c04f62f925b54fe511920f37e6407f864631\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:02Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:02 crc kubenswrapper[4754]: I1005 20:55:02.998195 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:02Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:03 crc kubenswrapper[4754]: I1005 20:55:03.039144 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9ad070bb699b643d29b82bcc1126ce3b585f7492f5ae2817fa491b309b1691d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:03Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:03 crc kubenswrapper[4754]: I1005 20:55:03.077714 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-57sr4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e854c7b9-1229-4da4-80e9-fea6479459b8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceb577e1055d84007936a50d0b3d54b9ae0299a1fc1a6618b37163a6d3fbb9cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vw2ps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:58Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-57sr4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:03Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:03 crc kubenswrapper[4754]: I1005 20:55:03.096102 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-ss7lm" event={"ID":"4f9ef817-0622-4185-a7fb-57221690aa75","Type":"ContainerStarted","Data":"2ba613b0d7753a003c572c3733ed394707c45b6f8072cb7aa88f3c7312b8882b"} Oct 05 20:55:03 crc kubenswrapper[4754]: I1005 20:55:03.096185 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-ss7lm" event={"ID":"4f9ef817-0622-4185-a7fb-57221690aa75","Type":"ContainerStarted","Data":"5a6889b99050e0a0c9cae9a8c4b3f0d60a55c3ea879e2f2034c0aa5c2dd22d2d"} Oct 05 20:55:03 crc kubenswrapper[4754]: I1005 20:55:03.099480 4754 generic.go:334] "Generic (PLEG): container finished" podID="a68e9c89-465b-4c01-a470-e21a480ee465" containerID="6f6f08dd853b25881e2f3c66206aa80d0c493518284217780aa40ac4b69f0c0b" exitCode=0 Oct 05 20:55:03 crc kubenswrapper[4754]: I1005 20:55:03.099565 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-tskt7" event={"ID":"a68e9c89-465b-4c01-a470-e21a480ee465","Type":"ContainerDied","Data":"6f6f08dd853b25881e2f3c66206aa80d0c493518284217780aa40ac4b69f0c0b"} Oct 05 20:55:03 crc kubenswrapper[4754]: I1005 20:55:03.119346 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cn76s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02d5e3f9-73c2-4496-9aca-0787184aef19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7ba2b076839f4a5f586d2e6c4f4d7188254128b782bfd2180e22bf3ea5ffdd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpx5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cn76s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:03Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:03 crc kubenswrapper[4754]: I1005 20:55:03.162637 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7506d496-65ff-41dc-8230-24c8ee653a2a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33da5a6bb3cd3b042c193bc571e6622f2f87f27e560cdd0bbb9be31e2010d707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df6d700a947d19c84b4f65fd2dd2f8d1f4f1f80fa9b4800c2acfa8dae3912639\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35ca8ee77a0e58cf969bd67b277a3392614c4e73b963e6f6002097709370715d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8907b9b69b249c34ec6db1e460d641b31711b75d6d31d4caafd867339e6571d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8907b9b69b249c34ec6db1e460d641b31711b75d6d31d4caafd867339e6571d0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1005 20:54:52.589115 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 20:54:52.590181 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1304624737/tls.crt::/tmp/serving-cert-1304624737/tls.key\\\\\\\"\\\\nI1005 20:54:58.070544 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1005 20:54:58.090352 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1005 20:54:58.090382 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1005 20:54:58.090412 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1005 20:54:58.090418 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1005 20:54:58.144656 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1005 20:54:58.144688 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 20:54:58.144695 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 20:54:58.144699 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1005 20:54:58.144702 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1005 20:54:58.144705 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1005 20:54:58.144708 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1005 20:54:58.145024 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1005 20:54:58.148112 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4efb34f3493cc8854bc4cff6b29f63973c52b754404d7a21daec4e1f612d327d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54c568b637f3737667433047b956c7eb22fdf0f9301acdfd3b49fb84c39de6ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54c568b637f3737667433047b956c7eb22fdf0f9301acdfd3b49fb84c39de6ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:03Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:03 crc kubenswrapper[4754]: I1005 20:55:03.198471 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c76e6e4740035a6ca54191a770c4516e28101aae69afbb89739174c2f076ef5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9eededc0cb1af94fb7b5daba08785f6b077a85e9cb5ddd8b9f397b658194f61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:03Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:03 crc kubenswrapper[4754]: I1005 20:55:03.233400 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tskt7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a68e9c89-465b-4c01-a470-e21a480ee465\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d14ec99610d2167b484bd7712af7bb6d993d26c74b2828da1c89314f16fda22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d14ec99610d2167b484bd7712af7bb6d993d26c74b2828da1c89314f16fda22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f6f08dd853b25881e2f3c66206aa80d0c493518284217780aa40ac4b69f0c0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tskt7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:03Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:03 crc kubenswrapper[4754]: I1005 20:55:03.272164 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be95a413-6a59-45b4-84b7-b43cae694a26\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10b211cf6e47dbd1ee42c82e49de0a13a45581b92a0c4f6bbba0326c7149e41d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2ggg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4bb8c5903e8b70dd19eebb111097d96f725e17de44e06c3e74c067376f6c7360\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2ggg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-b2h9k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:03Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:03 crc kubenswrapper[4754]: I1005 20:55:03.317187 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e380f9b2a0e319dc325c54d1247a6db98346292548fa688cb41a2ca1b3048419\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:03Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:03 crc kubenswrapper[4754]: I1005 20:55:03.356244 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:03Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:03 crc kubenswrapper[4754]: I1005 20:55:03.411116 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9309179-b92a-4fa4-9fba-747b8f911936\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d969d3d50d6ad280fc6654df9441bab238bb538f277ed6d1273779813ffbaaed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://800a731ed8ef57714b2318af14054ef23f0368936630887a093fc77167b33d12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8cd9b08e83dc174413cc37891c14562816787028471a4201680c86dd620b7a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93b729f3f7f993db0d6bc3ba60c04930aada3cef8272ffcc3c6a8a242a9b1475\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dde1519111fc3af5d14463fdb58d4f5c62a7045e7de98c60981c7e4da37b0e7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ea160a5b37aea7f2970ae4dd2614b09d3dcf5dcbd1fc103b9997f8ecc3a804c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ea160a5b37aea7f2970ae4dd2614b09d3dcf5dcbd1fc103b9997f8ecc3a804c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://17ae6aa6c53e7dd061f70b41428ef50bbc7ffcc65352fad1a4a9d3c69dbd2ba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17ae6aa6c53e7dd061f70b41428ef50bbc7ffcc65352fad1a4a9d3c69dbd2ba9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://ad4cd69df7608c7d87d839fe09d4557c674c7efbc4730f4e4d3daaf9a26f5aee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad4cd69df7608c7d87d839fe09d4557c674c7efbc4730f4e4d3daaf9a26f5aee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:36Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:03Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:03 crc kubenswrapper[4754]: I1005 20:55:03.439150 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:03Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:03 crc kubenswrapper[4754]: I1005 20:55:03.477352 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b44790fb-fecb-4ec3-9816-8e23dd72ab33\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35ebeb7a87d5d47b6567917d38eca38a3e622c497fba58971296c4609c287423\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35ebeb7a87d5d47b6567917d38eca38a3e622c497fba58971296c4609c287423\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7rhps\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:03Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:03 crc kubenswrapper[4754]: I1005 20:55:03.513194 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cn76s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02d5e3f9-73c2-4496-9aca-0787184aef19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7ba2b076839f4a5f586d2e6c4f4d7188254128b782bfd2180e22bf3ea5ffdd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpx5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cn76s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:03Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:03 crc kubenswrapper[4754]: I1005 20:55:03.555319 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ss7lm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f9ef817-0622-4185-a7fb-57221690aa75\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ba613b0d7753a003c572c3733ed394707c45b6f8072cb7aa88f3c7312b8882b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wpwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:55:02Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ss7lm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:03Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:03 crc kubenswrapper[4754]: I1005 20:55:03.599146 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"934bcd66-9c03-48b9-82dd-5093338f9c8f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9042df2f43243d3d79ba9a3940f8dd4f6c7a89ef9e3510e81fe60b1944d7e83b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5366bbf46a91667355f43f5df7335a536b395cda788b90b35ef18e2b53bc4c02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://600bc3bb551b747cf2f18e26745883232d7a37ac5877fad0c67100e52d7a43aa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78f74982180048eb50b9be79fa72c04f62f925b54fe511920f37e6407f864631\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:03Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:03 crc kubenswrapper[4754]: I1005 20:55:03.640488 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:03Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:03 crc kubenswrapper[4754]: I1005 20:55:03.672154 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9ad070bb699b643d29b82bcc1126ce3b585f7492f5ae2817fa491b309b1691d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:03Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:03 crc kubenswrapper[4754]: I1005 20:55:03.713821 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-57sr4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e854c7b9-1229-4da4-80e9-fea6479459b8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceb577e1055d84007936a50d0b3d54b9ae0299a1fc1a6618b37163a6d3fbb9cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vw2ps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:58Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-57sr4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:03Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:03 crc kubenswrapper[4754]: I1005 20:55:03.761126 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7506d496-65ff-41dc-8230-24c8ee653a2a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33da5a6bb3cd3b042c193bc571e6622f2f87f27e560cdd0bbb9be31e2010d707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df6d700a947d19c84b4f65fd2dd2f8d1f4f1f80fa9b4800c2acfa8dae3912639\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35ca8ee77a0e58cf969bd67b277a3392614c4e73b963e6f6002097709370715d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8907b9b69b249c34ec6db1e460d641b31711b75d6d31d4caafd867339e6571d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8907b9b69b249c34ec6db1e460d641b31711b75d6d31d4caafd867339e6571d0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1005 20:54:52.589115 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 20:54:52.590181 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1304624737/tls.crt::/tmp/serving-cert-1304624737/tls.key\\\\\\\"\\\\nI1005 20:54:58.070544 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1005 20:54:58.090352 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1005 20:54:58.090382 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1005 20:54:58.090412 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1005 20:54:58.090418 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1005 20:54:58.144656 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1005 20:54:58.144688 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 20:54:58.144695 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 20:54:58.144699 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1005 20:54:58.144702 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1005 20:54:58.144705 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1005 20:54:58.144708 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1005 20:54:58.145024 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1005 20:54:58.148112 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4efb34f3493cc8854bc4cff6b29f63973c52b754404d7a21daec4e1f612d327d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54c568b637f3737667433047b956c7eb22fdf0f9301acdfd3b49fb84c39de6ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54c568b637f3737667433047b956c7eb22fdf0f9301acdfd3b49fb84c39de6ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:03Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:03 crc kubenswrapper[4754]: I1005 20:55:03.797795 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c76e6e4740035a6ca54191a770c4516e28101aae69afbb89739174c2f076ef5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9eededc0cb1af94fb7b5daba08785f6b077a85e9cb5ddd8b9f397b658194f61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:03Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:03 crc kubenswrapper[4754]: I1005 20:55:03.841310 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tskt7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a68e9c89-465b-4c01-a470-e21a480ee465\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d14ec99610d2167b484bd7712af7bb6d993d26c74b2828da1c89314f16fda22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d14ec99610d2167b484bd7712af7bb6d993d26c74b2828da1c89314f16fda22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f6f08dd853b25881e2f3c66206aa80d0c493518284217780aa40ac4b69f0c0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f6f08dd853b25881e2f3c66206aa80d0c493518284217780aa40ac4b69f0c0b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tskt7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:03Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:03 crc kubenswrapper[4754]: I1005 20:55:03.875472 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be95a413-6a59-45b4-84b7-b43cae694a26\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10b211cf6e47dbd1ee42c82e49de0a13a45581b92a0c4f6bbba0326c7149e41d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2ggg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4bb8c5903e8b70dd19eebb111097d96f725e17de44e06c3e74c067376f6c7360\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2ggg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-b2h9k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:03Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:03 crc kubenswrapper[4754]: I1005 20:55:03.919522 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e380f9b2a0e319dc325c54d1247a6db98346292548fa688cb41a2ca1b3048419\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:03Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:03 crc kubenswrapper[4754]: I1005 20:55:03.958422 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:03Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:04 crc kubenswrapper[4754]: I1005 20:55:04.009817 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9309179-b92a-4fa4-9fba-747b8f911936\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d969d3d50d6ad280fc6654df9441bab238bb538f277ed6d1273779813ffbaaed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://800a731ed8ef57714b2318af14054ef23f0368936630887a093fc77167b33d12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8cd9b08e83dc174413cc37891c14562816787028471a4201680c86dd620b7a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93b729f3f7f993db0d6bc3ba60c04930aada3cef8272ffcc3c6a8a242a9b1475\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dde1519111fc3af5d14463fdb58d4f5c62a7045e7de98c60981c7e4da37b0e7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ea160a5b37aea7f2970ae4dd2614b09d3dcf5dcbd1fc103b9997f8ecc3a804c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ea160a5b37aea7f2970ae4dd2614b09d3dcf5dcbd1fc103b9997f8ecc3a804c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://17ae6aa6c53e7dd061f70b41428ef50bbc7ffcc65352fad1a4a9d3c69dbd2ba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17ae6aa6c53e7dd061f70b41428ef50bbc7ffcc65352fad1a4a9d3c69dbd2ba9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://ad4cd69df7608c7d87d839fe09d4557c674c7efbc4730f4e4d3daaf9a26f5aee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad4cd69df7608c7d87d839fe09d4557c674c7efbc4730f4e4d3daaf9a26f5aee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:36Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:04Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:04 crc kubenswrapper[4754]: I1005 20:55:04.039545 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:04Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:04 crc kubenswrapper[4754]: I1005 20:55:04.082316 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b44790fb-fecb-4ec3-9816-8e23dd72ab33\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35ebeb7a87d5d47b6567917d38eca38a3e622c497fba58971296c4609c287423\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35ebeb7a87d5d47b6567917d38eca38a3e622c497fba58971296c4609c287423\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7rhps\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:04Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:04 crc kubenswrapper[4754]: I1005 20:55:04.107661 4754 generic.go:334] "Generic (PLEG): container finished" podID="a68e9c89-465b-4c01-a470-e21a480ee465" containerID="71026d00ec1a3d91b019deb3c9f43ee52fb05b8f2d6374e1a5b762c9e5c1b10d" exitCode=0 Oct 05 20:55:04 crc kubenswrapper[4754]: I1005 20:55:04.107745 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-tskt7" event={"ID":"a68e9c89-465b-4c01-a470-e21a480ee465","Type":"ContainerDied","Data":"71026d00ec1a3d91b019deb3c9f43ee52fb05b8f2d6374e1a5b762c9e5c1b10d"} Oct 05 20:55:04 crc kubenswrapper[4754]: I1005 20:55:04.131927 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:04Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:04 crc kubenswrapper[4754]: I1005 20:55:04.159588 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e380f9b2a0e319dc325c54d1247a6db98346292548fa688cb41a2ca1b3048419\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:04Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:04 crc kubenswrapper[4754]: I1005 20:55:04.195277 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 20:55:04 crc kubenswrapper[4754]: I1005 20:55:04.200289 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:04 crc kubenswrapper[4754]: I1005 20:55:04.200336 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:04 crc kubenswrapper[4754]: I1005 20:55:04.200354 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:04 crc kubenswrapper[4754]: I1005 20:55:04.205302 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9309179-b92a-4fa4-9fba-747b8f911936\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d969d3d50d6ad280fc6654df9441bab238bb538f277ed6d1273779813ffbaaed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://800a731ed8ef57714b2318af14054ef23f0368936630887a093fc77167b33d12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8cd9b08e83dc174413cc37891c14562816787028471a4201680c86dd620b7a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93b729f3f7f993db0d6bc3ba60c04930aada3cef8272ffcc3c6a8a242a9b1475\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dde1519111fc3af5d14463fdb58d4f5c62a7045e7de98c60981c7e4da37b0e7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ea160a5b37aea7f2970ae4dd2614b09d3dcf5dcbd1fc103b9997f8ecc3a804c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ea160a5b37aea7f2970ae4dd2614b09d3dcf5dcbd1fc103b9997f8ecc3a804c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://17ae6aa6c53e7dd061f70b41428ef50bbc7ffcc65352fad1a4a9d3c69dbd2ba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17ae6aa6c53e7dd061f70b41428ef50bbc7ffcc65352fad1a4a9d3c69dbd2ba9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://ad4cd69df7608c7d87d839fe09d4557c674c7efbc4730f4e4d3daaf9a26f5aee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad4cd69df7608c7d87d839fe09d4557c674c7efbc4730f4e4d3daaf9a26f5aee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:36Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:04Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:04 crc kubenswrapper[4754]: I1005 20:55:04.209596 4754 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 05 20:55:04 crc kubenswrapper[4754]: I1005 20:55:04.235803 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:04Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:04 crc kubenswrapper[4754]: I1005 20:55:04.285598 4754 kubelet_node_status.go:115] "Node was previously registered" node="crc" Oct 05 20:55:04 crc kubenswrapper[4754]: I1005 20:55:04.286026 4754 kubelet_node_status.go:79] "Successfully registered node" node="crc" Oct 05 20:55:04 crc kubenswrapper[4754]: I1005 20:55:04.287755 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:04 crc kubenswrapper[4754]: I1005 20:55:04.287838 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:04 crc kubenswrapper[4754]: I1005 20:55:04.287862 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:04 crc kubenswrapper[4754]: I1005 20:55:04.287897 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:04 crc kubenswrapper[4754]: I1005 20:55:04.287934 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:04Z","lastTransitionTime":"2025-10-05T20:55:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:04 crc kubenswrapper[4754]: E1005 20:55:04.305947 4754 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T20:55:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T20:55:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T20:55:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T20:55:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ede2b442-76b0-443a-a450-252ee8711c80\\\",\\\"systemUUID\\\":\\\"72d081dd-fe7e-42a6-9cf0-f0ee7f733ac9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:04Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:04 crc kubenswrapper[4754]: I1005 20:55:04.317219 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:04 crc kubenswrapper[4754]: I1005 20:55:04.317285 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:04 crc kubenswrapper[4754]: I1005 20:55:04.317305 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:04 crc kubenswrapper[4754]: I1005 20:55:04.317330 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:04 crc kubenswrapper[4754]: I1005 20:55:04.317349 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:04Z","lastTransitionTime":"2025-10-05T20:55:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:04 crc kubenswrapper[4754]: I1005 20:55:04.327684 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b44790fb-fecb-4ec3-9816-8e23dd72ab33\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35ebeb7a87d5d47b6567917d38eca38a3e622c497fba58971296c4609c287423\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35ebeb7a87d5d47b6567917d38eca38a3e622c497fba58971296c4609c287423\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7rhps\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:04Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:04 crc kubenswrapper[4754]: E1005 20:55:04.334486 4754 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T20:55:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T20:55:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T20:55:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T20:55:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ede2b442-76b0-443a-a450-252ee8711c80\\\",\\\"systemUUID\\\":\\\"72d081dd-fe7e-42a6-9cf0-f0ee7f733ac9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:04Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:04 crc kubenswrapper[4754]: I1005 20:55:04.340552 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:04 crc kubenswrapper[4754]: I1005 20:55:04.340611 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:04 crc kubenswrapper[4754]: I1005 20:55:04.340627 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:04 crc kubenswrapper[4754]: I1005 20:55:04.340651 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:04 crc kubenswrapper[4754]: I1005 20:55:04.340667 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:04Z","lastTransitionTime":"2025-10-05T20:55:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:04 crc kubenswrapper[4754]: I1005 20:55:04.349842 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-57sr4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e854c7b9-1229-4da4-80e9-fea6479459b8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceb577e1055d84007936a50d0b3d54b9ae0299a1fc1a6618b37163a6d3fbb9cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vw2ps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:58Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-57sr4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:04Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:04 crc kubenswrapper[4754]: E1005 20:55:04.354171 4754 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T20:55:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T20:55:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T20:55:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T20:55:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ede2b442-76b0-443a-a450-252ee8711c80\\\",\\\"systemUUID\\\":\\\"72d081dd-fe7e-42a6-9cf0-f0ee7f733ac9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:04Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:04 crc kubenswrapper[4754]: I1005 20:55:04.358579 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:04 crc kubenswrapper[4754]: I1005 20:55:04.358619 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:04 crc kubenswrapper[4754]: I1005 20:55:04.358631 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:04 crc kubenswrapper[4754]: I1005 20:55:04.358656 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:04 crc kubenswrapper[4754]: I1005 20:55:04.358672 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:04Z","lastTransitionTime":"2025-10-05T20:55:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:04 crc kubenswrapper[4754]: E1005 20:55:04.372940 4754 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T20:55:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T20:55:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T20:55:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T20:55:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ede2b442-76b0-443a-a450-252ee8711c80\\\",\\\"systemUUID\\\":\\\"72d081dd-fe7e-42a6-9cf0-f0ee7f733ac9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:04Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:04 crc kubenswrapper[4754]: I1005 20:55:04.377148 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:04 crc kubenswrapper[4754]: I1005 20:55:04.377198 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:04 crc kubenswrapper[4754]: I1005 20:55:04.377210 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:04 crc kubenswrapper[4754]: I1005 20:55:04.377230 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:04 crc kubenswrapper[4754]: I1005 20:55:04.377243 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:04Z","lastTransitionTime":"2025-10-05T20:55:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:04 crc kubenswrapper[4754]: E1005 20:55:04.392580 4754 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T20:55:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T20:55:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T20:55:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T20:55:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ede2b442-76b0-443a-a450-252ee8711c80\\\",\\\"systemUUID\\\":\\\"72d081dd-fe7e-42a6-9cf0-f0ee7f733ac9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:04Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:04 crc kubenswrapper[4754]: E1005 20:55:04.392709 4754 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 05 20:55:04 crc kubenswrapper[4754]: I1005 20:55:04.394993 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:04 crc kubenswrapper[4754]: I1005 20:55:04.395055 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:04 crc kubenswrapper[4754]: I1005 20:55:04.395070 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:04 crc kubenswrapper[4754]: I1005 20:55:04.395090 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:04 crc kubenswrapper[4754]: I1005 20:55:04.395101 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:04Z","lastTransitionTime":"2025-10-05T20:55:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:04 crc kubenswrapper[4754]: I1005 20:55:04.394979 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cn76s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02d5e3f9-73c2-4496-9aca-0787184aef19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7ba2b076839f4a5f586d2e6c4f4d7188254128b782bfd2180e22bf3ea5ffdd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpx5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cn76s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:04Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:04 crc kubenswrapper[4754]: I1005 20:55:04.429458 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ss7lm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f9ef817-0622-4185-a7fb-57221690aa75\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ba613b0d7753a003c572c3733ed394707c45b6f8072cb7aa88f3c7312b8882b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wpwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:55:02Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ss7lm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:04Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:04 crc kubenswrapper[4754]: I1005 20:55:04.473395 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"934bcd66-9c03-48b9-82dd-5093338f9c8f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9042df2f43243d3d79ba9a3940f8dd4f6c7a89ef9e3510e81fe60b1944d7e83b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5366bbf46a91667355f43f5df7335a536b395cda788b90b35ef18e2b53bc4c02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://600bc3bb551b747cf2f18e26745883232d7a37ac5877fad0c67100e52d7a43aa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78f74982180048eb50b9be79fa72c04f62f925b54fe511920f37e6407f864631\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:04Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:04 crc kubenswrapper[4754]: I1005 20:55:04.498940 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:04 crc kubenswrapper[4754]: I1005 20:55:04.499025 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:04 crc kubenswrapper[4754]: I1005 20:55:04.499056 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:04 crc kubenswrapper[4754]: I1005 20:55:04.499094 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:04 crc kubenswrapper[4754]: I1005 20:55:04.499120 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:04Z","lastTransitionTime":"2025-10-05T20:55:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:04 crc kubenswrapper[4754]: I1005 20:55:04.512185 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:04Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:04 crc kubenswrapper[4754]: I1005 20:55:04.550782 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9ad070bb699b643d29b82bcc1126ce3b585f7492f5ae2817fa491b309b1691d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:04Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:04 crc kubenswrapper[4754]: I1005 20:55:04.593880 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be95a413-6a59-45b4-84b7-b43cae694a26\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10b211cf6e47dbd1ee42c82e49de0a13a45581b92a0c4f6bbba0326c7149e41d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2ggg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4bb8c5903e8b70dd19eebb111097d96f725e17de44e06c3e74c067376f6c7360\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2ggg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-b2h9k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:04Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:04 crc kubenswrapper[4754]: I1005 20:55:04.602024 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:04 crc kubenswrapper[4754]: I1005 20:55:04.602059 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:04 crc kubenswrapper[4754]: I1005 20:55:04.602069 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:04 crc kubenswrapper[4754]: I1005 20:55:04.602087 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:04 crc kubenswrapper[4754]: I1005 20:55:04.602102 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:04Z","lastTransitionTime":"2025-10-05T20:55:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:04 crc kubenswrapper[4754]: I1005 20:55:04.635114 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7506d496-65ff-41dc-8230-24c8ee653a2a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33da5a6bb3cd3b042c193bc571e6622f2f87f27e560cdd0bbb9be31e2010d707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df6d700a947d19c84b4f65fd2dd2f8d1f4f1f80fa9b4800c2acfa8dae3912639\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35ca8ee77a0e58cf969bd67b277a3392614c4e73b963e6f6002097709370715d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8907b9b69b249c34ec6db1e460d641b31711b75d6d31d4caafd867339e6571d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8907b9b69b249c34ec6db1e460d641b31711b75d6d31d4caafd867339e6571d0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1005 20:54:52.589115 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 20:54:52.590181 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1304624737/tls.crt::/tmp/serving-cert-1304624737/tls.key\\\\\\\"\\\\nI1005 20:54:58.070544 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1005 20:54:58.090352 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1005 20:54:58.090382 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1005 20:54:58.090412 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1005 20:54:58.090418 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1005 20:54:58.144656 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1005 20:54:58.144688 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 20:54:58.144695 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 20:54:58.144699 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1005 20:54:58.144702 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1005 20:54:58.144705 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1005 20:54:58.144708 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1005 20:54:58.145024 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1005 20:54:58.148112 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4efb34f3493cc8854bc4cff6b29f63973c52b754404d7a21daec4e1f612d327d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54c568b637f3737667433047b956c7eb22fdf0f9301acdfd3b49fb84c39de6ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54c568b637f3737667433047b956c7eb22fdf0f9301acdfd3b49fb84c39de6ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:04Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:04 crc kubenswrapper[4754]: I1005 20:55:04.673176 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c76e6e4740035a6ca54191a770c4516e28101aae69afbb89739174c2f076ef5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9eededc0cb1af94fb7b5daba08785f6b077a85e9cb5ddd8b9f397b658194f61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:04Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:04 crc kubenswrapper[4754]: I1005 20:55:04.705602 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:04 crc kubenswrapper[4754]: I1005 20:55:04.705643 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:04 crc kubenswrapper[4754]: I1005 20:55:04.705655 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:04 crc kubenswrapper[4754]: I1005 20:55:04.705677 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:04 crc kubenswrapper[4754]: I1005 20:55:04.705690 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:04Z","lastTransitionTime":"2025-10-05T20:55:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:04 crc kubenswrapper[4754]: I1005 20:55:04.725324 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tskt7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a68e9c89-465b-4c01-a470-e21a480ee465\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d14ec99610d2167b484bd7712af7bb6d993d26c74b2828da1c89314f16fda22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d14ec99610d2167b484bd7712af7bb6d993d26c74b2828da1c89314f16fda22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f6f08dd853b25881e2f3c66206aa80d0c493518284217780aa40ac4b69f0c0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f6f08dd853b25881e2f3c66206aa80d0c493518284217780aa40ac4b69f0c0b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://71026d00ec1a3d91b019deb3c9f43ee52fb05b8f2d6374e1a5b762c9e5c1b10d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71026d00ec1a3d91b019deb3c9f43ee52fb05b8f2d6374e1a5b762c9e5c1b10d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tskt7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:04Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:04 crc kubenswrapper[4754]: I1005 20:55:04.808456 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:04 crc kubenswrapper[4754]: I1005 20:55:04.808864 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:04 crc kubenswrapper[4754]: I1005 20:55:04.808874 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:04 crc kubenswrapper[4754]: I1005 20:55:04.808890 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:04 crc kubenswrapper[4754]: I1005 20:55:04.808901 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:04Z","lastTransitionTime":"2025-10-05T20:55:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:04 crc kubenswrapper[4754]: I1005 20:55:04.838868 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 20:55:04 crc kubenswrapper[4754]: I1005 20:55:04.838968 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 20:55:04 crc kubenswrapper[4754]: E1005 20:55:04.839055 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 20:55:04 crc kubenswrapper[4754]: E1005 20:55:04.839180 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 20:55:04 crc kubenswrapper[4754]: I1005 20:55:04.838857 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 20:55:04 crc kubenswrapper[4754]: E1005 20:55:04.846023 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 20:55:04 crc kubenswrapper[4754]: I1005 20:55:04.911566 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:04 crc kubenswrapper[4754]: I1005 20:55:04.911628 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:04 crc kubenswrapper[4754]: I1005 20:55:04.911649 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:04 crc kubenswrapper[4754]: I1005 20:55:04.911675 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:04 crc kubenswrapper[4754]: I1005 20:55:04.911695 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:04Z","lastTransitionTime":"2025-10-05T20:55:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:05 crc kubenswrapper[4754]: I1005 20:55:05.013964 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:05 crc kubenswrapper[4754]: I1005 20:55:05.014015 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:05 crc kubenswrapper[4754]: I1005 20:55:05.014027 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:05 crc kubenswrapper[4754]: I1005 20:55:05.014045 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:05 crc kubenswrapper[4754]: I1005 20:55:05.014060 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:05Z","lastTransitionTime":"2025-10-05T20:55:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:05 crc kubenswrapper[4754]: I1005 20:55:05.125989 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:05 crc kubenswrapper[4754]: I1005 20:55:05.126030 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:05 crc kubenswrapper[4754]: I1005 20:55:05.126022 4754 generic.go:334] "Generic (PLEG): container finished" podID="a68e9c89-465b-4c01-a470-e21a480ee465" containerID="5b9910d2470acb16ebb436d7030b3161061607cb2a5ef50d482d8d5c4f48a8ac" exitCode=0 Oct 05 20:55:05 crc kubenswrapper[4754]: I1005 20:55:05.126043 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:05 crc kubenswrapper[4754]: I1005 20:55:05.126139 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:05 crc kubenswrapper[4754]: I1005 20:55:05.126153 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:05Z","lastTransitionTime":"2025-10-05T20:55:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:05 crc kubenswrapper[4754]: I1005 20:55:05.126274 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-tskt7" event={"ID":"a68e9c89-465b-4c01-a470-e21a480ee465","Type":"ContainerDied","Data":"5b9910d2470acb16ebb436d7030b3161061607cb2a5ef50d482d8d5c4f48a8ac"} Oct 05 20:55:05 crc kubenswrapper[4754]: I1005 20:55:05.133281 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" event={"ID":"b44790fb-fecb-4ec3-9816-8e23dd72ab33","Type":"ContainerStarted","Data":"a2acf9637fa25e5d3ffeb05224c11218c47defc679568fb928a498869613de4e"} Oct 05 20:55:05 crc kubenswrapper[4754]: I1005 20:55:05.156590 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b44790fb-fecb-4ec3-9816-8e23dd72ab33\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35ebeb7a87d5d47b6567917d38eca38a3e622c497fba58971296c4609c287423\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35ebeb7a87d5d47b6567917d38eca38a3e622c497fba58971296c4609c287423\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7rhps\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:05Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:05 crc kubenswrapper[4754]: I1005 20:55:05.183260 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9309179-b92a-4fa4-9fba-747b8f911936\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d969d3d50d6ad280fc6654df9441bab238bb538f277ed6d1273779813ffbaaed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://800a731ed8ef57714b2318af14054ef23f0368936630887a093fc77167b33d12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8cd9b08e83dc174413cc37891c14562816787028471a4201680c86dd620b7a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93b729f3f7f993db0d6bc3ba60c04930aada3cef8272ffcc3c6a8a242a9b1475\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dde1519111fc3af5d14463fdb58d4f5c62a7045e7de98c60981c7e4da37b0e7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ea160a5b37aea7f2970ae4dd2614b09d3dcf5dcbd1fc103b9997f8ecc3a804c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ea160a5b37aea7f2970ae4dd2614b09d3dcf5dcbd1fc103b9997f8ecc3a804c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://17ae6aa6c53e7dd061f70b41428ef50bbc7ffcc65352fad1a4a9d3c69dbd2ba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17ae6aa6c53e7dd061f70b41428ef50bbc7ffcc65352fad1a4a9d3c69dbd2ba9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://ad4cd69df7608c7d87d839fe09d4557c674c7efbc4730f4e4d3daaf9a26f5aee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad4cd69df7608c7d87d839fe09d4557c674c7efbc4730f4e4d3daaf9a26f5aee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:36Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:05Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:05 crc kubenswrapper[4754]: I1005 20:55:05.203968 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:05Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:05 crc kubenswrapper[4754]: I1005 20:55:05.224946 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9ad070bb699b643d29b82bcc1126ce3b585f7492f5ae2817fa491b309b1691d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:05Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:05 crc kubenswrapper[4754]: I1005 20:55:05.233903 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:05 crc kubenswrapper[4754]: I1005 20:55:05.233954 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:05 crc kubenswrapper[4754]: I1005 20:55:05.233969 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:05 crc kubenswrapper[4754]: I1005 20:55:05.233993 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:05 crc kubenswrapper[4754]: I1005 20:55:05.234007 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:05Z","lastTransitionTime":"2025-10-05T20:55:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:05 crc kubenswrapper[4754]: I1005 20:55:05.241226 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-57sr4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e854c7b9-1229-4da4-80e9-fea6479459b8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceb577e1055d84007936a50d0b3d54b9ae0299a1fc1a6618b37163a6d3fbb9cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vw2ps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:58Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-57sr4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:05Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:05 crc kubenswrapper[4754]: I1005 20:55:05.262916 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cn76s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02d5e3f9-73c2-4496-9aca-0787184aef19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7ba2b076839f4a5f586d2e6c4f4d7188254128b782bfd2180e22bf3ea5ffdd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpx5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cn76s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:05Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:05 crc kubenswrapper[4754]: I1005 20:55:05.283920 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ss7lm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f9ef817-0622-4185-a7fb-57221690aa75\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ba613b0d7753a003c572c3733ed394707c45b6f8072cb7aa88f3c7312b8882b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wpwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:55:02Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ss7lm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:05Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:05 crc kubenswrapper[4754]: I1005 20:55:05.312133 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"934bcd66-9c03-48b9-82dd-5093338f9c8f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9042df2f43243d3d79ba9a3940f8dd4f6c7a89ef9e3510e81fe60b1944d7e83b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5366bbf46a91667355f43f5df7335a536b395cda788b90b35ef18e2b53bc4c02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://600bc3bb551b747cf2f18e26745883232d7a37ac5877fad0c67100e52d7a43aa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78f74982180048eb50b9be79fa72c04f62f925b54fe511920f37e6407f864631\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:05Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:05 crc kubenswrapper[4754]: I1005 20:55:05.341425 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:05Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:05 crc kubenswrapper[4754]: I1005 20:55:05.342285 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:05 crc kubenswrapper[4754]: I1005 20:55:05.342356 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:05 crc kubenswrapper[4754]: I1005 20:55:05.342377 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:05 crc kubenswrapper[4754]: I1005 20:55:05.342403 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:05 crc kubenswrapper[4754]: I1005 20:55:05.342421 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:05Z","lastTransitionTime":"2025-10-05T20:55:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:05 crc kubenswrapper[4754]: I1005 20:55:05.363109 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tskt7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a68e9c89-465b-4c01-a470-e21a480ee465\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d14ec99610d2167b484bd7712af7bb6d993d26c74b2828da1c89314f16fda22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d14ec99610d2167b484bd7712af7bb6d993d26c74b2828da1c89314f16fda22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f6f08dd853b25881e2f3c66206aa80d0c493518284217780aa40ac4b69f0c0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f6f08dd853b25881e2f3c66206aa80d0c493518284217780aa40ac4b69f0c0b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://71026d00ec1a3d91b019deb3c9f43ee52fb05b8f2d6374e1a5b762c9e5c1b10d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71026d00ec1a3d91b019deb3c9f43ee52fb05b8f2d6374e1a5b762c9e5c1b10d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b9910d2470acb16ebb436d7030b3161061607cb2a5ef50d482d8d5c4f48a8ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b9910d2470acb16ebb436d7030b3161061607cb2a5ef50d482d8d5c4f48a8ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tskt7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:05Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:05 crc kubenswrapper[4754]: I1005 20:55:05.378596 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be95a413-6a59-45b4-84b7-b43cae694a26\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10b211cf6e47dbd1ee42c82e49de0a13a45581b92a0c4f6bbba0326c7149e41d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2ggg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4bb8c5903e8b70dd19eebb111097d96f725e17de44e06c3e74c067376f6c7360\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2ggg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-b2h9k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:05Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:05 crc kubenswrapper[4754]: I1005 20:55:05.396416 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7506d496-65ff-41dc-8230-24c8ee653a2a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33da5a6bb3cd3b042c193bc571e6622f2f87f27e560cdd0bbb9be31e2010d707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df6d700a947d19c84b4f65fd2dd2f8d1f4f1f80fa9b4800c2acfa8dae3912639\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35ca8ee77a0e58cf969bd67b277a3392614c4e73b963e6f6002097709370715d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8907b9b69b249c34ec6db1e460d641b31711b75d6d31d4caafd867339e6571d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8907b9b69b249c34ec6db1e460d641b31711b75d6d31d4caafd867339e6571d0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1005 20:54:52.589115 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 20:54:52.590181 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1304624737/tls.crt::/tmp/serving-cert-1304624737/tls.key\\\\\\\"\\\\nI1005 20:54:58.070544 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1005 20:54:58.090352 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1005 20:54:58.090382 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1005 20:54:58.090412 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1005 20:54:58.090418 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1005 20:54:58.144656 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1005 20:54:58.144688 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 20:54:58.144695 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 20:54:58.144699 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1005 20:54:58.144702 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1005 20:54:58.144705 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1005 20:54:58.144708 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1005 20:54:58.145024 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1005 20:54:58.148112 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4efb34f3493cc8854bc4cff6b29f63973c52b754404d7a21daec4e1f612d327d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54c568b637f3737667433047b956c7eb22fdf0f9301acdfd3b49fb84c39de6ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54c568b637f3737667433047b956c7eb22fdf0f9301acdfd3b49fb84c39de6ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:05Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:05 crc kubenswrapper[4754]: I1005 20:55:05.415029 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c76e6e4740035a6ca54191a770c4516e28101aae69afbb89739174c2f076ef5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9eededc0cb1af94fb7b5daba08785f6b077a85e9cb5ddd8b9f397b658194f61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:05Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:05 crc kubenswrapper[4754]: I1005 20:55:05.430592 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e380f9b2a0e319dc325c54d1247a6db98346292548fa688cb41a2ca1b3048419\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:05Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:05 crc kubenswrapper[4754]: I1005 20:55:05.446046 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:05 crc kubenswrapper[4754]: I1005 20:55:05.446077 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:05 crc kubenswrapper[4754]: I1005 20:55:05.446086 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:05 crc kubenswrapper[4754]: I1005 20:55:05.446101 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:05 crc kubenswrapper[4754]: I1005 20:55:05.446111 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:05Z","lastTransitionTime":"2025-10-05T20:55:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:05 crc kubenswrapper[4754]: I1005 20:55:05.448463 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:05Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:05 crc kubenswrapper[4754]: I1005 20:55:05.550030 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:05 crc kubenswrapper[4754]: I1005 20:55:05.550086 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:05 crc kubenswrapper[4754]: I1005 20:55:05.550104 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:05 crc kubenswrapper[4754]: I1005 20:55:05.550130 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:05 crc kubenswrapper[4754]: I1005 20:55:05.550148 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:05Z","lastTransitionTime":"2025-10-05T20:55:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:05 crc kubenswrapper[4754]: I1005 20:55:05.653126 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:05 crc kubenswrapper[4754]: I1005 20:55:05.653169 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:05 crc kubenswrapper[4754]: I1005 20:55:05.653178 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:05 crc kubenswrapper[4754]: I1005 20:55:05.653195 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:05 crc kubenswrapper[4754]: I1005 20:55:05.653207 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:05Z","lastTransitionTime":"2025-10-05T20:55:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:05 crc kubenswrapper[4754]: I1005 20:55:05.756253 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:05 crc kubenswrapper[4754]: I1005 20:55:05.756287 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:05 crc kubenswrapper[4754]: I1005 20:55:05.756297 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:05 crc kubenswrapper[4754]: I1005 20:55:05.756313 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:05 crc kubenswrapper[4754]: I1005 20:55:05.756323 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:05Z","lastTransitionTime":"2025-10-05T20:55:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:05 crc kubenswrapper[4754]: I1005 20:55:05.859456 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:05 crc kubenswrapper[4754]: I1005 20:55:05.859567 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:05 crc kubenswrapper[4754]: I1005 20:55:05.859592 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:05 crc kubenswrapper[4754]: I1005 20:55:05.859621 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:05 crc kubenswrapper[4754]: I1005 20:55:05.859640 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:05Z","lastTransitionTime":"2025-10-05T20:55:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:05 crc kubenswrapper[4754]: I1005 20:55:05.962983 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:05 crc kubenswrapper[4754]: I1005 20:55:05.963033 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:05 crc kubenswrapper[4754]: I1005 20:55:05.963044 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:05 crc kubenswrapper[4754]: I1005 20:55:05.963062 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:05 crc kubenswrapper[4754]: I1005 20:55:05.963075 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:05Z","lastTransitionTime":"2025-10-05T20:55:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:06 crc kubenswrapper[4754]: I1005 20:55:06.065806 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:06 crc kubenswrapper[4754]: I1005 20:55:06.065875 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:06 crc kubenswrapper[4754]: I1005 20:55:06.065894 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:06 crc kubenswrapper[4754]: I1005 20:55:06.065924 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:06 crc kubenswrapper[4754]: I1005 20:55:06.065946 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:06Z","lastTransitionTime":"2025-10-05T20:55:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:06 crc kubenswrapper[4754]: I1005 20:55:06.140375 4754 generic.go:334] "Generic (PLEG): container finished" podID="a68e9c89-465b-4c01-a470-e21a480ee465" containerID="b0a93b61eacc55111995e69c15f89a210b8aa15468ac9ac053b7cf6b007d301e" exitCode=0 Oct 05 20:55:06 crc kubenswrapper[4754]: I1005 20:55:06.140438 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-tskt7" event={"ID":"a68e9c89-465b-4c01-a470-e21a480ee465","Type":"ContainerDied","Data":"b0a93b61eacc55111995e69c15f89a210b8aa15468ac9ac053b7cf6b007d301e"} Oct 05 20:55:06 crc kubenswrapper[4754]: I1005 20:55:06.163211 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e380f9b2a0e319dc325c54d1247a6db98346292548fa688cb41a2ca1b3048419\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:06Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:06 crc kubenswrapper[4754]: I1005 20:55:06.180764 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:06 crc kubenswrapper[4754]: I1005 20:55:06.180809 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:06 crc kubenswrapper[4754]: I1005 20:55:06.180823 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:06 crc kubenswrapper[4754]: I1005 20:55:06.180842 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:06 crc kubenswrapper[4754]: I1005 20:55:06.180855 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:06Z","lastTransitionTime":"2025-10-05T20:55:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:06 crc kubenswrapper[4754]: I1005 20:55:06.190085 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:06Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:06 crc kubenswrapper[4754]: I1005 20:55:06.213938 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9309179-b92a-4fa4-9fba-747b8f911936\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d969d3d50d6ad280fc6654df9441bab238bb538f277ed6d1273779813ffbaaed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://800a731ed8ef57714b2318af14054ef23f0368936630887a093fc77167b33d12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8cd9b08e83dc174413cc37891c14562816787028471a4201680c86dd620b7a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93b729f3f7f993db0d6bc3ba60c04930aada3cef8272ffcc3c6a8a242a9b1475\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dde1519111fc3af5d14463fdb58d4f5c62a7045e7de98c60981c7e4da37b0e7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ea160a5b37aea7f2970ae4dd2614b09d3dcf5dcbd1fc103b9997f8ecc3a804c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ea160a5b37aea7f2970ae4dd2614b09d3dcf5dcbd1fc103b9997f8ecc3a804c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://17ae6aa6c53e7dd061f70b41428ef50bbc7ffcc65352fad1a4a9d3c69dbd2ba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17ae6aa6c53e7dd061f70b41428ef50bbc7ffcc65352fad1a4a9d3c69dbd2ba9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://ad4cd69df7608c7d87d839fe09d4557c674c7efbc4730f4e4d3daaf9a26f5aee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad4cd69df7608c7d87d839fe09d4557c674c7efbc4730f4e4d3daaf9a26f5aee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:36Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:06Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:06 crc kubenswrapper[4754]: I1005 20:55:06.243237 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:06Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:06 crc kubenswrapper[4754]: I1005 20:55:06.277186 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b44790fb-fecb-4ec3-9816-8e23dd72ab33\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35ebeb7a87d5d47b6567917d38eca38a3e622c497fba58971296c4609c287423\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35ebeb7a87d5d47b6567917d38eca38a3e622c497fba58971296c4609c287423\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7rhps\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:06Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:06 crc kubenswrapper[4754]: I1005 20:55:06.283657 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:06 crc kubenswrapper[4754]: I1005 20:55:06.283708 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:06 crc kubenswrapper[4754]: I1005 20:55:06.283723 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:06 crc kubenswrapper[4754]: I1005 20:55:06.283742 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:06 crc kubenswrapper[4754]: I1005 20:55:06.283757 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:06Z","lastTransitionTime":"2025-10-05T20:55:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:06 crc kubenswrapper[4754]: I1005 20:55:06.292272 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"934bcd66-9c03-48b9-82dd-5093338f9c8f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9042df2f43243d3d79ba9a3940f8dd4f6c7a89ef9e3510e81fe60b1944d7e83b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5366bbf46a91667355f43f5df7335a536b395cda788b90b35ef18e2b53bc4c02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://600bc3bb551b747cf2f18e26745883232d7a37ac5877fad0c67100e52d7a43aa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78f74982180048eb50b9be79fa72c04f62f925b54fe511920f37e6407f864631\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:06Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:06 crc kubenswrapper[4754]: I1005 20:55:06.305373 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:06Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:06 crc kubenswrapper[4754]: I1005 20:55:06.318364 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9ad070bb699b643d29b82bcc1126ce3b585f7492f5ae2817fa491b309b1691d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:06Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:06 crc kubenswrapper[4754]: I1005 20:55:06.328986 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-57sr4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e854c7b9-1229-4da4-80e9-fea6479459b8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceb577e1055d84007936a50d0b3d54b9ae0299a1fc1a6618b37163a6d3fbb9cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vw2ps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:58Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-57sr4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:06Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:06 crc kubenswrapper[4754]: I1005 20:55:06.342425 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cn76s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02d5e3f9-73c2-4496-9aca-0787184aef19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7ba2b076839f4a5f586d2e6c4f4d7188254128b782bfd2180e22bf3ea5ffdd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpx5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cn76s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:06Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:06 crc kubenswrapper[4754]: I1005 20:55:06.355521 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ss7lm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f9ef817-0622-4185-a7fb-57221690aa75\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ba613b0d7753a003c572c3733ed394707c45b6f8072cb7aa88f3c7312b8882b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wpwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:55:02Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ss7lm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:06Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:06 crc kubenswrapper[4754]: I1005 20:55:06.374415 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7506d496-65ff-41dc-8230-24c8ee653a2a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33da5a6bb3cd3b042c193bc571e6622f2f87f27e560cdd0bbb9be31e2010d707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df6d700a947d19c84b4f65fd2dd2f8d1f4f1f80fa9b4800c2acfa8dae3912639\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35ca8ee77a0e58cf969bd67b277a3392614c4e73b963e6f6002097709370715d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8907b9b69b249c34ec6db1e460d641b31711b75d6d31d4caafd867339e6571d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8907b9b69b249c34ec6db1e460d641b31711b75d6d31d4caafd867339e6571d0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1005 20:54:52.589115 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 20:54:52.590181 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1304624737/tls.crt::/tmp/serving-cert-1304624737/tls.key\\\\\\\"\\\\nI1005 20:54:58.070544 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1005 20:54:58.090352 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1005 20:54:58.090382 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1005 20:54:58.090412 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1005 20:54:58.090418 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1005 20:54:58.144656 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1005 20:54:58.144688 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 20:54:58.144695 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 20:54:58.144699 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1005 20:54:58.144702 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1005 20:54:58.144705 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1005 20:54:58.144708 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1005 20:54:58.145024 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1005 20:54:58.148112 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4efb34f3493cc8854bc4cff6b29f63973c52b754404d7a21daec4e1f612d327d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54c568b637f3737667433047b956c7eb22fdf0f9301acdfd3b49fb84c39de6ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54c568b637f3737667433047b956c7eb22fdf0f9301acdfd3b49fb84c39de6ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:06Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:06 crc kubenswrapper[4754]: I1005 20:55:06.386039 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:06 crc kubenswrapper[4754]: I1005 20:55:06.386312 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:06 crc kubenswrapper[4754]: I1005 20:55:06.386321 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:06 crc kubenswrapper[4754]: I1005 20:55:06.386339 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:06 crc kubenswrapper[4754]: I1005 20:55:06.386352 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:06Z","lastTransitionTime":"2025-10-05T20:55:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:06 crc kubenswrapper[4754]: I1005 20:55:06.389338 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c76e6e4740035a6ca54191a770c4516e28101aae69afbb89739174c2f076ef5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9eededc0cb1af94fb7b5daba08785f6b077a85e9cb5ddd8b9f397b658194f61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:06Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:06 crc kubenswrapper[4754]: I1005 20:55:06.405594 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tskt7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a68e9c89-465b-4c01-a470-e21a480ee465\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d14ec99610d2167b484bd7712af7bb6d993d26c74b2828da1c89314f16fda22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d14ec99610d2167b484bd7712af7bb6d993d26c74b2828da1c89314f16fda22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f6f08dd853b25881e2f3c66206aa80d0c493518284217780aa40ac4b69f0c0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f6f08dd853b25881e2f3c66206aa80d0c493518284217780aa40ac4b69f0c0b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://71026d00ec1a3d91b019deb3c9f43ee52fb05b8f2d6374e1a5b762c9e5c1b10d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71026d00ec1a3d91b019deb3c9f43ee52fb05b8f2d6374e1a5b762c9e5c1b10d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b9910d2470acb16ebb436d7030b3161061607cb2a5ef50d482d8d5c4f48a8ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b9910d2470acb16ebb436d7030b3161061607cb2a5ef50d482d8d5c4f48a8ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0a93b61eacc55111995e69c15f89a210b8aa15468ac9ac053b7cf6b007d301e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b0a93b61eacc55111995e69c15f89a210b8aa15468ac9ac053b7cf6b007d301e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tskt7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:06Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:06 crc kubenswrapper[4754]: I1005 20:55:06.416635 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be95a413-6a59-45b4-84b7-b43cae694a26\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10b211cf6e47dbd1ee42c82e49de0a13a45581b92a0c4f6bbba0326c7149e41d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2ggg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4bb8c5903e8b70dd19eebb111097d96f725e17de44e06c3e74c067376f6c7360\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2ggg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-b2h9k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:06Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:06 crc kubenswrapper[4754]: I1005 20:55:06.481095 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 20:55:06 crc kubenswrapper[4754]: E1005 20:55:06.481297 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 20:55:14.481261395 +0000 UTC m=+38.385380105 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:55:06 crc kubenswrapper[4754]: I1005 20:55:06.489258 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:06 crc kubenswrapper[4754]: I1005 20:55:06.489292 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:06 crc kubenswrapper[4754]: I1005 20:55:06.489303 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:06 crc kubenswrapper[4754]: I1005 20:55:06.489325 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:06 crc kubenswrapper[4754]: I1005 20:55:06.489337 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:06Z","lastTransitionTime":"2025-10-05T20:55:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:06 crc kubenswrapper[4754]: I1005 20:55:06.582124 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 20:55:06 crc kubenswrapper[4754]: I1005 20:55:06.582186 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 20:55:06 crc kubenswrapper[4754]: I1005 20:55:06.582223 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 20:55:06 crc kubenswrapper[4754]: I1005 20:55:06.582267 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 20:55:06 crc kubenswrapper[4754]: E1005 20:55:06.582352 4754 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 05 20:55:06 crc kubenswrapper[4754]: E1005 20:55:06.582376 4754 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 05 20:55:06 crc kubenswrapper[4754]: E1005 20:55:06.582428 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-05 20:55:14.582407068 +0000 UTC m=+38.486525798 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 05 20:55:06 crc kubenswrapper[4754]: E1005 20:55:06.582479 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-05 20:55:14.582464749 +0000 UTC m=+38.486583449 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 05 20:55:06 crc kubenswrapper[4754]: E1005 20:55:06.582445 4754 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 05 20:55:06 crc kubenswrapper[4754]: E1005 20:55:06.582540 4754 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 05 20:55:06 crc kubenswrapper[4754]: E1005 20:55:06.582556 4754 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 05 20:55:06 crc kubenswrapper[4754]: E1005 20:55:06.582596 4754 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 05 20:55:06 crc kubenswrapper[4754]: E1005 20:55:06.582644 4754 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 05 20:55:06 crc kubenswrapper[4754]: E1005 20:55:06.582670 4754 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 05 20:55:06 crc kubenswrapper[4754]: E1005 20:55:06.582614 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-05 20:55:14.582605583 +0000 UTC m=+38.486724293 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 05 20:55:06 crc kubenswrapper[4754]: E1005 20:55:06.582757 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-05 20:55:14.582728936 +0000 UTC m=+38.486847686 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 05 20:55:06 crc kubenswrapper[4754]: I1005 20:55:06.593482 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:06 crc kubenswrapper[4754]: I1005 20:55:06.593567 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:06 crc kubenswrapper[4754]: I1005 20:55:06.593584 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:06 crc kubenswrapper[4754]: I1005 20:55:06.593611 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:06 crc kubenswrapper[4754]: I1005 20:55:06.593632 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:06Z","lastTransitionTime":"2025-10-05T20:55:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:06 crc kubenswrapper[4754]: I1005 20:55:06.696304 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:06 crc kubenswrapper[4754]: I1005 20:55:06.696357 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:06 crc kubenswrapper[4754]: I1005 20:55:06.696367 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:06 crc kubenswrapper[4754]: I1005 20:55:06.696390 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:06 crc kubenswrapper[4754]: I1005 20:55:06.696405 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:06Z","lastTransitionTime":"2025-10-05T20:55:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:06 crc kubenswrapper[4754]: I1005 20:55:06.799963 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:06 crc kubenswrapper[4754]: I1005 20:55:06.800058 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:06 crc kubenswrapper[4754]: I1005 20:55:06.800082 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:06 crc kubenswrapper[4754]: I1005 20:55:06.800108 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:06 crc kubenswrapper[4754]: I1005 20:55:06.800126 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:06Z","lastTransitionTime":"2025-10-05T20:55:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:06 crc kubenswrapper[4754]: I1005 20:55:06.836545 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 20:55:06 crc kubenswrapper[4754]: E1005 20:55:06.836751 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 20:55:06 crc kubenswrapper[4754]: I1005 20:55:06.837692 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 20:55:06 crc kubenswrapper[4754]: E1005 20:55:06.837840 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 20:55:06 crc kubenswrapper[4754]: I1005 20:55:06.838087 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 20:55:06 crc kubenswrapper[4754]: E1005 20:55:06.838235 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 20:55:06 crc kubenswrapper[4754]: I1005 20:55:06.861984 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9ad070bb699b643d29b82bcc1126ce3b585f7492f5ae2817fa491b309b1691d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:06Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:06 crc kubenswrapper[4754]: I1005 20:55:06.872523 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-57sr4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e854c7b9-1229-4da4-80e9-fea6479459b8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceb577e1055d84007936a50d0b3d54b9ae0299a1fc1a6618b37163a6d3fbb9cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vw2ps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:58Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-57sr4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:06Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:06 crc kubenswrapper[4754]: I1005 20:55:06.884537 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cn76s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02d5e3f9-73c2-4496-9aca-0787184aef19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7ba2b076839f4a5f586d2e6c4f4d7188254128b782bfd2180e22bf3ea5ffdd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpx5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cn76s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:06Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:06 crc kubenswrapper[4754]: I1005 20:55:06.894004 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ss7lm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f9ef817-0622-4185-a7fb-57221690aa75\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ba613b0d7753a003c572c3733ed394707c45b6f8072cb7aa88f3c7312b8882b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wpwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:55:02Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ss7lm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:06Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:06 crc kubenswrapper[4754]: I1005 20:55:06.903154 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:06 crc kubenswrapper[4754]: I1005 20:55:06.903189 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:06 crc kubenswrapper[4754]: I1005 20:55:06.903201 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:06 crc kubenswrapper[4754]: I1005 20:55:06.903226 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:06 crc kubenswrapper[4754]: I1005 20:55:06.903240 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:06Z","lastTransitionTime":"2025-10-05T20:55:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:06 crc kubenswrapper[4754]: I1005 20:55:06.914909 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"934bcd66-9c03-48b9-82dd-5093338f9c8f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9042df2f43243d3d79ba9a3940f8dd4f6c7a89ef9e3510e81fe60b1944d7e83b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5366bbf46a91667355f43f5df7335a536b395cda788b90b35ef18e2b53bc4c02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://600bc3bb551b747cf2f18e26745883232d7a37ac5877fad0c67100e52d7a43aa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78f74982180048eb50b9be79fa72c04f62f925b54fe511920f37e6407f864631\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:06Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:06 crc kubenswrapper[4754]: I1005 20:55:06.931901 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:06Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:06 crc kubenswrapper[4754]: I1005 20:55:06.950815 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tskt7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a68e9c89-465b-4c01-a470-e21a480ee465\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d14ec99610d2167b484bd7712af7bb6d993d26c74b2828da1c89314f16fda22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d14ec99610d2167b484bd7712af7bb6d993d26c74b2828da1c89314f16fda22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f6f08dd853b25881e2f3c66206aa80d0c493518284217780aa40ac4b69f0c0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f6f08dd853b25881e2f3c66206aa80d0c493518284217780aa40ac4b69f0c0b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://71026d00ec1a3d91b019deb3c9f43ee52fb05b8f2d6374e1a5b762c9e5c1b10d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71026d00ec1a3d91b019deb3c9f43ee52fb05b8f2d6374e1a5b762c9e5c1b10d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b9910d2470acb16ebb436d7030b3161061607cb2a5ef50d482d8d5c4f48a8ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b9910d2470acb16ebb436d7030b3161061607cb2a5ef50d482d8d5c4f48a8ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0a93b61eacc55111995e69c15f89a210b8aa15468ac9ac053b7cf6b007d301e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b0a93b61eacc55111995e69c15f89a210b8aa15468ac9ac053b7cf6b007d301e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tskt7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:06Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:06 crc kubenswrapper[4754]: I1005 20:55:06.968242 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be95a413-6a59-45b4-84b7-b43cae694a26\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10b211cf6e47dbd1ee42c82e49de0a13a45581b92a0c4f6bbba0326c7149e41d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2ggg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4bb8c5903e8b70dd19eebb111097d96f725e17de44e06c3e74c067376f6c7360\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2ggg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-b2h9k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:06Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:06 crc kubenswrapper[4754]: I1005 20:55:06.991015 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7506d496-65ff-41dc-8230-24c8ee653a2a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33da5a6bb3cd3b042c193bc571e6622f2f87f27e560cdd0bbb9be31e2010d707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df6d700a947d19c84b4f65fd2dd2f8d1f4f1f80fa9b4800c2acfa8dae3912639\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35ca8ee77a0e58cf969bd67b277a3392614c4e73b963e6f6002097709370715d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8907b9b69b249c34ec6db1e460d641b31711b75d6d31d4caafd867339e6571d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8907b9b69b249c34ec6db1e460d641b31711b75d6d31d4caafd867339e6571d0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1005 20:54:52.589115 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 20:54:52.590181 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1304624737/tls.crt::/tmp/serving-cert-1304624737/tls.key\\\\\\\"\\\\nI1005 20:54:58.070544 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1005 20:54:58.090352 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1005 20:54:58.090382 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1005 20:54:58.090412 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1005 20:54:58.090418 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1005 20:54:58.144656 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1005 20:54:58.144688 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 20:54:58.144695 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 20:54:58.144699 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1005 20:54:58.144702 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1005 20:54:58.144705 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1005 20:54:58.144708 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1005 20:54:58.145024 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1005 20:54:58.148112 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4efb34f3493cc8854bc4cff6b29f63973c52b754404d7a21daec4e1f612d327d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54c568b637f3737667433047b956c7eb22fdf0f9301acdfd3b49fb84c39de6ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54c568b637f3737667433047b956c7eb22fdf0f9301acdfd3b49fb84c39de6ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:06Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:07 crc kubenswrapper[4754]: I1005 20:55:07.006313 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:07 crc kubenswrapper[4754]: I1005 20:55:07.006356 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:07 crc kubenswrapper[4754]: I1005 20:55:07.006369 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:07 crc kubenswrapper[4754]: I1005 20:55:07.006389 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:07 crc kubenswrapper[4754]: I1005 20:55:07.006405 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:07Z","lastTransitionTime":"2025-10-05T20:55:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:07 crc kubenswrapper[4754]: I1005 20:55:07.008053 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c76e6e4740035a6ca54191a770c4516e28101aae69afbb89739174c2f076ef5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9eededc0cb1af94fb7b5daba08785f6b077a85e9cb5ddd8b9f397b658194f61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:07Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:07 crc kubenswrapper[4754]: I1005 20:55:07.028232 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e380f9b2a0e319dc325c54d1247a6db98346292548fa688cb41a2ca1b3048419\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:07Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:07 crc kubenswrapper[4754]: I1005 20:55:07.049108 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:07Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:07 crc kubenswrapper[4754]: I1005 20:55:07.069476 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b44790fb-fecb-4ec3-9816-8e23dd72ab33\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35ebeb7a87d5d47b6567917d38eca38a3e622c497fba58971296c4609c287423\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35ebeb7a87d5d47b6567917d38eca38a3e622c497fba58971296c4609c287423\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7rhps\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:07Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:07 crc kubenswrapper[4754]: I1005 20:55:07.089865 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9309179-b92a-4fa4-9fba-747b8f911936\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d969d3d50d6ad280fc6654df9441bab238bb538f277ed6d1273779813ffbaaed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://800a731ed8ef57714b2318af14054ef23f0368936630887a093fc77167b33d12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8cd9b08e83dc174413cc37891c14562816787028471a4201680c86dd620b7a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93b729f3f7f993db0d6bc3ba60c04930aada3cef8272ffcc3c6a8a242a9b1475\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dde1519111fc3af5d14463fdb58d4f5c62a7045e7de98c60981c7e4da37b0e7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ea160a5b37aea7f2970ae4dd2614b09d3dcf5dcbd1fc103b9997f8ecc3a804c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ea160a5b37aea7f2970ae4dd2614b09d3dcf5dcbd1fc103b9997f8ecc3a804c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://17ae6aa6c53e7dd061f70b41428ef50bbc7ffcc65352fad1a4a9d3c69dbd2ba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17ae6aa6c53e7dd061f70b41428ef50bbc7ffcc65352fad1a4a9d3c69dbd2ba9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://ad4cd69df7608c7d87d839fe09d4557c674c7efbc4730f4e4d3daaf9a26f5aee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad4cd69df7608c7d87d839fe09d4557c674c7efbc4730f4e4d3daaf9a26f5aee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:36Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:07Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:07 crc kubenswrapper[4754]: I1005 20:55:07.108207 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:07Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:07 crc kubenswrapper[4754]: I1005 20:55:07.109555 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:07 crc kubenswrapper[4754]: I1005 20:55:07.109713 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:07 crc kubenswrapper[4754]: I1005 20:55:07.109818 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:07 crc kubenswrapper[4754]: I1005 20:55:07.109906 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:07 crc kubenswrapper[4754]: I1005 20:55:07.109989 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:07Z","lastTransitionTime":"2025-10-05T20:55:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:07 crc kubenswrapper[4754]: I1005 20:55:07.149783 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" event={"ID":"b44790fb-fecb-4ec3-9816-8e23dd72ab33","Type":"ContainerStarted","Data":"c154d33bd80e9c252f693ffe0ec79b7bb6dbf1563c76343e8b3aba65f53b24b7"} Oct 05 20:55:07 crc kubenswrapper[4754]: I1005 20:55:07.150244 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" Oct 05 20:55:07 crc kubenswrapper[4754]: I1005 20:55:07.158238 4754 generic.go:334] "Generic (PLEG): container finished" podID="a68e9c89-465b-4c01-a470-e21a480ee465" containerID="cafb3208e7e3bcdd01898a68fc5bbf4cb52ca0800b879b77a60bb102794257f2" exitCode=0 Oct 05 20:55:07 crc kubenswrapper[4754]: I1005 20:55:07.158304 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-tskt7" event={"ID":"a68e9c89-465b-4c01-a470-e21a480ee465","Type":"ContainerDied","Data":"cafb3208e7e3bcdd01898a68fc5bbf4cb52ca0800b879b77a60bb102794257f2"} Oct 05 20:55:07 crc kubenswrapper[4754]: I1005 20:55:07.193996 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" Oct 05 20:55:07 crc kubenswrapper[4754]: I1005 20:55:07.196330 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9309179-b92a-4fa4-9fba-747b8f911936\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d969d3d50d6ad280fc6654df9441bab238bb538f277ed6d1273779813ffbaaed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://800a731ed8ef57714b2318af14054ef23f0368936630887a093fc77167b33d12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8cd9b08e83dc174413cc37891c14562816787028471a4201680c86dd620b7a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93b729f3f7f993db0d6bc3ba60c04930aada3cef8272ffcc3c6a8a242a9b1475\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dde1519111fc3af5d14463fdb58d4f5c62a7045e7de98c60981c7e4da37b0e7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ea160a5b37aea7f2970ae4dd2614b09d3dcf5dcbd1fc103b9997f8ecc3a804c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ea160a5b37aea7f2970ae4dd2614b09d3dcf5dcbd1fc103b9997f8ecc3a804c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://17ae6aa6c53e7dd061f70b41428ef50bbc7ffcc65352fad1a4a9d3c69dbd2ba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17ae6aa6c53e7dd061f70b41428ef50bbc7ffcc65352fad1a4a9d3c69dbd2ba9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://ad4cd69df7608c7d87d839fe09d4557c674c7efbc4730f4e4d3daaf9a26f5aee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad4cd69df7608c7d87d839fe09d4557c674c7efbc4730f4e4d3daaf9a26f5aee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:36Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:07Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:07 crc kubenswrapper[4754]: I1005 20:55:07.213668 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:07 crc kubenswrapper[4754]: I1005 20:55:07.213724 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:07 crc kubenswrapper[4754]: I1005 20:55:07.213744 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:07 crc kubenswrapper[4754]: I1005 20:55:07.213771 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:07 crc kubenswrapper[4754]: I1005 20:55:07.213792 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:07Z","lastTransitionTime":"2025-10-05T20:55:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:07 crc kubenswrapper[4754]: I1005 20:55:07.217659 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:07Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:07 crc kubenswrapper[4754]: I1005 20:55:07.245781 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b44790fb-fecb-4ec3-9816-8e23dd72ab33\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a7dc53e7f96e48213d724a2cb4859514df925005644108ae3a0c33b14dcf52d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfdd8b1e542ab676576c311e2923fb02861f92de91fe1f4b3e4b968c549c375d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3a63451ffaa14cacc3b78d11340f4ea81445fa3af96178e9b172d063d9697d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3cfd3f2274a8a46fa36660e94ac6c9c298c791283f12d52330c5594ac131998\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f626cb108716da0b595844ea0251b7a04d49b5908a1d6dd0aef5cb2d402c4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://822223f515b5add6d71d193fd5bc77df029b8d37a1cc07be4145b53c7226bf61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c154d33bd80e9c252f693ffe0ec79b7bb6dbf1563c76343e8b3aba65f53b24b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2acf9637fa25e5d3ffeb05224c11218c47defc679568fb928a498869613de4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35ebeb7a87d5d47b6567917d38eca38a3e622c497fba58971296c4609c287423\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35ebeb7a87d5d47b6567917d38eca38a3e622c497fba58971296c4609c287423\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7rhps\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:07Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:07 crc kubenswrapper[4754]: I1005 20:55:07.263810 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cn76s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02d5e3f9-73c2-4496-9aca-0787184aef19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7ba2b076839f4a5f586d2e6c4f4d7188254128b782bfd2180e22bf3ea5ffdd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpx5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cn76s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:07Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:07 crc kubenswrapper[4754]: I1005 20:55:07.274468 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ss7lm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f9ef817-0622-4185-a7fb-57221690aa75\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ba613b0d7753a003c572c3733ed394707c45b6f8072cb7aa88f3c7312b8882b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wpwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:55:02Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ss7lm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:07Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:07 crc kubenswrapper[4754]: I1005 20:55:07.288611 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"934bcd66-9c03-48b9-82dd-5093338f9c8f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9042df2f43243d3d79ba9a3940f8dd4f6c7a89ef9e3510e81fe60b1944d7e83b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5366bbf46a91667355f43f5df7335a536b395cda788b90b35ef18e2b53bc4c02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://600bc3bb551b747cf2f18e26745883232d7a37ac5877fad0c67100e52d7a43aa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78f74982180048eb50b9be79fa72c04f62f925b54fe511920f37e6407f864631\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:07Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:07 crc kubenswrapper[4754]: I1005 20:55:07.303554 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:07Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:07 crc kubenswrapper[4754]: I1005 20:55:07.318692 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:07 crc kubenswrapper[4754]: I1005 20:55:07.318759 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:07 crc kubenswrapper[4754]: I1005 20:55:07.318800 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:07 crc kubenswrapper[4754]: I1005 20:55:07.318865 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:07 crc kubenswrapper[4754]: I1005 20:55:07.318881 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:07Z","lastTransitionTime":"2025-10-05T20:55:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:07 crc kubenswrapper[4754]: I1005 20:55:07.319857 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9ad070bb699b643d29b82bcc1126ce3b585f7492f5ae2817fa491b309b1691d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:07Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:07 crc kubenswrapper[4754]: I1005 20:55:07.331387 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-57sr4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e854c7b9-1229-4da4-80e9-fea6479459b8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceb577e1055d84007936a50d0b3d54b9ae0299a1fc1a6618b37163a6d3fbb9cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vw2ps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:58Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-57sr4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:07Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:07 crc kubenswrapper[4754]: I1005 20:55:07.353878 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7506d496-65ff-41dc-8230-24c8ee653a2a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33da5a6bb3cd3b042c193bc571e6622f2f87f27e560cdd0bbb9be31e2010d707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df6d700a947d19c84b4f65fd2dd2f8d1f4f1f80fa9b4800c2acfa8dae3912639\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35ca8ee77a0e58cf969bd67b277a3392614c4e73b963e6f6002097709370715d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8907b9b69b249c34ec6db1e460d641b31711b75d6d31d4caafd867339e6571d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8907b9b69b249c34ec6db1e460d641b31711b75d6d31d4caafd867339e6571d0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1005 20:54:52.589115 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 20:54:52.590181 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1304624737/tls.crt::/tmp/serving-cert-1304624737/tls.key\\\\\\\"\\\\nI1005 20:54:58.070544 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1005 20:54:58.090352 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1005 20:54:58.090382 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1005 20:54:58.090412 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1005 20:54:58.090418 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1005 20:54:58.144656 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1005 20:54:58.144688 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 20:54:58.144695 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 20:54:58.144699 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1005 20:54:58.144702 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1005 20:54:58.144705 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1005 20:54:58.144708 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1005 20:54:58.145024 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1005 20:54:58.148112 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4efb34f3493cc8854bc4cff6b29f63973c52b754404d7a21daec4e1f612d327d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54c568b637f3737667433047b956c7eb22fdf0f9301acdfd3b49fb84c39de6ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54c568b637f3737667433047b956c7eb22fdf0f9301acdfd3b49fb84c39de6ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:07Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:07 crc kubenswrapper[4754]: I1005 20:55:07.371048 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c76e6e4740035a6ca54191a770c4516e28101aae69afbb89739174c2f076ef5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9eededc0cb1af94fb7b5daba08785f6b077a85e9cb5ddd8b9f397b658194f61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:07Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:07 crc kubenswrapper[4754]: I1005 20:55:07.391537 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tskt7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a68e9c89-465b-4c01-a470-e21a480ee465\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d14ec99610d2167b484bd7712af7bb6d993d26c74b2828da1c89314f16fda22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d14ec99610d2167b484bd7712af7bb6d993d26c74b2828da1c89314f16fda22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f6f08dd853b25881e2f3c66206aa80d0c493518284217780aa40ac4b69f0c0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f6f08dd853b25881e2f3c66206aa80d0c493518284217780aa40ac4b69f0c0b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://71026d00ec1a3d91b019deb3c9f43ee52fb05b8f2d6374e1a5b762c9e5c1b10d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71026d00ec1a3d91b019deb3c9f43ee52fb05b8f2d6374e1a5b762c9e5c1b10d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b9910d2470acb16ebb436d7030b3161061607cb2a5ef50d482d8d5c4f48a8ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b9910d2470acb16ebb436d7030b3161061607cb2a5ef50d482d8d5c4f48a8ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0a93b61eacc55111995e69c15f89a210b8aa15468ac9ac053b7cf6b007d301e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b0a93b61eacc55111995e69c15f89a210b8aa15468ac9ac053b7cf6b007d301e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tskt7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:07Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:07 crc kubenswrapper[4754]: I1005 20:55:07.405961 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be95a413-6a59-45b4-84b7-b43cae694a26\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10b211cf6e47dbd1ee42c82e49de0a13a45581b92a0c4f6bbba0326c7149e41d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2ggg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4bb8c5903e8b70dd19eebb111097d96f725e17de44e06c3e74c067376f6c7360\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2ggg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-b2h9k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:07Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:07 crc kubenswrapper[4754]: I1005 20:55:07.431041 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e380f9b2a0e319dc325c54d1247a6db98346292548fa688cb41a2ca1b3048419\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:07Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:07 crc kubenswrapper[4754]: I1005 20:55:07.431446 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:07 crc kubenswrapper[4754]: I1005 20:55:07.431474 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:07 crc kubenswrapper[4754]: I1005 20:55:07.431484 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:07 crc kubenswrapper[4754]: I1005 20:55:07.431515 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:07 crc kubenswrapper[4754]: I1005 20:55:07.431527 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:07Z","lastTransitionTime":"2025-10-05T20:55:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:07 crc kubenswrapper[4754]: I1005 20:55:07.447006 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:07Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:07 crc kubenswrapper[4754]: I1005 20:55:07.459648 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:07Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:07 crc kubenswrapper[4754]: I1005 20:55:07.473450 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e380f9b2a0e319dc325c54d1247a6db98346292548fa688cb41a2ca1b3048419\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:07Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:07 crc kubenswrapper[4754]: I1005 20:55:07.491631 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9309179-b92a-4fa4-9fba-747b8f911936\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d969d3d50d6ad280fc6654df9441bab238bb538f277ed6d1273779813ffbaaed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://800a731ed8ef57714b2318af14054ef23f0368936630887a093fc77167b33d12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8cd9b08e83dc174413cc37891c14562816787028471a4201680c86dd620b7a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93b729f3f7f993db0d6bc3ba60c04930aada3cef8272ffcc3c6a8a242a9b1475\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dde1519111fc3af5d14463fdb58d4f5c62a7045e7de98c60981c7e4da37b0e7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ea160a5b37aea7f2970ae4dd2614b09d3dcf5dcbd1fc103b9997f8ecc3a804c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ea160a5b37aea7f2970ae4dd2614b09d3dcf5dcbd1fc103b9997f8ecc3a804c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://17ae6aa6c53e7dd061f70b41428ef50bbc7ffcc65352fad1a4a9d3c69dbd2ba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17ae6aa6c53e7dd061f70b41428ef50bbc7ffcc65352fad1a4a9d3c69dbd2ba9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://ad4cd69df7608c7d87d839fe09d4557c674c7efbc4730f4e4d3daaf9a26f5aee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad4cd69df7608c7d87d839fe09d4557c674c7efbc4730f4e4d3daaf9a26f5aee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:36Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:07Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:07 crc kubenswrapper[4754]: I1005 20:55:07.507238 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:07Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:07 crc kubenswrapper[4754]: I1005 20:55:07.534532 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:07 crc kubenswrapper[4754]: I1005 20:55:07.534910 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:07 crc kubenswrapper[4754]: I1005 20:55:07.534987 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:07 crc kubenswrapper[4754]: I1005 20:55:07.535063 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:07 crc kubenswrapper[4754]: I1005 20:55:07.535137 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:07Z","lastTransitionTime":"2025-10-05T20:55:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:07 crc kubenswrapper[4754]: I1005 20:55:07.537705 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b44790fb-fecb-4ec3-9816-8e23dd72ab33\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a7dc53e7f96e48213d724a2cb4859514df925005644108ae3a0c33b14dcf52d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfdd8b1e542ab676576c311e2923fb02861f92de91fe1f4b3e4b968c549c375d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3a63451ffaa14cacc3b78d11340f4ea81445fa3af96178e9b172d063d9697d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3cfd3f2274a8a46fa36660e94ac6c9c298c791283f12d52330c5594ac131998\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f626cb108716da0b595844ea0251b7a04d49b5908a1d6dd0aef5cb2d402c4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://822223f515b5add6d71d193fd5bc77df029b8d37a1cc07be4145b53c7226bf61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c154d33bd80e9c252f693ffe0ec79b7bb6dbf1563c76343e8b3aba65f53b24b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2acf9637fa25e5d3ffeb05224c11218c47defc679568fb928a498869613de4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35ebeb7a87d5d47b6567917d38eca38a3e622c497fba58971296c4609c287423\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35ebeb7a87d5d47b6567917d38eca38a3e622c497fba58971296c4609c287423\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7rhps\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:07Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:07 crc kubenswrapper[4754]: I1005 20:55:07.553650 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-57sr4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e854c7b9-1229-4da4-80e9-fea6479459b8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceb577e1055d84007936a50d0b3d54b9ae0299a1fc1a6618b37163a6d3fbb9cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vw2ps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:58Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-57sr4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:07Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:07 crc kubenswrapper[4754]: I1005 20:55:07.568522 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cn76s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02d5e3f9-73c2-4496-9aca-0787184aef19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7ba2b076839f4a5f586d2e6c4f4d7188254128b782bfd2180e22bf3ea5ffdd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpx5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cn76s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:07Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:07 crc kubenswrapper[4754]: I1005 20:55:07.584532 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ss7lm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f9ef817-0622-4185-a7fb-57221690aa75\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ba613b0d7753a003c572c3733ed394707c45b6f8072cb7aa88f3c7312b8882b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wpwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:55:02Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ss7lm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:07Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:07 crc kubenswrapper[4754]: I1005 20:55:07.604429 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"934bcd66-9c03-48b9-82dd-5093338f9c8f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9042df2f43243d3d79ba9a3940f8dd4f6c7a89ef9e3510e81fe60b1944d7e83b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5366bbf46a91667355f43f5df7335a536b395cda788b90b35ef18e2b53bc4c02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://600bc3bb551b747cf2f18e26745883232d7a37ac5877fad0c67100e52d7a43aa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78f74982180048eb50b9be79fa72c04f62f925b54fe511920f37e6407f864631\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:07Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:07 crc kubenswrapper[4754]: I1005 20:55:07.621687 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:07Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:07 crc kubenswrapper[4754]: I1005 20:55:07.638223 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:07 crc kubenswrapper[4754]: I1005 20:55:07.638304 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:07 crc kubenswrapper[4754]: I1005 20:55:07.638319 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:07 crc kubenswrapper[4754]: I1005 20:55:07.638344 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:07 crc kubenswrapper[4754]: I1005 20:55:07.638360 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:07Z","lastTransitionTime":"2025-10-05T20:55:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:07 crc kubenswrapper[4754]: I1005 20:55:07.644046 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9ad070bb699b643d29b82bcc1126ce3b585f7492f5ae2817fa491b309b1691d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:07Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:07 crc kubenswrapper[4754]: I1005 20:55:07.657822 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be95a413-6a59-45b4-84b7-b43cae694a26\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10b211cf6e47dbd1ee42c82e49de0a13a45581b92a0c4f6bbba0326c7149e41d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2ggg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4bb8c5903e8b70dd19eebb111097d96f725e17de44e06c3e74c067376f6c7360\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2ggg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-b2h9k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:07Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:07 crc kubenswrapper[4754]: I1005 20:55:07.678887 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7506d496-65ff-41dc-8230-24c8ee653a2a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33da5a6bb3cd3b042c193bc571e6622f2f87f27e560cdd0bbb9be31e2010d707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df6d700a947d19c84b4f65fd2dd2f8d1f4f1f80fa9b4800c2acfa8dae3912639\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35ca8ee77a0e58cf969bd67b277a3392614c4e73b963e6f6002097709370715d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8907b9b69b249c34ec6db1e460d641b31711b75d6d31d4caafd867339e6571d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8907b9b69b249c34ec6db1e460d641b31711b75d6d31d4caafd867339e6571d0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1005 20:54:52.589115 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 20:54:52.590181 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1304624737/tls.crt::/tmp/serving-cert-1304624737/tls.key\\\\\\\"\\\\nI1005 20:54:58.070544 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1005 20:54:58.090352 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1005 20:54:58.090382 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1005 20:54:58.090412 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1005 20:54:58.090418 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1005 20:54:58.144656 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1005 20:54:58.144688 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 20:54:58.144695 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 20:54:58.144699 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1005 20:54:58.144702 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1005 20:54:58.144705 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1005 20:54:58.144708 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1005 20:54:58.145024 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1005 20:54:58.148112 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4efb34f3493cc8854bc4cff6b29f63973c52b754404d7a21daec4e1f612d327d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54c568b637f3737667433047b956c7eb22fdf0f9301acdfd3b49fb84c39de6ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54c568b637f3737667433047b956c7eb22fdf0f9301acdfd3b49fb84c39de6ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:07Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:07 crc kubenswrapper[4754]: I1005 20:55:07.696646 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c76e6e4740035a6ca54191a770c4516e28101aae69afbb89739174c2f076ef5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9eededc0cb1af94fb7b5daba08785f6b077a85e9cb5ddd8b9f397b658194f61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:07Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:07 crc kubenswrapper[4754]: I1005 20:55:07.714915 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tskt7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a68e9c89-465b-4c01-a470-e21a480ee465\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d14ec99610d2167b484bd7712af7bb6d993d26c74b2828da1c89314f16fda22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d14ec99610d2167b484bd7712af7bb6d993d26c74b2828da1c89314f16fda22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f6f08dd853b25881e2f3c66206aa80d0c493518284217780aa40ac4b69f0c0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f6f08dd853b25881e2f3c66206aa80d0c493518284217780aa40ac4b69f0c0b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://71026d00ec1a3d91b019deb3c9f43ee52fb05b8f2d6374e1a5b762c9e5c1b10d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71026d00ec1a3d91b019deb3c9f43ee52fb05b8f2d6374e1a5b762c9e5c1b10d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b9910d2470acb16ebb436d7030b3161061607cb2a5ef50d482d8d5c4f48a8ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b9910d2470acb16ebb436d7030b3161061607cb2a5ef50d482d8d5c4f48a8ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0a93b61eacc55111995e69c15f89a210b8aa15468ac9ac053b7cf6b007d301e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b0a93b61eacc55111995e69c15f89a210b8aa15468ac9ac053b7cf6b007d301e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cafb3208e7e3bcdd01898a68fc5bbf4cb52ca0800b879b77a60bb102794257f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cafb3208e7e3bcdd01898a68fc5bbf4cb52ca0800b879b77a60bb102794257f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tskt7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:07Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:07 crc kubenswrapper[4754]: I1005 20:55:07.741813 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:07 crc kubenswrapper[4754]: I1005 20:55:07.742168 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:07 crc kubenswrapper[4754]: I1005 20:55:07.742318 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:07 crc kubenswrapper[4754]: I1005 20:55:07.742406 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:07 crc kubenswrapper[4754]: I1005 20:55:07.742484 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:07Z","lastTransitionTime":"2025-10-05T20:55:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:07 crc kubenswrapper[4754]: I1005 20:55:07.846194 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:07 crc kubenswrapper[4754]: I1005 20:55:07.846258 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:07 crc kubenswrapper[4754]: I1005 20:55:07.846276 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:07 crc kubenswrapper[4754]: I1005 20:55:07.846303 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:07 crc kubenswrapper[4754]: I1005 20:55:07.846322 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:07Z","lastTransitionTime":"2025-10-05T20:55:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:07 crc kubenswrapper[4754]: I1005 20:55:07.950593 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:07 crc kubenswrapper[4754]: I1005 20:55:07.950650 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:07 crc kubenswrapper[4754]: I1005 20:55:07.950669 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:07 crc kubenswrapper[4754]: I1005 20:55:07.950692 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:07 crc kubenswrapper[4754]: I1005 20:55:07.950709 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:07Z","lastTransitionTime":"2025-10-05T20:55:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:08 crc kubenswrapper[4754]: I1005 20:55:08.054097 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:08 crc kubenswrapper[4754]: I1005 20:55:08.054652 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:08 crc kubenswrapper[4754]: I1005 20:55:08.054820 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:08 crc kubenswrapper[4754]: I1005 20:55:08.054957 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:08 crc kubenswrapper[4754]: I1005 20:55:08.055075 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:08Z","lastTransitionTime":"2025-10-05T20:55:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:08 crc kubenswrapper[4754]: I1005 20:55:08.159597 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:08 crc kubenswrapper[4754]: I1005 20:55:08.159684 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:08 crc kubenswrapper[4754]: I1005 20:55:08.159702 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:08 crc kubenswrapper[4754]: I1005 20:55:08.159731 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:08 crc kubenswrapper[4754]: I1005 20:55:08.159751 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:08Z","lastTransitionTime":"2025-10-05T20:55:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:08 crc kubenswrapper[4754]: I1005 20:55:08.169306 4754 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 05 20:55:08 crc kubenswrapper[4754]: I1005 20:55:08.169283 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-tskt7" event={"ID":"a68e9c89-465b-4c01-a470-e21a480ee465","Type":"ContainerStarted","Data":"ba6edba14cf321b4cd2f7bb44067a3ecf49f1594e76c0197e807ff3daa45e9b7"} Oct 05 20:55:08 crc kubenswrapper[4754]: I1005 20:55:08.171079 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" Oct 05 20:55:08 crc kubenswrapper[4754]: I1005 20:55:08.208569 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" Oct 05 20:55:08 crc kubenswrapper[4754]: I1005 20:55:08.209796 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9309179-b92a-4fa4-9fba-747b8f911936\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d969d3d50d6ad280fc6654df9441bab238bb538f277ed6d1273779813ffbaaed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://800a731ed8ef57714b2318af14054ef23f0368936630887a093fc77167b33d12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8cd9b08e83dc174413cc37891c14562816787028471a4201680c86dd620b7a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93b729f3f7f993db0d6bc3ba60c04930aada3cef8272ffcc3c6a8a242a9b1475\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dde1519111fc3af5d14463fdb58d4f5c62a7045e7de98c60981c7e4da37b0e7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ea160a5b37aea7f2970ae4dd2614b09d3dcf5dcbd1fc103b9997f8ecc3a804c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ea160a5b37aea7f2970ae4dd2614b09d3dcf5dcbd1fc103b9997f8ecc3a804c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://17ae6aa6c53e7dd061f70b41428ef50bbc7ffcc65352fad1a4a9d3c69dbd2ba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17ae6aa6c53e7dd061f70b41428ef50bbc7ffcc65352fad1a4a9d3c69dbd2ba9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://ad4cd69df7608c7d87d839fe09d4557c674c7efbc4730f4e4d3daaf9a26f5aee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad4cd69df7608c7d87d839fe09d4557c674c7efbc4730f4e4d3daaf9a26f5aee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:36Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:08Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:08 crc kubenswrapper[4754]: I1005 20:55:08.229403 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:08Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:08 crc kubenswrapper[4754]: I1005 20:55:08.253842 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b44790fb-fecb-4ec3-9816-8e23dd72ab33\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a7dc53e7f96e48213d724a2cb4859514df925005644108ae3a0c33b14dcf52d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfdd8b1e542ab676576c311e2923fb02861f92de91fe1f4b3e4b968c549c375d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3a63451ffaa14cacc3b78d11340f4ea81445fa3af96178e9b172d063d9697d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3cfd3f2274a8a46fa36660e94ac6c9c298c791283f12d52330c5594ac131998\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f626cb108716da0b595844ea0251b7a04d49b5908a1d6dd0aef5cb2d402c4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://822223f515b5add6d71d193fd5bc77df029b8d37a1cc07be4145b53c7226bf61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c154d33bd80e9c252f693ffe0ec79b7bb6dbf1563c76343e8b3aba65f53b24b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2acf9637fa25e5d3ffeb05224c11218c47defc679568fb928a498869613de4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35ebeb7a87d5d47b6567917d38eca38a3e622c497fba58971296c4609c287423\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35ebeb7a87d5d47b6567917d38eca38a3e622c497fba58971296c4609c287423\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7rhps\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:08Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:08 crc kubenswrapper[4754]: I1005 20:55:08.263198 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:08 crc kubenswrapper[4754]: I1005 20:55:08.263396 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:08 crc kubenswrapper[4754]: I1005 20:55:08.263484 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:08 crc kubenswrapper[4754]: I1005 20:55:08.263624 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:08 crc kubenswrapper[4754]: I1005 20:55:08.263713 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:08Z","lastTransitionTime":"2025-10-05T20:55:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:08 crc kubenswrapper[4754]: I1005 20:55:08.273775 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"934bcd66-9c03-48b9-82dd-5093338f9c8f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9042df2f43243d3d79ba9a3940f8dd4f6c7a89ef9e3510e81fe60b1944d7e83b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5366bbf46a91667355f43f5df7335a536b395cda788b90b35ef18e2b53bc4c02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://600bc3bb551b747cf2f18e26745883232d7a37ac5877fad0c67100e52d7a43aa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78f74982180048eb50b9be79fa72c04f62f925b54fe511920f37e6407f864631\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:08Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:08 crc kubenswrapper[4754]: I1005 20:55:08.289788 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:08Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:08 crc kubenswrapper[4754]: I1005 20:55:08.313275 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9ad070bb699b643d29b82bcc1126ce3b585f7492f5ae2817fa491b309b1691d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:08Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:08 crc kubenswrapper[4754]: I1005 20:55:08.326741 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-57sr4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e854c7b9-1229-4da4-80e9-fea6479459b8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceb577e1055d84007936a50d0b3d54b9ae0299a1fc1a6618b37163a6d3fbb9cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vw2ps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:58Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-57sr4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:08Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:08 crc kubenswrapper[4754]: I1005 20:55:08.341860 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cn76s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02d5e3f9-73c2-4496-9aca-0787184aef19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7ba2b076839f4a5f586d2e6c4f4d7188254128b782bfd2180e22bf3ea5ffdd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpx5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cn76s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:08Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:08 crc kubenswrapper[4754]: I1005 20:55:08.356019 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ss7lm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f9ef817-0622-4185-a7fb-57221690aa75\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ba613b0d7753a003c572c3733ed394707c45b6f8072cb7aa88f3c7312b8882b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wpwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:55:02Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ss7lm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:08Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:08 crc kubenswrapper[4754]: I1005 20:55:08.368621 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:08 crc kubenswrapper[4754]: I1005 20:55:08.368702 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:08 crc kubenswrapper[4754]: I1005 20:55:08.368723 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:08 crc kubenswrapper[4754]: I1005 20:55:08.368755 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:08 crc kubenswrapper[4754]: I1005 20:55:08.368777 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:08Z","lastTransitionTime":"2025-10-05T20:55:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:08 crc kubenswrapper[4754]: I1005 20:55:08.375849 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7506d496-65ff-41dc-8230-24c8ee653a2a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33da5a6bb3cd3b042c193bc571e6622f2f87f27e560cdd0bbb9be31e2010d707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df6d700a947d19c84b4f65fd2dd2f8d1f4f1f80fa9b4800c2acfa8dae3912639\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35ca8ee77a0e58cf969bd67b277a3392614c4e73b963e6f6002097709370715d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8907b9b69b249c34ec6db1e460d641b31711b75d6d31d4caafd867339e6571d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8907b9b69b249c34ec6db1e460d641b31711b75d6d31d4caafd867339e6571d0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1005 20:54:52.589115 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 20:54:52.590181 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1304624737/tls.crt::/tmp/serving-cert-1304624737/tls.key\\\\\\\"\\\\nI1005 20:54:58.070544 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1005 20:54:58.090352 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1005 20:54:58.090382 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1005 20:54:58.090412 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1005 20:54:58.090418 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1005 20:54:58.144656 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1005 20:54:58.144688 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 20:54:58.144695 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 20:54:58.144699 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1005 20:54:58.144702 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1005 20:54:58.144705 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1005 20:54:58.144708 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1005 20:54:58.145024 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1005 20:54:58.148112 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4efb34f3493cc8854bc4cff6b29f63973c52b754404d7a21daec4e1f612d327d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54c568b637f3737667433047b956c7eb22fdf0f9301acdfd3b49fb84c39de6ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54c568b637f3737667433047b956c7eb22fdf0f9301acdfd3b49fb84c39de6ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:08Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:08 crc kubenswrapper[4754]: I1005 20:55:08.389056 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c76e6e4740035a6ca54191a770c4516e28101aae69afbb89739174c2f076ef5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9eededc0cb1af94fb7b5daba08785f6b077a85e9cb5ddd8b9f397b658194f61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:08Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:08 crc kubenswrapper[4754]: I1005 20:55:08.406025 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tskt7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a68e9c89-465b-4c01-a470-e21a480ee465\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba6edba14cf321b4cd2f7bb44067a3ecf49f1594e76c0197e807ff3daa45e9b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d14ec99610d2167b484bd7712af7bb6d993d26c74b2828da1c89314f16fda22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d14ec99610d2167b484bd7712af7bb6d993d26c74b2828da1c89314f16fda22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f6f08dd853b25881e2f3c66206aa80d0c493518284217780aa40ac4b69f0c0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f6f08dd853b25881e2f3c66206aa80d0c493518284217780aa40ac4b69f0c0b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://71026d00ec1a3d91b019deb3c9f43ee52fb05b8f2d6374e1a5b762c9e5c1b10d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71026d00ec1a3d91b019deb3c9f43ee52fb05b8f2d6374e1a5b762c9e5c1b10d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b9910d2470acb16ebb436d7030b3161061607cb2a5ef50d482d8d5c4f48a8ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b9910d2470acb16ebb436d7030b3161061607cb2a5ef50d482d8d5c4f48a8ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0a93b61eacc55111995e69c15f89a210b8aa15468ac9ac053b7cf6b007d301e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b0a93b61eacc55111995e69c15f89a210b8aa15468ac9ac053b7cf6b007d301e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cafb3208e7e3bcdd01898a68fc5bbf4cb52ca0800b879b77a60bb102794257f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cafb3208e7e3bcdd01898a68fc5bbf4cb52ca0800b879b77a60bb102794257f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tskt7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:08Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:08 crc kubenswrapper[4754]: I1005 20:55:08.423532 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be95a413-6a59-45b4-84b7-b43cae694a26\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10b211cf6e47dbd1ee42c82e49de0a13a45581b92a0c4f6bbba0326c7149e41d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2ggg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4bb8c5903e8b70dd19eebb111097d96f725e17de44e06c3e74c067376f6c7360\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2ggg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-b2h9k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:08Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:08 crc kubenswrapper[4754]: I1005 20:55:08.442710 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e380f9b2a0e319dc325c54d1247a6db98346292548fa688cb41a2ca1b3048419\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:08Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:08 crc kubenswrapper[4754]: I1005 20:55:08.463658 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:08Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:08 crc kubenswrapper[4754]: I1005 20:55:08.472481 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:08 crc kubenswrapper[4754]: I1005 20:55:08.472567 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:08 crc kubenswrapper[4754]: I1005 20:55:08.472588 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:08 crc kubenswrapper[4754]: I1005 20:55:08.472615 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:08 crc kubenswrapper[4754]: I1005 20:55:08.472638 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:08Z","lastTransitionTime":"2025-10-05T20:55:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:08 crc kubenswrapper[4754]: I1005 20:55:08.486510 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e380f9b2a0e319dc325c54d1247a6db98346292548fa688cb41a2ca1b3048419\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:08Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:08 crc kubenswrapper[4754]: I1005 20:55:08.501078 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:08Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:08 crc kubenswrapper[4754]: I1005 20:55:08.535374 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9309179-b92a-4fa4-9fba-747b8f911936\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d969d3d50d6ad280fc6654df9441bab238bb538f277ed6d1273779813ffbaaed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://800a731ed8ef57714b2318af14054ef23f0368936630887a093fc77167b33d12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8cd9b08e83dc174413cc37891c14562816787028471a4201680c86dd620b7a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93b729f3f7f993db0d6bc3ba60c04930aada3cef8272ffcc3c6a8a242a9b1475\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dde1519111fc3af5d14463fdb58d4f5c62a7045e7de98c60981c7e4da37b0e7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ea160a5b37aea7f2970ae4dd2614b09d3dcf5dcbd1fc103b9997f8ecc3a804c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ea160a5b37aea7f2970ae4dd2614b09d3dcf5dcbd1fc103b9997f8ecc3a804c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://17ae6aa6c53e7dd061f70b41428ef50bbc7ffcc65352fad1a4a9d3c69dbd2ba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17ae6aa6c53e7dd061f70b41428ef50bbc7ffcc65352fad1a4a9d3c69dbd2ba9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://ad4cd69df7608c7d87d839fe09d4557c674c7efbc4730f4e4d3daaf9a26f5aee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad4cd69df7608c7d87d839fe09d4557c674c7efbc4730f4e4d3daaf9a26f5aee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:36Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:08Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:08 crc kubenswrapper[4754]: I1005 20:55:08.553425 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:08Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:08 crc kubenswrapper[4754]: I1005 20:55:08.576253 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:08 crc kubenswrapper[4754]: I1005 20:55:08.576297 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:08 crc kubenswrapper[4754]: I1005 20:55:08.576316 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:08 crc kubenswrapper[4754]: I1005 20:55:08.576342 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:08 crc kubenswrapper[4754]: I1005 20:55:08.576363 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:08Z","lastTransitionTime":"2025-10-05T20:55:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:08 crc kubenswrapper[4754]: I1005 20:55:08.595783 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b44790fb-fecb-4ec3-9816-8e23dd72ab33\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a7dc53e7f96e48213d724a2cb4859514df925005644108ae3a0c33b14dcf52d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfdd8b1e542ab676576c311e2923fb02861f92de91fe1f4b3e4b968c549c375d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3a63451ffaa14cacc3b78d11340f4ea81445fa3af96178e9b172d063d9697d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3cfd3f2274a8a46fa36660e94ac6c9c298c791283f12d52330c5594ac131998\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f626cb108716da0b595844ea0251b7a04d49b5908a1d6dd0aef5cb2d402c4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://822223f515b5add6d71d193fd5bc77df029b8d37a1cc07be4145b53c7226bf61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c154d33bd80e9c252f693ffe0ec79b7bb6dbf1563c76343e8b3aba65f53b24b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2acf9637fa25e5d3ffeb05224c11218c47defc679568fb928a498869613de4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35ebeb7a87d5d47b6567917d38eca38a3e622c497fba58971296c4609c287423\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35ebeb7a87d5d47b6567917d38eca38a3e622c497fba58971296c4609c287423\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7rhps\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:08Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:08 crc kubenswrapper[4754]: I1005 20:55:08.617103 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"934bcd66-9c03-48b9-82dd-5093338f9c8f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9042df2f43243d3d79ba9a3940f8dd4f6c7a89ef9e3510e81fe60b1944d7e83b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5366bbf46a91667355f43f5df7335a536b395cda788b90b35ef18e2b53bc4c02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://600bc3bb551b747cf2f18e26745883232d7a37ac5877fad0c67100e52d7a43aa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78f74982180048eb50b9be79fa72c04f62f925b54fe511920f37e6407f864631\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:08Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:08 crc kubenswrapper[4754]: I1005 20:55:08.642434 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:08Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:08 crc kubenswrapper[4754]: I1005 20:55:08.658949 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9ad070bb699b643d29b82bcc1126ce3b585f7492f5ae2817fa491b309b1691d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:08Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:08 crc kubenswrapper[4754]: I1005 20:55:08.673191 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-57sr4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e854c7b9-1229-4da4-80e9-fea6479459b8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceb577e1055d84007936a50d0b3d54b9ae0299a1fc1a6618b37163a6d3fbb9cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vw2ps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:58Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-57sr4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:08Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:08 crc kubenswrapper[4754]: I1005 20:55:08.679369 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:08 crc kubenswrapper[4754]: I1005 20:55:08.679413 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:08 crc kubenswrapper[4754]: I1005 20:55:08.679429 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:08 crc kubenswrapper[4754]: I1005 20:55:08.679454 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:08 crc kubenswrapper[4754]: I1005 20:55:08.679469 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:08Z","lastTransitionTime":"2025-10-05T20:55:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:08 crc kubenswrapper[4754]: I1005 20:55:08.724122 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cn76s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02d5e3f9-73c2-4496-9aca-0787184aef19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7ba2b076839f4a5f586d2e6c4f4d7188254128b782bfd2180e22bf3ea5ffdd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpx5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cn76s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:08Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:08 crc kubenswrapper[4754]: I1005 20:55:08.755269 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ss7lm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f9ef817-0622-4185-a7fb-57221690aa75\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ba613b0d7753a003c572c3733ed394707c45b6f8072cb7aa88f3c7312b8882b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wpwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:55:02Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ss7lm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:08Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:08 crc kubenswrapper[4754]: I1005 20:55:08.782037 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:08 crc kubenswrapper[4754]: I1005 20:55:08.782330 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:08 crc kubenswrapper[4754]: I1005 20:55:08.782393 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:08 crc kubenswrapper[4754]: I1005 20:55:08.782463 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:08 crc kubenswrapper[4754]: I1005 20:55:08.782587 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:08Z","lastTransitionTime":"2025-10-05T20:55:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:08 crc kubenswrapper[4754]: I1005 20:55:08.795845 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7506d496-65ff-41dc-8230-24c8ee653a2a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33da5a6bb3cd3b042c193bc571e6622f2f87f27e560cdd0bbb9be31e2010d707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df6d700a947d19c84b4f65fd2dd2f8d1f4f1f80fa9b4800c2acfa8dae3912639\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35ca8ee77a0e58cf969bd67b277a3392614c4e73b963e6f6002097709370715d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8907b9b69b249c34ec6db1e460d641b31711b75d6d31d4caafd867339e6571d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8907b9b69b249c34ec6db1e460d641b31711b75d6d31d4caafd867339e6571d0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1005 20:54:52.589115 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 20:54:52.590181 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1304624737/tls.crt::/tmp/serving-cert-1304624737/tls.key\\\\\\\"\\\\nI1005 20:54:58.070544 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1005 20:54:58.090352 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1005 20:54:58.090382 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1005 20:54:58.090412 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1005 20:54:58.090418 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1005 20:54:58.144656 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1005 20:54:58.144688 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 20:54:58.144695 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 20:54:58.144699 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1005 20:54:58.144702 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1005 20:54:58.144705 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1005 20:54:58.144708 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1005 20:54:58.145024 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1005 20:54:58.148112 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4efb34f3493cc8854bc4cff6b29f63973c52b754404d7a21daec4e1f612d327d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54c568b637f3737667433047b956c7eb22fdf0f9301acdfd3b49fb84c39de6ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54c568b637f3737667433047b956c7eb22fdf0f9301acdfd3b49fb84c39de6ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:08Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:08 crc kubenswrapper[4754]: I1005 20:55:08.835675 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c76e6e4740035a6ca54191a770c4516e28101aae69afbb89739174c2f076ef5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9eededc0cb1af94fb7b5daba08785f6b077a85e9cb5ddd8b9f397b658194f61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:08Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:08 crc kubenswrapper[4754]: I1005 20:55:08.836304 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 20:55:08 crc kubenswrapper[4754]: I1005 20:55:08.836366 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 20:55:08 crc kubenswrapper[4754]: I1005 20:55:08.836371 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 20:55:08 crc kubenswrapper[4754]: E1005 20:55:08.836480 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 20:55:08 crc kubenswrapper[4754]: E1005 20:55:08.836605 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 20:55:08 crc kubenswrapper[4754]: E1005 20:55:08.836772 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 20:55:08 crc kubenswrapper[4754]: I1005 20:55:08.876014 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tskt7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a68e9c89-465b-4c01-a470-e21a480ee465\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba6edba14cf321b4cd2f7bb44067a3ecf49f1594e76c0197e807ff3daa45e9b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d14ec99610d2167b484bd7712af7bb6d993d26c74b2828da1c89314f16fda22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d14ec99610d2167b484bd7712af7bb6d993d26c74b2828da1c89314f16fda22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f6f08dd853b25881e2f3c66206aa80d0c493518284217780aa40ac4b69f0c0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f6f08dd853b25881e2f3c66206aa80d0c493518284217780aa40ac4b69f0c0b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://71026d00ec1a3d91b019deb3c9f43ee52fb05b8f2d6374e1a5b762c9e5c1b10d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71026d00ec1a3d91b019deb3c9f43ee52fb05b8f2d6374e1a5b762c9e5c1b10d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b9910d2470acb16ebb436d7030b3161061607cb2a5ef50d482d8d5c4f48a8ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b9910d2470acb16ebb436d7030b3161061607cb2a5ef50d482d8d5c4f48a8ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0a93b61eacc55111995e69c15f89a210b8aa15468ac9ac053b7cf6b007d301e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b0a93b61eacc55111995e69c15f89a210b8aa15468ac9ac053b7cf6b007d301e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cafb3208e7e3bcdd01898a68fc5bbf4cb52ca0800b879b77a60bb102794257f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cafb3208e7e3bcdd01898a68fc5bbf4cb52ca0800b879b77a60bb102794257f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tskt7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:08Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:08 crc kubenswrapper[4754]: I1005 20:55:08.886959 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:08 crc kubenswrapper[4754]: I1005 20:55:08.887009 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:08 crc kubenswrapper[4754]: I1005 20:55:08.887022 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:08 crc kubenswrapper[4754]: I1005 20:55:08.887043 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:08 crc kubenswrapper[4754]: I1005 20:55:08.887056 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:08Z","lastTransitionTime":"2025-10-05T20:55:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:08 crc kubenswrapper[4754]: I1005 20:55:08.913992 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be95a413-6a59-45b4-84b7-b43cae694a26\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10b211cf6e47dbd1ee42c82e49de0a13a45581b92a0c4f6bbba0326c7149e41d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2ggg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4bb8c5903e8b70dd19eebb111097d96f725e17de44e06c3e74c067376f6c7360\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2ggg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-b2h9k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:08Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:08 crc kubenswrapper[4754]: I1005 20:55:08.990732 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:08 crc kubenswrapper[4754]: I1005 20:55:08.990773 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:08 crc kubenswrapper[4754]: I1005 20:55:08.990786 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:08 crc kubenswrapper[4754]: I1005 20:55:08.990805 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:08 crc kubenswrapper[4754]: I1005 20:55:08.990818 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:08Z","lastTransitionTime":"2025-10-05T20:55:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:09 crc kubenswrapper[4754]: I1005 20:55:09.094438 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:09 crc kubenswrapper[4754]: I1005 20:55:09.095107 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:09 crc kubenswrapper[4754]: I1005 20:55:09.095120 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:09 crc kubenswrapper[4754]: I1005 20:55:09.095140 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:09 crc kubenswrapper[4754]: I1005 20:55:09.095154 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:09Z","lastTransitionTime":"2025-10-05T20:55:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:09 crc kubenswrapper[4754]: I1005 20:55:09.172632 4754 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 05 20:55:09 crc kubenswrapper[4754]: I1005 20:55:09.200956 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:09 crc kubenswrapper[4754]: I1005 20:55:09.201011 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:09 crc kubenswrapper[4754]: I1005 20:55:09.201025 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:09 crc kubenswrapper[4754]: I1005 20:55:09.201048 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:09 crc kubenswrapper[4754]: I1005 20:55:09.201065 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:09Z","lastTransitionTime":"2025-10-05T20:55:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:09 crc kubenswrapper[4754]: I1005 20:55:09.303302 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:09 crc kubenswrapper[4754]: I1005 20:55:09.303358 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:09 crc kubenswrapper[4754]: I1005 20:55:09.303370 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:09 crc kubenswrapper[4754]: I1005 20:55:09.303393 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:09 crc kubenswrapper[4754]: I1005 20:55:09.303407 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:09Z","lastTransitionTime":"2025-10-05T20:55:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:09 crc kubenswrapper[4754]: I1005 20:55:09.407207 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:09 crc kubenswrapper[4754]: I1005 20:55:09.407277 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:09 crc kubenswrapper[4754]: I1005 20:55:09.407299 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:09 crc kubenswrapper[4754]: I1005 20:55:09.407329 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:09 crc kubenswrapper[4754]: I1005 20:55:09.407354 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:09Z","lastTransitionTime":"2025-10-05T20:55:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:09 crc kubenswrapper[4754]: I1005 20:55:09.510652 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:09 crc kubenswrapper[4754]: I1005 20:55:09.510727 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:09 crc kubenswrapper[4754]: I1005 20:55:09.510746 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:09 crc kubenswrapper[4754]: I1005 20:55:09.510782 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:09 crc kubenswrapper[4754]: I1005 20:55:09.510805 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:09Z","lastTransitionTime":"2025-10-05T20:55:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:09 crc kubenswrapper[4754]: I1005 20:55:09.614235 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:09 crc kubenswrapper[4754]: I1005 20:55:09.614317 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:09 crc kubenswrapper[4754]: I1005 20:55:09.614341 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:09 crc kubenswrapper[4754]: I1005 20:55:09.614374 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:09 crc kubenswrapper[4754]: I1005 20:55:09.614398 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:09Z","lastTransitionTime":"2025-10-05T20:55:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:09 crc kubenswrapper[4754]: I1005 20:55:09.719017 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:09 crc kubenswrapper[4754]: I1005 20:55:09.719074 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:09 crc kubenswrapper[4754]: I1005 20:55:09.719085 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:09 crc kubenswrapper[4754]: I1005 20:55:09.719111 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:09 crc kubenswrapper[4754]: I1005 20:55:09.719124 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:09Z","lastTransitionTime":"2025-10-05T20:55:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:09 crc kubenswrapper[4754]: I1005 20:55:09.823184 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:09 crc kubenswrapper[4754]: I1005 20:55:09.823443 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:09 crc kubenswrapper[4754]: I1005 20:55:09.823463 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:09 crc kubenswrapper[4754]: I1005 20:55:09.823526 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:09 crc kubenswrapper[4754]: I1005 20:55:09.823554 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:09Z","lastTransitionTime":"2025-10-05T20:55:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:09 crc kubenswrapper[4754]: I1005 20:55:09.927349 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:09 crc kubenswrapper[4754]: I1005 20:55:09.927417 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:09 crc kubenswrapper[4754]: I1005 20:55:09.927434 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:09 crc kubenswrapper[4754]: I1005 20:55:09.927462 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:09 crc kubenswrapper[4754]: I1005 20:55:09.927481 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:09Z","lastTransitionTime":"2025-10-05T20:55:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:10 crc kubenswrapper[4754]: I1005 20:55:10.031455 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:10 crc kubenswrapper[4754]: I1005 20:55:10.031572 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:10 crc kubenswrapper[4754]: I1005 20:55:10.031603 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:10 crc kubenswrapper[4754]: I1005 20:55:10.031637 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:10 crc kubenswrapper[4754]: I1005 20:55:10.031661 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:10Z","lastTransitionTime":"2025-10-05T20:55:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:10 crc kubenswrapper[4754]: I1005 20:55:10.134409 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:10 crc kubenswrapper[4754]: I1005 20:55:10.134485 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:10 crc kubenswrapper[4754]: I1005 20:55:10.134534 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:10 crc kubenswrapper[4754]: I1005 20:55:10.134569 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:10 crc kubenswrapper[4754]: I1005 20:55:10.134587 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:10Z","lastTransitionTime":"2025-10-05T20:55:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:10 crc kubenswrapper[4754]: I1005 20:55:10.178331 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7rhps_b44790fb-fecb-4ec3-9816-8e23dd72ab33/ovnkube-controller/0.log" Oct 05 20:55:10 crc kubenswrapper[4754]: I1005 20:55:10.183145 4754 generic.go:334] "Generic (PLEG): container finished" podID="b44790fb-fecb-4ec3-9816-8e23dd72ab33" containerID="c154d33bd80e9c252f693ffe0ec79b7bb6dbf1563c76343e8b3aba65f53b24b7" exitCode=1 Oct 05 20:55:10 crc kubenswrapper[4754]: I1005 20:55:10.183209 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" event={"ID":"b44790fb-fecb-4ec3-9816-8e23dd72ab33","Type":"ContainerDied","Data":"c154d33bd80e9c252f693ffe0ec79b7bb6dbf1563c76343e8b3aba65f53b24b7"} Oct 05 20:55:10 crc kubenswrapper[4754]: I1005 20:55:10.184668 4754 scope.go:117] "RemoveContainer" containerID="c154d33bd80e9c252f693ffe0ec79b7bb6dbf1563c76343e8b3aba65f53b24b7" Oct 05 20:55:10 crc kubenswrapper[4754]: I1005 20:55:10.206091 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ss7lm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f9ef817-0622-4185-a7fb-57221690aa75\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ba613b0d7753a003c572c3733ed394707c45b6f8072cb7aa88f3c7312b8882b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wpwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:55:02Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ss7lm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:10Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:10 crc kubenswrapper[4754]: I1005 20:55:10.228313 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"934bcd66-9c03-48b9-82dd-5093338f9c8f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9042df2f43243d3d79ba9a3940f8dd4f6c7a89ef9e3510e81fe60b1944d7e83b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5366bbf46a91667355f43f5df7335a536b395cda788b90b35ef18e2b53bc4c02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://600bc3bb551b747cf2f18e26745883232d7a37ac5877fad0c67100e52d7a43aa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78f74982180048eb50b9be79fa72c04f62f925b54fe511920f37e6407f864631\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:10Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:10 crc kubenswrapper[4754]: I1005 20:55:10.241008 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:10 crc kubenswrapper[4754]: I1005 20:55:10.245098 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:10 crc kubenswrapper[4754]: I1005 20:55:10.245138 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:10 crc kubenswrapper[4754]: I1005 20:55:10.245178 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:10 crc kubenswrapper[4754]: I1005 20:55:10.245202 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:10Z","lastTransitionTime":"2025-10-05T20:55:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:10 crc kubenswrapper[4754]: I1005 20:55:10.250689 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:10Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:10 crc kubenswrapper[4754]: I1005 20:55:10.272362 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9ad070bb699b643d29b82bcc1126ce3b585f7492f5ae2817fa491b309b1691d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:10Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:10 crc kubenswrapper[4754]: I1005 20:55:10.290395 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-57sr4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e854c7b9-1229-4da4-80e9-fea6479459b8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceb577e1055d84007936a50d0b3d54b9ae0299a1fc1a6618b37163a6d3fbb9cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vw2ps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:58Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-57sr4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:10Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:10 crc kubenswrapper[4754]: I1005 20:55:10.319543 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cn76s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02d5e3f9-73c2-4496-9aca-0787184aef19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7ba2b076839f4a5f586d2e6c4f4d7188254128b782bfd2180e22bf3ea5ffdd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpx5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cn76s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:10Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:10 crc kubenswrapper[4754]: I1005 20:55:10.340216 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7506d496-65ff-41dc-8230-24c8ee653a2a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33da5a6bb3cd3b042c193bc571e6622f2f87f27e560cdd0bbb9be31e2010d707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df6d700a947d19c84b4f65fd2dd2f8d1f4f1f80fa9b4800c2acfa8dae3912639\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35ca8ee77a0e58cf969bd67b277a3392614c4e73b963e6f6002097709370715d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8907b9b69b249c34ec6db1e460d641b31711b75d6d31d4caafd867339e6571d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8907b9b69b249c34ec6db1e460d641b31711b75d6d31d4caafd867339e6571d0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1005 20:54:52.589115 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 20:54:52.590181 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1304624737/tls.crt::/tmp/serving-cert-1304624737/tls.key\\\\\\\"\\\\nI1005 20:54:58.070544 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1005 20:54:58.090352 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1005 20:54:58.090382 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1005 20:54:58.090412 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1005 20:54:58.090418 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1005 20:54:58.144656 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1005 20:54:58.144688 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 20:54:58.144695 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 20:54:58.144699 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1005 20:54:58.144702 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1005 20:54:58.144705 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1005 20:54:58.144708 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1005 20:54:58.145024 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1005 20:54:58.148112 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4efb34f3493cc8854bc4cff6b29f63973c52b754404d7a21daec4e1f612d327d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54c568b637f3737667433047b956c7eb22fdf0f9301acdfd3b49fb84c39de6ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54c568b637f3737667433047b956c7eb22fdf0f9301acdfd3b49fb84c39de6ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:10Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:10 crc kubenswrapper[4754]: I1005 20:55:10.348425 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:10 crc kubenswrapper[4754]: I1005 20:55:10.348471 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:10 crc kubenswrapper[4754]: I1005 20:55:10.348483 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:10 crc kubenswrapper[4754]: I1005 20:55:10.348554 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:10 crc kubenswrapper[4754]: I1005 20:55:10.348760 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:10Z","lastTransitionTime":"2025-10-05T20:55:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:10 crc kubenswrapper[4754]: I1005 20:55:10.358995 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c76e6e4740035a6ca54191a770c4516e28101aae69afbb89739174c2f076ef5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9eededc0cb1af94fb7b5daba08785f6b077a85e9cb5ddd8b9f397b658194f61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:10Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:10 crc kubenswrapper[4754]: I1005 20:55:10.379197 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tskt7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a68e9c89-465b-4c01-a470-e21a480ee465\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba6edba14cf321b4cd2f7bb44067a3ecf49f1594e76c0197e807ff3daa45e9b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d14ec99610d2167b484bd7712af7bb6d993d26c74b2828da1c89314f16fda22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d14ec99610d2167b484bd7712af7bb6d993d26c74b2828da1c89314f16fda22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f6f08dd853b25881e2f3c66206aa80d0c493518284217780aa40ac4b69f0c0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f6f08dd853b25881e2f3c66206aa80d0c493518284217780aa40ac4b69f0c0b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://71026d00ec1a3d91b019deb3c9f43ee52fb05b8f2d6374e1a5b762c9e5c1b10d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71026d00ec1a3d91b019deb3c9f43ee52fb05b8f2d6374e1a5b762c9e5c1b10d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b9910d2470acb16ebb436d7030b3161061607cb2a5ef50d482d8d5c4f48a8ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b9910d2470acb16ebb436d7030b3161061607cb2a5ef50d482d8d5c4f48a8ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0a93b61eacc55111995e69c15f89a210b8aa15468ac9ac053b7cf6b007d301e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b0a93b61eacc55111995e69c15f89a210b8aa15468ac9ac053b7cf6b007d301e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cafb3208e7e3bcdd01898a68fc5bbf4cb52ca0800b879b77a60bb102794257f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cafb3208e7e3bcdd01898a68fc5bbf4cb52ca0800b879b77a60bb102794257f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tskt7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:10Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:10 crc kubenswrapper[4754]: I1005 20:55:10.394485 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be95a413-6a59-45b4-84b7-b43cae694a26\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10b211cf6e47dbd1ee42c82e49de0a13a45581b92a0c4f6bbba0326c7149e41d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2ggg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4bb8c5903e8b70dd19eebb111097d96f725e17de44e06c3e74c067376f6c7360\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2ggg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-b2h9k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:10Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:10 crc kubenswrapper[4754]: I1005 20:55:10.395434 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" Oct 05 20:55:10 crc kubenswrapper[4754]: I1005 20:55:10.415854 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e380f9b2a0e319dc325c54d1247a6db98346292548fa688cb41a2ca1b3048419\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:10Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:10 crc kubenswrapper[4754]: I1005 20:55:10.434864 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:10Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:10 crc kubenswrapper[4754]: I1005 20:55:10.451611 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:10 crc kubenswrapper[4754]: I1005 20:55:10.451639 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:10 crc kubenswrapper[4754]: I1005 20:55:10.451649 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:10 crc kubenswrapper[4754]: I1005 20:55:10.451664 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:10 crc kubenswrapper[4754]: I1005 20:55:10.451676 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:10Z","lastTransitionTime":"2025-10-05T20:55:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:10 crc kubenswrapper[4754]: I1005 20:55:10.459097 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9309179-b92a-4fa4-9fba-747b8f911936\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d969d3d50d6ad280fc6654df9441bab238bb538f277ed6d1273779813ffbaaed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://800a731ed8ef57714b2318af14054ef23f0368936630887a093fc77167b33d12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8cd9b08e83dc174413cc37891c14562816787028471a4201680c86dd620b7a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93b729f3f7f993db0d6bc3ba60c04930aada3cef8272ffcc3c6a8a242a9b1475\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dde1519111fc3af5d14463fdb58d4f5c62a7045e7de98c60981c7e4da37b0e7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ea160a5b37aea7f2970ae4dd2614b09d3dcf5dcbd1fc103b9997f8ecc3a804c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ea160a5b37aea7f2970ae4dd2614b09d3dcf5dcbd1fc103b9997f8ecc3a804c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://17ae6aa6c53e7dd061f70b41428ef50bbc7ffcc65352fad1a4a9d3c69dbd2ba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17ae6aa6c53e7dd061f70b41428ef50bbc7ffcc65352fad1a4a9d3c69dbd2ba9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://ad4cd69df7608c7d87d839fe09d4557c674c7efbc4730f4e4d3daaf9a26f5aee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad4cd69df7608c7d87d839fe09d4557c674c7efbc4730f4e4d3daaf9a26f5aee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:36Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:10Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:10 crc kubenswrapper[4754]: I1005 20:55:10.475884 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:10Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:10 crc kubenswrapper[4754]: I1005 20:55:10.503292 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b44790fb-fecb-4ec3-9816-8e23dd72ab33\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a7dc53e7f96e48213d724a2cb4859514df925005644108ae3a0c33b14dcf52d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfdd8b1e542ab676576c311e2923fb02861f92de91fe1f4b3e4b968c549c375d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3a63451ffaa14cacc3b78d11340f4ea81445fa3af96178e9b172d063d9697d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3cfd3f2274a8a46fa36660e94ac6c9c298c791283f12d52330c5594ac131998\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f626cb108716da0b595844ea0251b7a04d49b5908a1d6dd0aef5cb2d402c4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://822223f515b5add6d71d193fd5bc77df029b8d37a1cc07be4145b53c7226bf61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c154d33bd80e9c252f693ffe0ec79b7bb6dbf1563c76343e8b3aba65f53b24b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c154d33bd80e9c252f693ffe0ec79b7bb6dbf1563c76343e8b3aba65f53b24b7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-05T20:55:09Z\\\",\\\"message\\\":\\\"ressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1005 20:55:09.342783 5992 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1005 20:55:09.342837 5992 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1005 20:55:09.342971 5992 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1005 20:55:09.343568 5992 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1005 20:55:09.343630 5992 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1005 20:55:09.343661 5992 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1005 20:55:09.343670 5992 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1005 20:55:09.343709 5992 factory.go:656] Stopping watch factory\\\\nI1005 20:55:09.343727 5992 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1005 20:55:09.343739 5992 ovnkube.go:599] Stopped ovnkube\\\\nI1005 20:55:09.343743 5992 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1005 20:55:09.343733 5992 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1005 20\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2acf9637fa25e5d3ffeb05224c11218c47defc679568fb928a498869613de4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35ebeb7a87d5d47b6567917d38eca38a3e622c497fba58971296c4609c287423\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35ebeb7a87d5d47b6567917d38eca38a3e622c497fba58971296c4609c287423\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7rhps\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:10Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:10 crc kubenswrapper[4754]: I1005 20:55:10.554259 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:10 crc kubenswrapper[4754]: I1005 20:55:10.554316 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:10 crc kubenswrapper[4754]: I1005 20:55:10.554327 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:10 crc kubenswrapper[4754]: I1005 20:55:10.554347 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:10 crc kubenswrapper[4754]: I1005 20:55:10.554361 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:10Z","lastTransitionTime":"2025-10-05T20:55:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:10 crc kubenswrapper[4754]: I1005 20:55:10.657620 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:10 crc kubenswrapper[4754]: I1005 20:55:10.657648 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:10 crc kubenswrapper[4754]: I1005 20:55:10.657656 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:10 crc kubenswrapper[4754]: I1005 20:55:10.657670 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:10 crc kubenswrapper[4754]: I1005 20:55:10.657681 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:10Z","lastTransitionTime":"2025-10-05T20:55:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:10 crc kubenswrapper[4754]: I1005 20:55:10.760008 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:10 crc kubenswrapper[4754]: I1005 20:55:10.760047 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:10 crc kubenswrapper[4754]: I1005 20:55:10.760060 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:10 crc kubenswrapper[4754]: I1005 20:55:10.760077 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:10 crc kubenswrapper[4754]: I1005 20:55:10.760088 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:10Z","lastTransitionTime":"2025-10-05T20:55:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:10 crc kubenswrapper[4754]: I1005 20:55:10.836575 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 20:55:10 crc kubenswrapper[4754]: E1005 20:55:10.836809 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 20:55:10 crc kubenswrapper[4754]: I1005 20:55:10.837213 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 20:55:10 crc kubenswrapper[4754]: I1005 20:55:10.837343 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 20:55:10 crc kubenswrapper[4754]: E1005 20:55:10.837399 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 20:55:10 crc kubenswrapper[4754]: E1005 20:55:10.837536 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 20:55:10 crc kubenswrapper[4754]: I1005 20:55:10.863129 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:10 crc kubenswrapper[4754]: I1005 20:55:10.863178 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:10 crc kubenswrapper[4754]: I1005 20:55:10.863188 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:10 crc kubenswrapper[4754]: I1005 20:55:10.863206 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:10 crc kubenswrapper[4754]: I1005 20:55:10.863218 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:10Z","lastTransitionTime":"2025-10-05T20:55:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:10 crc kubenswrapper[4754]: I1005 20:55:10.966042 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:10 crc kubenswrapper[4754]: I1005 20:55:10.966096 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:10 crc kubenswrapper[4754]: I1005 20:55:10.966110 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:10 crc kubenswrapper[4754]: I1005 20:55:10.966131 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:10 crc kubenswrapper[4754]: I1005 20:55:10.966146 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:10Z","lastTransitionTime":"2025-10-05T20:55:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:11 crc kubenswrapper[4754]: I1005 20:55:11.069727 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:11 crc kubenswrapper[4754]: I1005 20:55:11.069791 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:11 crc kubenswrapper[4754]: I1005 20:55:11.069808 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:11 crc kubenswrapper[4754]: I1005 20:55:11.069867 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:11 crc kubenswrapper[4754]: I1005 20:55:11.069884 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:11Z","lastTransitionTime":"2025-10-05T20:55:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:11 crc kubenswrapper[4754]: I1005 20:55:11.173484 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:11 crc kubenswrapper[4754]: I1005 20:55:11.173554 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:11 crc kubenswrapper[4754]: I1005 20:55:11.173567 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:11 crc kubenswrapper[4754]: I1005 20:55:11.173587 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:11 crc kubenswrapper[4754]: I1005 20:55:11.173601 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:11Z","lastTransitionTime":"2025-10-05T20:55:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:11 crc kubenswrapper[4754]: I1005 20:55:11.190472 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7rhps_b44790fb-fecb-4ec3-9816-8e23dd72ab33/ovnkube-controller/0.log" Oct 05 20:55:11 crc kubenswrapper[4754]: I1005 20:55:11.194020 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" event={"ID":"b44790fb-fecb-4ec3-9816-8e23dd72ab33","Type":"ContainerStarted","Data":"2cbed906d679176209043edd5dcc649f5ba7d5848a10cfd09efa9e656dcacfef"} Oct 05 20:55:11 crc kubenswrapper[4754]: I1005 20:55:11.194888 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" Oct 05 20:55:11 crc kubenswrapper[4754]: I1005 20:55:11.216190 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9309179-b92a-4fa4-9fba-747b8f911936\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d969d3d50d6ad280fc6654df9441bab238bb538f277ed6d1273779813ffbaaed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://800a731ed8ef57714b2318af14054ef23f0368936630887a093fc77167b33d12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8cd9b08e83dc174413cc37891c14562816787028471a4201680c86dd620b7a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93b729f3f7f993db0d6bc3ba60c04930aada3cef8272ffcc3c6a8a242a9b1475\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dde1519111fc3af5d14463fdb58d4f5c62a7045e7de98c60981c7e4da37b0e7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ea160a5b37aea7f2970ae4dd2614b09d3dcf5dcbd1fc103b9997f8ecc3a804c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ea160a5b37aea7f2970ae4dd2614b09d3dcf5dcbd1fc103b9997f8ecc3a804c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://17ae6aa6c53e7dd061f70b41428ef50bbc7ffcc65352fad1a4a9d3c69dbd2ba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17ae6aa6c53e7dd061f70b41428ef50bbc7ffcc65352fad1a4a9d3c69dbd2ba9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://ad4cd69df7608c7d87d839fe09d4557c674c7efbc4730f4e4d3daaf9a26f5aee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad4cd69df7608c7d87d839fe09d4557c674c7efbc4730f4e4d3daaf9a26f5aee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:36Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:11Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:11 crc kubenswrapper[4754]: I1005 20:55:11.235278 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:11Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:11 crc kubenswrapper[4754]: I1005 20:55:11.259486 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b44790fb-fecb-4ec3-9816-8e23dd72ab33\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a7dc53e7f96e48213d724a2cb4859514df925005644108ae3a0c33b14dcf52d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfdd8b1e542ab676576c311e2923fb02861f92de91fe1f4b3e4b968c549c375d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3a63451ffaa14cacc3b78d11340f4ea81445fa3af96178e9b172d063d9697d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3cfd3f2274a8a46fa36660e94ac6c9c298c791283f12d52330c5594ac131998\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f626cb108716da0b595844ea0251b7a04d49b5908a1d6dd0aef5cb2d402c4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://822223f515b5add6d71d193fd5bc77df029b8d37a1cc07be4145b53c7226bf61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cbed906d679176209043edd5dcc649f5ba7d5848a10cfd09efa9e656dcacfef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c154d33bd80e9c252f693ffe0ec79b7bb6dbf1563c76343e8b3aba65f53b24b7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-05T20:55:09Z\\\",\\\"message\\\":\\\"ressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1005 20:55:09.342783 5992 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1005 20:55:09.342837 5992 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1005 20:55:09.342971 5992 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1005 20:55:09.343568 5992 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1005 20:55:09.343630 5992 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1005 20:55:09.343661 5992 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1005 20:55:09.343670 5992 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1005 20:55:09.343709 5992 factory.go:656] Stopping watch factory\\\\nI1005 20:55:09.343727 5992 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1005 20:55:09.343739 5992 ovnkube.go:599] Stopped ovnkube\\\\nI1005 20:55:09.343743 5992 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1005 20:55:09.343733 5992 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1005 20\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:06Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2acf9637fa25e5d3ffeb05224c11218c47defc679568fb928a498869613de4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35ebeb7a87d5d47b6567917d38eca38a3e622c497fba58971296c4609c287423\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35ebeb7a87d5d47b6567917d38eca38a3e622c497fba58971296c4609c287423\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7rhps\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:11Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:11 crc kubenswrapper[4754]: I1005 20:55:11.269708 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ss7lm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f9ef817-0622-4185-a7fb-57221690aa75\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ba613b0d7753a003c572c3733ed394707c45b6f8072cb7aa88f3c7312b8882b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wpwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:55:02Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ss7lm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:11Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:11 crc kubenswrapper[4754]: I1005 20:55:11.276694 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:11 crc kubenswrapper[4754]: I1005 20:55:11.276738 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:11 crc kubenswrapper[4754]: I1005 20:55:11.276749 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:11 crc kubenswrapper[4754]: I1005 20:55:11.276768 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:11 crc kubenswrapper[4754]: I1005 20:55:11.276783 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:11Z","lastTransitionTime":"2025-10-05T20:55:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:11 crc kubenswrapper[4754]: I1005 20:55:11.288336 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"934bcd66-9c03-48b9-82dd-5093338f9c8f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9042df2f43243d3d79ba9a3940f8dd4f6c7a89ef9e3510e81fe60b1944d7e83b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5366bbf46a91667355f43f5df7335a536b395cda788b90b35ef18e2b53bc4c02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://600bc3bb551b747cf2f18e26745883232d7a37ac5877fad0c67100e52d7a43aa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78f74982180048eb50b9be79fa72c04f62f925b54fe511920f37e6407f864631\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:11Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:11 crc kubenswrapper[4754]: I1005 20:55:11.308918 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:11Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:11 crc kubenswrapper[4754]: I1005 20:55:11.325433 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9ad070bb699b643d29b82bcc1126ce3b585f7492f5ae2817fa491b309b1691d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:11Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:11 crc kubenswrapper[4754]: I1005 20:55:11.342898 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-57sr4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e854c7b9-1229-4da4-80e9-fea6479459b8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceb577e1055d84007936a50d0b3d54b9ae0299a1fc1a6618b37163a6d3fbb9cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vw2ps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:58Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-57sr4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:11Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:11 crc kubenswrapper[4754]: I1005 20:55:11.365657 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cn76s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02d5e3f9-73c2-4496-9aca-0787184aef19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7ba2b076839f4a5f586d2e6c4f4d7188254128b782bfd2180e22bf3ea5ffdd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpx5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cn76s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:11Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:11 crc kubenswrapper[4754]: I1005 20:55:11.379986 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:11 crc kubenswrapper[4754]: I1005 20:55:11.380148 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:11 crc kubenswrapper[4754]: I1005 20:55:11.380246 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:11 crc kubenswrapper[4754]: I1005 20:55:11.380339 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:11 crc kubenswrapper[4754]: I1005 20:55:11.380434 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:11Z","lastTransitionTime":"2025-10-05T20:55:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:11 crc kubenswrapper[4754]: I1005 20:55:11.385883 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7506d496-65ff-41dc-8230-24c8ee653a2a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33da5a6bb3cd3b042c193bc571e6622f2f87f27e560cdd0bbb9be31e2010d707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df6d700a947d19c84b4f65fd2dd2f8d1f4f1f80fa9b4800c2acfa8dae3912639\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35ca8ee77a0e58cf969bd67b277a3392614c4e73b963e6f6002097709370715d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8907b9b69b249c34ec6db1e460d641b31711b75d6d31d4caafd867339e6571d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8907b9b69b249c34ec6db1e460d641b31711b75d6d31d4caafd867339e6571d0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1005 20:54:52.589115 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 20:54:52.590181 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1304624737/tls.crt::/tmp/serving-cert-1304624737/tls.key\\\\\\\"\\\\nI1005 20:54:58.070544 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1005 20:54:58.090352 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1005 20:54:58.090382 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1005 20:54:58.090412 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1005 20:54:58.090418 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1005 20:54:58.144656 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1005 20:54:58.144688 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 20:54:58.144695 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 20:54:58.144699 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1005 20:54:58.144702 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1005 20:54:58.144705 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1005 20:54:58.144708 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1005 20:54:58.145024 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1005 20:54:58.148112 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4efb34f3493cc8854bc4cff6b29f63973c52b754404d7a21daec4e1f612d327d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54c568b637f3737667433047b956c7eb22fdf0f9301acdfd3b49fb84c39de6ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54c568b637f3737667433047b956c7eb22fdf0f9301acdfd3b49fb84c39de6ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:11Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:11 crc kubenswrapper[4754]: I1005 20:55:11.406226 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c76e6e4740035a6ca54191a770c4516e28101aae69afbb89739174c2f076ef5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9eededc0cb1af94fb7b5daba08785f6b077a85e9cb5ddd8b9f397b658194f61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:11Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:11 crc kubenswrapper[4754]: I1005 20:55:11.427042 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tskt7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a68e9c89-465b-4c01-a470-e21a480ee465\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba6edba14cf321b4cd2f7bb44067a3ecf49f1594e76c0197e807ff3daa45e9b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d14ec99610d2167b484bd7712af7bb6d993d26c74b2828da1c89314f16fda22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d14ec99610d2167b484bd7712af7bb6d993d26c74b2828da1c89314f16fda22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f6f08dd853b25881e2f3c66206aa80d0c493518284217780aa40ac4b69f0c0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f6f08dd853b25881e2f3c66206aa80d0c493518284217780aa40ac4b69f0c0b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://71026d00ec1a3d91b019deb3c9f43ee52fb05b8f2d6374e1a5b762c9e5c1b10d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71026d00ec1a3d91b019deb3c9f43ee52fb05b8f2d6374e1a5b762c9e5c1b10d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b9910d2470acb16ebb436d7030b3161061607cb2a5ef50d482d8d5c4f48a8ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b9910d2470acb16ebb436d7030b3161061607cb2a5ef50d482d8d5c4f48a8ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0a93b61eacc55111995e69c15f89a210b8aa15468ac9ac053b7cf6b007d301e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b0a93b61eacc55111995e69c15f89a210b8aa15468ac9ac053b7cf6b007d301e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cafb3208e7e3bcdd01898a68fc5bbf4cb52ca0800b879b77a60bb102794257f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cafb3208e7e3bcdd01898a68fc5bbf4cb52ca0800b879b77a60bb102794257f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tskt7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:11Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:11 crc kubenswrapper[4754]: I1005 20:55:11.448975 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be95a413-6a59-45b4-84b7-b43cae694a26\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10b211cf6e47dbd1ee42c82e49de0a13a45581b92a0c4f6bbba0326c7149e41d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2ggg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4bb8c5903e8b70dd19eebb111097d96f725e17de44e06c3e74c067376f6c7360\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2ggg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-b2h9k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:11Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:11 crc kubenswrapper[4754]: I1005 20:55:11.471526 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e380f9b2a0e319dc325c54d1247a6db98346292548fa688cb41a2ca1b3048419\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:11Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:11 crc kubenswrapper[4754]: I1005 20:55:11.482921 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:11 crc kubenswrapper[4754]: I1005 20:55:11.482965 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:11 crc kubenswrapper[4754]: I1005 20:55:11.482977 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:11 crc kubenswrapper[4754]: I1005 20:55:11.483001 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:11 crc kubenswrapper[4754]: I1005 20:55:11.483017 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:11Z","lastTransitionTime":"2025-10-05T20:55:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:11 crc kubenswrapper[4754]: I1005 20:55:11.490929 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:11Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:11 crc kubenswrapper[4754]: I1005 20:55:11.586984 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:11 crc kubenswrapper[4754]: I1005 20:55:11.587029 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:11 crc kubenswrapper[4754]: I1005 20:55:11.587047 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:11 crc kubenswrapper[4754]: I1005 20:55:11.587075 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:11 crc kubenswrapper[4754]: I1005 20:55:11.587095 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:11Z","lastTransitionTime":"2025-10-05T20:55:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:11 crc kubenswrapper[4754]: I1005 20:55:11.690764 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:11 crc kubenswrapper[4754]: I1005 20:55:11.690835 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:11 crc kubenswrapper[4754]: I1005 20:55:11.690857 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:11 crc kubenswrapper[4754]: I1005 20:55:11.690953 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:11 crc kubenswrapper[4754]: I1005 20:55:11.690975 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:11Z","lastTransitionTime":"2025-10-05T20:55:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:11 crc kubenswrapper[4754]: I1005 20:55:11.795203 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:11 crc kubenswrapper[4754]: I1005 20:55:11.795530 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:11 crc kubenswrapper[4754]: I1005 20:55:11.795703 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:11 crc kubenswrapper[4754]: I1005 20:55:11.795875 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:11 crc kubenswrapper[4754]: I1005 20:55:11.796018 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:11Z","lastTransitionTime":"2025-10-05T20:55:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:11 crc kubenswrapper[4754]: I1005 20:55:11.900131 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:11 crc kubenswrapper[4754]: I1005 20:55:11.900209 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:11 crc kubenswrapper[4754]: I1005 20:55:11.900235 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:11 crc kubenswrapper[4754]: I1005 20:55:11.900264 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:11 crc kubenswrapper[4754]: I1005 20:55:11.900289 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:11Z","lastTransitionTime":"2025-10-05T20:55:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:12 crc kubenswrapper[4754]: I1005 20:55:12.003802 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:12 crc kubenswrapper[4754]: I1005 20:55:12.004322 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:12 crc kubenswrapper[4754]: I1005 20:55:12.004488 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:12 crc kubenswrapper[4754]: I1005 20:55:12.004673 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:12 crc kubenswrapper[4754]: I1005 20:55:12.004798 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:12Z","lastTransitionTime":"2025-10-05T20:55:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:12 crc kubenswrapper[4754]: I1005 20:55:12.108430 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:12 crc kubenswrapper[4754]: I1005 20:55:12.108555 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:12 crc kubenswrapper[4754]: I1005 20:55:12.108577 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:12 crc kubenswrapper[4754]: I1005 20:55:12.108608 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:12 crc kubenswrapper[4754]: I1005 20:55:12.108628 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:12Z","lastTransitionTime":"2025-10-05T20:55:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:12 crc kubenswrapper[4754]: I1005 20:55:12.201689 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7rhps_b44790fb-fecb-4ec3-9816-8e23dd72ab33/ovnkube-controller/1.log" Oct 05 20:55:12 crc kubenswrapper[4754]: I1005 20:55:12.202460 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7rhps_b44790fb-fecb-4ec3-9816-8e23dd72ab33/ovnkube-controller/0.log" Oct 05 20:55:12 crc kubenswrapper[4754]: I1005 20:55:12.206569 4754 generic.go:334] "Generic (PLEG): container finished" podID="b44790fb-fecb-4ec3-9816-8e23dd72ab33" containerID="2cbed906d679176209043edd5dcc649f5ba7d5848a10cfd09efa9e656dcacfef" exitCode=1 Oct 05 20:55:12 crc kubenswrapper[4754]: I1005 20:55:12.206620 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" event={"ID":"b44790fb-fecb-4ec3-9816-8e23dd72ab33","Type":"ContainerDied","Data":"2cbed906d679176209043edd5dcc649f5ba7d5848a10cfd09efa9e656dcacfef"} Oct 05 20:55:12 crc kubenswrapper[4754]: I1005 20:55:12.206672 4754 scope.go:117] "RemoveContainer" containerID="c154d33bd80e9c252f693ffe0ec79b7bb6dbf1563c76343e8b3aba65f53b24b7" Oct 05 20:55:12 crc kubenswrapper[4754]: I1005 20:55:12.208031 4754 scope.go:117] "RemoveContainer" containerID="2cbed906d679176209043edd5dcc649f5ba7d5848a10cfd09efa9e656dcacfef" Oct 05 20:55:12 crc kubenswrapper[4754]: E1005 20:55:12.208484 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-7rhps_openshift-ovn-kubernetes(b44790fb-fecb-4ec3-9816-8e23dd72ab33)\"" pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" podUID="b44790fb-fecb-4ec3-9816-8e23dd72ab33" Oct 05 20:55:12 crc kubenswrapper[4754]: I1005 20:55:12.212239 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:12 crc kubenswrapper[4754]: I1005 20:55:12.214521 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:12 crc kubenswrapper[4754]: I1005 20:55:12.214556 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:12 crc kubenswrapper[4754]: I1005 20:55:12.214584 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:12 crc kubenswrapper[4754]: I1005 20:55:12.214632 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:12Z","lastTransitionTime":"2025-10-05T20:55:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:12 crc kubenswrapper[4754]: I1005 20:55:12.226772 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ss7lm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f9ef817-0622-4185-a7fb-57221690aa75\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ba613b0d7753a003c572c3733ed394707c45b6f8072cb7aa88f3c7312b8882b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wpwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:55:02Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ss7lm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:12Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:12 crc kubenswrapper[4754]: I1005 20:55:12.255906 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"934bcd66-9c03-48b9-82dd-5093338f9c8f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9042df2f43243d3d79ba9a3940f8dd4f6c7a89ef9e3510e81fe60b1944d7e83b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5366bbf46a91667355f43f5df7335a536b395cda788b90b35ef18e2b53bc4c02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://600bc3bb551b747cf2f18e26745883232d7a37ac5877fad0c67100e52d7a43aa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78f74982180048eb50b9be79fa72c04f62f925b54fe511920f37e6407f864631\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:12Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:12 crc kubenswrapper[4754]: I1005 20:55:12.279234 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:12Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:12 crc kubenswrapper[4754]: I1005 20:55:12.302469 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9ad070bb699b643d29b82bcc1126ce3b585f7492f5ae2817fa491b309b1691d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:12Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:12 crc kubenswrapper[4754]: I1005 20:55:12.318098 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:12 crc kubenswrapper[4754]: I1005 20:55:12.318152 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:12 crc kubenswrapper[4754]: I1005 20:55:12.318170 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:12 crc kubenswrapper[4754]: I1005 20:55:12.318199 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:12 crc kubenswrapper[4754]: I1005 20:55:12.318218 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:12Z","lastTransitionTime":"2025-10-05T20:55:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:12 crc kubenswrapper[4754]: I1005 20:55:12.326944 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-57sr4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e854c7b9-1229-4da4-80e9-fea6479459b8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceb577e1055d84007936a50d0b3d54b9ae0299a1fc1a6618b37163a6d3fbb9cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vw2ps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:58Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-57sr4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:12Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:12 crc kubenswrapper[4754]: I1005 20:55:12.347651 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cn76s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02d5e3f9-73c2-4496-9aca-0787184aef19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7ba2b076839f4a5f586d2e6c4f4d7188254128b782bfd2180e22bf3ea5ffdd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpx5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cn76s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:12Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:12 crc kubenswrapper[4754]: I1005 20:55:12.369306 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7506d496-65ff-41dc-8230-24c8ee653a2a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33da5a6bb3cd3b042c193bc571e6622f2f87f27e560cdd0bbb9be31e2010d707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df6d700a947d19c84b4f65fd2dd2f8d1f4f1f80fa9b4800c2acfa8dae3912639\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35ca8ee77a0e58cf969bd67b277a3392614c4e73b963e6f6002097709370715d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8907b9b69b249c34ec6db1e460d641b31711b75d6d31d4caafd867339e6571d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8907b9b69b249c34ec6db1e460d641b31711b75d6d31d4caafd867339e6571d0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1005 20:54:52.589115 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 20:54:52.590181 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1304624737/tls.crt::/tmp/serving-cert-1304624737/tls.key\\\\\\\"\\\\nI1005 20:54:58.070544 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1005 20:54:58.090352 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1005 20:54:58.090382 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1005 20:54:58.090412 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1005 20:54:58.090418 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1005 20:54:58.144656 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1005 20:54:58.144688 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 20:54:58.144695 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 20:54:58.144699 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1005 20:54:58.144702 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1005 20:54:58.144705 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1005 20:54:58.144708 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1005 20:54:58.145024 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1005 20:54:58.148112 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4efb34f3493cc8854bc4cff6b29f63973c52b754404d7a21daec4e1f612d327d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54c568b637f3737667433047b956c7eb22fdf0f9301acdfd3b49fb84c39de6ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54c568b637f3737667433047b956c7eb22fdf0f9301acdfd3b49fb84c39de6ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:12Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:12 crc kubenswrapper[4754]: I1005 20:55:12.391549 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c76e6e4740035a6ca54191a770c4516e28101aae69afbb89739174c2f076ef5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9eededc0cb1af94fb7b5daba08785f6b077a85e9cb5ddd8b9f397b658194f61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:12Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:12 crc kubenswrapper[4754]: I1005 20:55:12.411543 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tskt7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a68e9c89-465b-4c01-a470-e21a480ee465\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba6edba14cf321b4cd2f7bb44067a3ecf49f1594e76c0197e807ff3daa45e9b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d14ec99610d2167b484bd7712af7bb6d993d26c74b2828da1c89314f16fda22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d14ec99610d2167b484bd7712af7bb6d993d26c74b2828da1c89314f16fda22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f6f08dd853b25881e2f3c66206aa80d0c493518284217780aa40ac4b69f0c0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f6f08dd853b25881e2f3c66206aa80d0c493518284217780aa40ac4b69f0c0b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://71026d00ec1a3d91b019deb3c9f43ee52fb05b8f2d6374e1a5b762c9e5c1b10d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71026d00ec1a3d91b019deb3c9f43ee52fb05b8f2d6374e1a5b762c9e5c1b10d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b9910d2470acb16ebb436d7030b3161061607cb2a5ef50d482d8d5c4f48a8ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b9910d2470acb16ebb436d7030b3161061607cb2a5ef50d482d8d5c4f48a8ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0a93b61eacc55111995e69c15f89a210b8aa15468ac9ac053b7cf6b007d301e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b0a93b61eacc55111995e69c15f89a210b8aa15468ac9ac053b7cf6b007d301e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cafb3208e7e3bcdd01898a68fc5bbf4cb52ca0800b879b77a60bb102794257f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cafb3208e7e3bcdd01898a68fc5bbf4cb52ca0800b879b77a60bb102794257f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tskt7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:12Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:12 crc kubenswrapper[4754]: I1005 20:55:12.421449 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:12 crc kubenswrapper[4754]: I1005 20:55:12.421522 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:12 crc kubenswrapper[4754]: I1005 20:55:12.421547 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:12 crc kubenswrapper[4754]: I1005 20:55:12.421574 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:12 crc kubenswrapper[4754]: I1005 20:55:12.421594 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:12Z","lastTransitionTime":"2025-10-05T20:55:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:12 crc kubenswrapper[4754]: I1005 20:55:12.431916 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be95a413-6a59-45b4-84b7-b43cae694a26\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10b211cf6e47dbd1ee42c82e49de0a13a45581b92a0c4f6bbba0326c7149e41d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2ggg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4bb8c5903e8b70dd19eebb111097d96f725e17de44e06c3e74c067376f6c7360\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2ggg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-b2h9k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:12Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:12 crc kubenswrapper[4754]: I1005 20:55:12.449403 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e380f9b2a0e319dc325c54d1247a6db98346292548fa688cb41a2ca1b3048419\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:12Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:12 crc kubenswrapper[4754]: I1005 20:55:12.468432 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:12Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:12 crc kubenswrapper[4754]: I1005 20:55:12.496800 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9309179-b92a-4fa4-9fba-747b8f911936\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d969d3d50d6ad280fc6654df9441bab238bb538f277ed6d1273779813ffbaaed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://800a731ed8ef57714b2318af14054ef23f0368936630887a093fc77167b33d12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8cd9b08e83dc174413cc37891c14562816787028471a4201680c86dd620b7a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93b729f3f7f993db0d6bc3ba60c04930aada3cef8272ffcc3c6a8a242a9b1475\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dde1519111fc3af5d14463fdb58d4f5c62a7045e7de98c60981c7e4da37b0e7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ea160a5b37aea7f2970ae4dd2614b09d3dcf5dcbd1fc103b9997f8ecc3a804c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ea160a5b37aea7f2970ae4dd2614b09d3dcf5dcbd1fc103b9997f8ecc3a804c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://17ae6aa6c53e7dd061f70b41428ef50bbc7ffcc65352fad1a4a9d3c69dbd2ba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17ae6aa6c53e7dd061f70b41428ef50bbc7ffcc65352fad1a4a9d3c69dbd2ba9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://ad4cd69df7608c7d87d839fe09d4557c674c7efbc4730f4e4d3daaf9a26f5aee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad4cd69df7608c7d87d839fe09d4557c674c7efbc4730f4e4d3daaf9a26f5aee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:36Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:12Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:12 crc kubenswrapper[4754]: I1005 20:55:12.517144 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:12Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:12 crc kubenswrapper[4754]: I1005 20:55:12.525188 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:12 crc kubenswrapper[4754]: I1005 20:55:12.525238 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:12 crc kubenswrapper[4754]: I1005 20:55:12.525256 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:12 crc kubenswrapper[4754]: I1005 20:55:12.525285 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:12 crc kubenswrapper[4754]: I1005 20:55:12.525306 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:12Z","lastTransitionTime":"2025-10-05T20:55:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:12 crc kubenswrapper[4754]: I1005 20:55:12.549082 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b44790fb-fecb-4ec3-9816-8e23dd72ab33\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a7dc53e7f96e48213d724a2cb4859514df925005644108ae3a0c33b14dcf52d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfdd8b1e542ab676576c311e2923fb02861f92de91fe1f4b3e4b968c549c375d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3a63451ffaa14cacc3b78d11340f4ea81445fa3af96178e9b172d063d9697d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3cfd3f2274a8a46fa36660e94ac6c9c298c791283f12d52330c5594ac131998\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f626cb108716da0b595844ea0251b7a04d49b5908a1d6dd0aef5cb2d402c4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://822223f515b5add6d71d193fd5bc77df029b8d37a1cc07be4145b53c7226bf61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cbed906d679176209043edd5dcc649f5ba7d5848a10cfd09efa9e656dcacfef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c154d33bd80e9c252f693ffe0ec79b7bb6dbf1563c76343e8b3aba65f53b24b7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-05T20:55:09Z\\\",\\\"message\\\":\\\"ressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1005 20:55:09.342783 5992 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1005 20:55:09.342837 5992 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1005 20:55:09.342971 5992 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1005 20:55:09.343568 5992 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1005 20:55:09.343630 5992 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1005 20:55:09.343661 5992 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1005 20:55:09.343670 5992 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1005 20:55:09.343709 5992 factory.go:656] Stopping watch factory\\\\nI1005 20:55:09.343727 5992 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1005 20:55:09.343739 5992 ovnkube.go:599] Stopped ovnkube\\\\nI1005 20:55:09.343743 5992 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1005 20:55:09.343733 5992 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1005 20\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:06Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cbed906d679176209043edd5dcc649f5ba7d5848a10cfd09efa9e656dcacfef\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-05T20:55:11Z\\\",\\\"message\\\":\\\"gressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1005 20:55:11.136172 6140 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1005 20:55:11.136316 6140 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1005 20:55:11.136624 6140 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1005 20:55:11.136664 6140 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1005 20:55:11.136709 6140 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1005 20:55:11.136749 6140 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1005 20:55:11.138444 6140 factory.go:656] Stopping watch factory\\\\nI1005 20:55:11.148575 6140 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1005 20:55:11.148612 6140 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1005 20:55:11.148701 6140 ovnkube.go:599] Stopped ovnkube\\\\nI1005 20:55:11.148748 6140 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1005 20:55:11.148871 6140 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2acf9637fa25e5d3ffeb05224c11218c47defc679568fb928a498869613de4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35ebeb7a87d5d47b6567917d38eca38a3e622c497fba58971296c4609c287423\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35ebeb7a87d5d47b6567917d38eca38a3e622c497fba58971296c4609c287423\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7rhps\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:12Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:12 crc kubenswrapper[4754]: I1005 20:55:12.629154 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:12 crc kubenswrapper[4754]: I1005 20:55:12.629218 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:12 crc kubenswrapper[4754]: I1005 20:55:12.629248 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:12 crc kubenswrapper[4754]: I1005 20:55:12.629271 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:12 crc kubenswrapper[4754]: I1005 20:55:12.629288 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:12Z","lastTransitionTime":"2025-10-05T20:55:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:12 crc kubenswrapper[4754]: I1005 20:55:12.732411 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:12 crc kubenswrapper[4754]: I1005 20:55:12.732473 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:12 crc kubenswrapper[4754]: I1005 20:55:12.732520 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:12 crc kubenswrapper[4754]: I1005 20:55:12.732542 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:12 crc kubenswrapper[4754]: I1005 20:55:12.732556 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:12Z","lastTransitionTime":"2025-10-05T20:55:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:12 crc kubenswrapper[4754]: I1005 20:55:12.836444 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 20:55:12 crc kubenswrapper[4754]: I1005 20:55:12.836624 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:12 crc kubenswrapper[4754]: I1005 20:55:12.836685 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:12 crc kubenswrapper[4754]: I1005 20:55:12.836704 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:12 crc kubenswrapper[4754]: I1005 20:55:12.836732 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:12 crc kubenswrapper[4754]: I1005 20:55:12.836751 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:12Z","lastTransitionTime":"2025-10-05T20:55:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:12 crc kubenswrapper[4754]: I1005 20:55:12.837278 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 20:55:12 crc kubenswrapper[4754]: I1005 20:55:12.837292 4754 scope.go:117] "RemoveContainer" containerID="8907b9b69b249c34ec6db1e460d641b31711b75d6d31d4caafd867339e6571d0" Oct 05 20:55:12 crc kubenswrapper[4754]: I1005 20:55:12.837348 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 20:55:12 crc kubenswrapper[4754]: E1005 20:55:12.837477 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 20:55:12 crc kubenswrapper[4754]: E1005 20:55:12.837637 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 20:55:12 crc kubenswrapper[4754]: E1005 20:55:12.837730 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 20:55:12 crc kubenswrapper[4754]: I1005 20:55:12.885476 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rddkl"] Oct 05 20:55:12 crc kubenswrapper[4754]: I1005 20:55:12.886165 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rddkl" Oct 05 20:55:12 crc kubenswrapper[4754]: I1005 20:55:12.893716 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Oct 05 20:55:12 crc kubenswrapper[4754]: I1005 20:55:12.893902 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Oct 05 20:55:12 crc kubenswrapper[4754]: I1005 20:55:12.927039 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9309179-b92a-4fa4-9fba-747b8f911936\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d969d3d50d6ad280fc6654df9441bab238bb538f277ed6d1273779813ffbaaed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://800a731ed8ef57714b2318af14054ef23f0368936630887a093fc77167b33d12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8cd9b08e83dc174413cc37891c14562816787028471a4201680c86dd620b7a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93b729f3f7f993db0d6bc3ba60c04930aada3cef8272ffcc3c6a8a242a9b1475\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dde1519111fc3af5d14463fdb58d4f5c62a7045e7de98c60981c7e4da37b0e7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ea160a5b37aea7f2970ae4dd2614b09d3dcf5dcbd1fc103b9997f8ecc3a804c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ea160a5b37aea7f2970ae4dd2614b09d3dcf5dcbd1fc103b9997f8ecc3a804c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://17ae6aa6c53e7dd061f70b41428ef50bbc7ffcc65352fad1a4a9d3c69dbd2ba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17ae6aa6c53e7dd061f70b41428ef50bbc7ffcc65352fad1a4a9d3c69dbd2ba9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://ad4cd69df7608c7d87d839fe09d4557c674c7efbc4730f4e4d3daaf9a26f5aee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad4cd69df7608c7d87d839fe09d4557c674c7efbc4730f4e4d3daaf9a26f5aee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:36Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:12Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:12 crc kubenswrapper[4754]: I1005 20:55:12.940594 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:12 crc kubenswrapper[4754]: I1005 20:55:12.940646 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:12 crc kubenswrapper[4754]: I1005 20:55:12.940666 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:12 crc kubenswrapper[4754]: I1005 20:55:12.940693 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:12 crc kubenswrapper[4754]: I1005 20:55:12.940712 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:12Z","lastTransitionTime":"2025-10-05T20:55:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:12 crc kubenswrapper[4754]: I1005 20:55:12.951952 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:12Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:12 crc kubenswrapper[4754]: I1005 20:55:12.965387 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cqv8k\" (UniqueName: \"kubernetes.io/projected/15afc2a1-5cb7-41c6-954b-446c36822f50-kube-api-access-cqv8k\") pod \"ovnkube-control-plane-749d76644c-rddkl\" (UID: \"15afc2a1-5cb7-41c6-954b-446c36822f50\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rddkl" Oct 05 20:55:12 crc kubenswrapper[4754]: I1005 20:55:12.965485 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/15afc2a1-5cb7-41c6-954b-446c36822f50-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-rddkl\" (UID: \"15afc2a1-5cb7-41c6-954b-446c36822f50\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rddkl" Oct 05 20:55:12 crc kubenswrapper[4754]: I1005 20:55:12.965535 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/15afc2a1-5cb7-41c6-954b-446c36822f50-env-overrides\") pod \"ovnkube-control-plane-749d76644c-rddkl\" (UID: \"15afc2a1-5cb7-41c6-954b-446c36822f50\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rddkl" Oct 05 20:55:12 crc kubenswrapper[4754]: I1005 20:55:12.965566 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/15afc2a1-5cb7-41c6-954b-446c36822f50-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-rddkl\" (UID: \"15afc2a1-5cb7-41c6-954b-446c36822f50\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rddkl" Oct 05 20:55:12 crc kubenswrapper[4754]: I1005 20:55:12.984082 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b44790fb-fecb-4ec3-9816-8e23dd72ab33\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a7dc53e7f96e48213d724a2cb4859514df925005644108ae3a0c33b14dcf52d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfdd8b1e542ab676576c311e2923fb02861f92de91fe1f4b3e4b968c549c375d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3a63451ffaa14cacc3b78d11340f4ea81445fa3af96178e9b172d063d9697d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3cfd3f2274a8a46fa36660e94ac6c9c298c791283f12d52330c5594ac131998\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f626cb108716da0b595844ea0251b7a04d49b5908a1d6dd0aef5cb2d402c4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://822223f515b5add6d71d193fd5bc77df029b8d37a1cc07be4145b53c7226bf61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cbed906d679176209043edd5dcc649f5ba7d5848a10cfd09efa9e656dcacfef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c154d33bd80e9c252f693ffe0ec79b7bb6dbf1563c76343e8b3aba65f53b24b7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-05T20:55:09Z\\\",\\\"message\\\":\\\"ressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1005 20:55:09.342783 5992 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1005 20:55:09.342837 5992 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1005 20:55:09.342971 5992 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1005 20:55:09.343568 5992 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1005 20:55:09.343630 5992 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1005 20:55:09.343661 5992 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1005 20:55:09.343670 5992 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1005 20:55:09.343709 5992 factory.go:656] Stopping watch factory\\\\nI1005 20:55:09.343727 5992 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1005 20:55:09.343739 5992 ovnkube.go:599] Stopped ovnkube\\\\nI1005 20:55:09.343743 5992 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1005 20:55:09.343733 5992 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1005 20\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:06Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cbed906d679176209043edd5dcc649f5ba7d5848a10cfd09efa9e656dcacfef\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-05T20:55:11Z\\\",\\\"message\\\":\\\"gressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1005 20:55:11.136172 6140 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1005 20:55:11.136316 6140 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1005 20:55:11.136624 6140 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1005 20:55:11.136664 6140 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1005 20:55:11.136709 6140 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1005 20:55:11.136749 6140 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1005 20:55:11.138444 6140 factory.go:656] Stopping watch factory\\\\nI1005 20:55:11.148575 6140 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1005 20:55:11.148612 6140 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1005 20:55:11.148701 6140 ovnkube.go:599] Stopped ovnkube\\\\nI1005 20:55:11.148748 6140 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1005 20:55:11.148871 6140 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2acf9637fa25e5d3ffeb05224c11218c47defc679568fb928a498869613de4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35ebeb7a87d5d47b6567917d38eca38a3e622c497fba58971296c4609c287423\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35ebeb7a87d5d47b6567917d38eca38a3e622c497fba58971296c4609c287423\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7rhps\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:12Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:13 crc kubenswrapper[4754]: I1005 20:55:13.006577 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cn76s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02d5e3f9-73c2-4496-9aca-0787184aef19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7ba2b076839f4a5f586d2e6c4f4d7188254128b782bfd2180e22bf3ea5ffdd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpx5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cn76s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:13Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:13 crc kubenswrapper[4754]: I1005 20:55:13.022751 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ss7lm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f9ef817-0622-4185-a7fb-57221690aa75\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ba613b0d7753a003c572c3733ed394707c45b6f8072cb7aa88f3c7312b8882b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wpwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:55:02Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ss7lm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:13Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:13 crc kubenswrapper[4754]: I1005 20:55:13.039124 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"934bcd66-9c03-48b9-82dd-5093338f9c8f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9042df2f43243d3d79ba9a3940f8dd4f6c7a89ef9e3510e81fe60b1944d7e83b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5366bbf46a91667355f43f5df7335a536b395cda788b90b35ef18e2b53bc4c02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://600bc3bb551b747cf2f18e26745883232d7a37ac5877fad0c67100e52d7a43aa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78f74982180048eb50b9be79fa72c04f62f925b54fe511920f37e6407f864631\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:13Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:13 crc kubenswrapper[4754]: I1005 20:55:13.044774 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:13 crc kubenswrapper[4754]: I1005 20:55:13.044842 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:13 crc kubenswrapper[4754]: I1005 20:55:13.044864 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:13 crc kubenswrapper[4754]: I1005 20:55:13.044891 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:13 crc kubenswrapper[4754]: I1005 20:55:13.044910 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:13Z","lastTransitionTime":"2025-10-05T20:55:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:13 crc kubenswrapper[4754]: I1005 20:55:13.061557 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:13Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:13 crc kubenswrapper[4754]: I1005 20:55:13.066371 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/15afc2a1-5cb7-41c6-954b-446c36822f50-env-overrides\") pod \"ovnkube-control-plane-749d76644c-rddkl\" (UID: \"15afc2a1-5cb7-41c6-954b-446c36822f50\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rddkl" Oct 05 20:55:13 crc kubenswrapper[4754]: I1005 20:55:13.066438 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/15afc2a1-5cb7-41c6-954b-446c36822f50-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-rddkl\" (UID: \"15afc2a1-5cb7-41c6-954b-446c36822f50\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rddkl" Oct 05 20:55:13 crc kubenswrapper[4754]: I1005 20:55:13.066486 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqv8k\" (UniqueName: \"kubernetes.io/projected/15afc2a1-5cb7-41c6-954b-446c36822f50-kube-api-access-cqv8k\") pod \"ovnkube-control-plane-749d76644c-rddkl\" (UID: \"15afc2a1-5cb7-41c6-954b-446c36822f50\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rddkl" Oct 05 20:55:13 crc kubenswrapper[4754]: I1005 20:55:13.066708 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/15afc2a1-5cb7-41c6-954b-446c36822f50-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-rddkl\" (UID: \"15afc2a1-5cb7-41c6-954b-446c36822f50\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rddkl" Oct 05 20:55:13 crc kubenswrapper[4754]: I1005 20:55:13.067427 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/15afc2a1-5cb7-41c6-954b-446c36822f50-env-overrides\") pod \"ovnkube-control-plane-749d76644c-rddkl\" (UID: \"15afc2a1-5cb7-41c6-954b-446c36822f50\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rddkl" Oct 05 20:55:13 crc kubenswrapper[4754]: I1005 20:55:13.067943 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/15afc2a1-5cb7-41c6-954b-446c36822f50-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-rddkl\" (UID: \"15afc2a1-5cb7-41c6-954b-446c36822f50\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rddkl" Oct 05 20:55:13 crc kubenswrapper[4754]: I1005 20:55:13.076715 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/15afc2a1-5cb7-41c6-954b-446c36822f50-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-rddkl\" (UID: \"15afc2a1-5cb7-41c6-954b-446c36822f50\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rddkl" Oct 05 20:55:13 crc kubenswrapper[4754]: I1005 20:55:13.082294 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9ad070bb699b643d29b82bcc1126ce3b585f7492f5ae2817fa491b309b1691d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:13Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:13 crc kubenswrapper[4754]: I1005 20:55:13.095762 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-57sr4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e854c7b9-1229-4da4-80e9-fea6479459b8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceb577e1055d84007936a50d0b3d54b9ae0299a1fc1a6618b37163a6d3fbb9cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vw2ps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:58Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-57sr4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:13Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:13 crc kubenswrapper[4754]: I1005 20:55:13.098847 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqv8k\" (UniqueName: \"kubernetes.io/projected/15afc2a1-5cb7-41c6-954b-446c36822f50-kube-api-access-cqv8k\") pod \"ovnkube-control-plane-749d76644c-rddkl\" (UID: \"15afc2a1-5cb7-41c6-954b-446c36822f50\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rddkl" Oct 05 20:55:13 crc kubenswrapper[4754]: I1005 20:55:13.117693 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rddkl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"15afc2a1-5cb7-41c6-954b-446c36822f50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqv8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqv8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:55:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rddkl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:13Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:13 crc kubenswrapper[4754]: I1005 20:55:13.134252 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7506d496-65ff-41dc-8230-24c8ee653a2a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33da5a6bb3cd3b042c193bc571e6622f2f87f27e560cdd0bbb9be31e2010d707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df6d700a947d19c84b4f65fd2dd2f8d1f4f1f80fa9b4800c2acfa8dae3912639\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35ca8ee77a0e58cf969bd67b277a3392614c4e73b963e6f6002097709370715d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8907b9b69b249c34ec6db1e460d641b31711b75d6d31d4caafd867339e6571d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8907b9b69b249c34ec6db1e460d641b31711b75d6d31d4caafd867339e6571d0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1005 20:54:52.589115 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 20:54:52.590181 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1304624737/tls.crt::/tmp/serving-cert-1304624737/tls.key\\\\\\\"\\\\nI1005 20:54:58.070544 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1005 20:54:58.090352 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1005 20:54:58.090382 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1005 20:54:58.090412 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1005 20:54:58.090418 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1005 20:54:58.144656 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1005 20:54:58.144688 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 20:54:58.144695 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 20:54:58.144699 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1005 20:54:58.144702 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1005 20:54:58.144705 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1005 20:54:58.144708 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1005 20:54:58.145024 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1005 20:54:58.148112 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4efb34f3493cc8854bc4cff6b29f63973c52b754404d7a21daec4e1f612d327d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54c568b637f3737667433047b956c7eb22fdf0f9301acdfd3b49fb84c39de6ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54c568b637f3737667433047b956c7eb22fdf0f9301acdfd3b49fb84c39de6ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:13Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:13 crc kubenswrapper[4754]: I1005 20:55:13.148455 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:13 crc kubenswrapper[4754]: I1005 20:55:13.148853 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:13 crc kubenswrapper[4754]: I1005 20:55:13.148861 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:13 crc kubenswrapper[4754]: I1005 20:55:13.148878 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:13 crc kubenswrapper[4754]: I1005 20:55:13.148888 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:13Z","lastTransitionTime":"2025-10-05T20:55:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:13 crc kubenswrapper[4754]: I1005 20:55:13.158935 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c76e6e4740035a6ca54191a770c4516e28101aae69afbb89739174c2f076ef5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9eededc0cb1af94fb7b5daba08785f6b077a85e9cb5ddd8b9f397b658194f61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:13Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:13 crc kubenswrapper[4754]: I1005 20:55:13.179377 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tskt7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a68e9c89-465b-4c01-a470-e21a480ee465\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba6edba14cf321b4cd2f7bb44067a3ecf49f1594e76c0197e807ff3daa45e9b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d14ec99610d2167b484bd7712af7bb6d993d26c74b2828da1c89314f16fda22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d14ec99610d2167b484bd7712af7bb6d993d26c74b2828da1c89314f16fda22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f6f08dd853b25881e2f3c66206aa80d0c493518284217780aa40ac4b69f0c0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f6f08dd853b25881e2f3c66206aa80d0c493518284217780aa40ac4b69f0c0b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://71026d00ec1a3d91b019deb3c9f43ee52fb05b8f2d6374e1a5b762c9e5c1b10d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71026d00ec1a3d91b019deb3c9f43ee52fb05b8f2d6374e1a5b762c9e5c1b10d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b9910d2470acb16ebb436d7030b3161061607cb2a5ef50d482d8d5c4f48a8ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b9910d2470acb16ebb436d7030b3161061607cb2a5ef50d482d8d5c4f48a8ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0a93b61eacc55111995e69c15f89a210b8aa15468ac9ac053b7cf6b007d301e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b0a93b61eacc55111995e69c15f89a210b8aa15468ac9ac053b7cf6b007d301e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cafb3208e7e3bcdd01898a68fc5bbf4cb52ca0800b879b77a60bb102794257f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cafb3208e7e3bcdd01898a68fc5bbf4cb52ca0800b879b77a60bb102794257f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tskt7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:13Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:13 crc kubenswrapper[4754]: I1005 20:55:13.196791 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be95a413-6a59-45b4-84b7-b43cae694a26\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10b211cf6e47dbd1ee42c82e49de0a13a45581b92a0c4f6bbba0326c7149e41d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2ggg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4bb8c5903e8b70dd19eebb111097d96f725e17de44e06c3e74c067376f6c7360\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2ggg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-b2h9k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:13Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:13 crc kubenswrapper[4754]: I1005 20:55:13.212694 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7rhps_b44790fb-fecb-4ec3-9816-8e23dd72ab33/ovnkube-controller/1.log" Oct 05 20:55:13 crc kubenswrapper[4754]: I1005 20:55:13.217479 4754 scope.go:117] "RemoveContainer" containerID="2cbed906d679176209043edd5dcc649f5ba7d5848a10cfd09efa9e656dcacfef" Oct 05 20:55:13 crc kubenswrapper[4754]: E1005 20:55:13.217672 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-7rhps_openshift-ovn-kubernetes(b44790fb-fecb-4ec3-9816-8e23dd72ab33)\"" pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" podUID="b44790fb-fecb-4ec3-9816-8e23dd72ab33" Oct 05 20:55:13 crc kubenswrapper[4754]: I1005 20:55:13.218147 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rddkl" Oct 05 20:55:13 crc kubenswrapper[4754]: I1005 20:55:13.220579 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Oct 05 20:55:13 crc kubenswrapper[4754]: I1005 20:55:13.222240 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"e32c2154f46ba10c509fe40c357304aa8e82d04e9906bbaf2ccecd47b42279e6"} Oct 05 20:55:13 crc kubenswrapper[4754]: I1005 20:55:13.222783 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 05 20:55:13 crc kubenswrapper[4754]: I1005 20:55:13.233529 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e380f9b2a0e319dc325c54d1247a6db98346292548fa688cb41a2ca1b3048419\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:13Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:13 crc kubenswrapper[4754]: I1005 20:55:13.254398 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:13 crc kubenswrapper[4754]: I1005 20:55:13.254471 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:13 crc kubenswrapper[4754]: I1005 20:55:13.254485 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:13 crc kubenswrapper[4754]: I1005 20:55:13.254526 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:13 crc kubenswrapper[4754]: I1005 20:55:13.254543 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:13Z","lastTransitionTime":"2025-10-05T20:55:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:13 crc kubenswrapper[4754]: I1005 20:55:13.257476 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:13Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:13 crc kubenswrapper[4754]: I1005 20:55:13.285907 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7506d496-65ff-41dc-8230-24c8ee653a2a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33da5a6bb3cd3b042c193bc571e6622f2f87f27e560cdd0bbb9be31e2010d707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df6d700a947d19c84b4f65fd2dd2f8d1f4f1f80fa9b4800c2acfa8dae3912639\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35ca8ee77a0e58cf969bd67b277a3392614c4e73b963e6f6002097709370715d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e32c2154f46ba10c509fe40c357304aa8e82d04e9906bbaf2ccecd47b42279e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8907b9b69b249c34ec6db1e460d641b31711b75d6d31d4caafd867339e6571d0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1005 20:54:52.589115 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 20:54:52.590181 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1304624737/tls.crt::/tmp/serving-cert-1304624737/tls.key\\\\\\\"\\\\nI1005 20:54:58.070544 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1005 20:54:58.090352 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1005 20:54:58.090382 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1005 20:54:58.090412 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1005 20:54:58.090418 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1005 20:54:58.144656 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1005 20:54:58.144688 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 20:54:58.144695 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 20:54:58.144699 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1005 20:54:58.144702 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1005 20:54:58.144705 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1005 20:54:58.144708 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1005 20:54:58.145024 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1005 20:54:58.148112 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4efb34f3493cc8854bc4cff6b29f63973c52b754404d7a21daec4e1f612d327d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54c568b637f3737667433047b956c7eb22fdf0f9301acdfd3b49fb84c39de6ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54c568b637f3737667433047b956c7eb22fdf0f9301acdfd3b49fb84c39de6ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:13Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:13 crc kubenswrapper[4754]: I1005 20:55:13.301867 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c76e6e4740035a6ca54191a770c4516e28101aae69afbb89739174c2f076ef5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9eededc0cb1af94fb7b5daba08785f6b077a85e9cb5ddd8b9f397b658194f61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:13Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:13 crc kubenswrapper[4754]: I1005 20:55:13.322045 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tskt7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a68e9c89-465b-4c01-a470-e21a480ee465\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba6edba14cf321b4cd2f7bb44067a3ecf49f1594e76c0197e807ff3daa45e9b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d14ec99610d2167b484bd7712af7bb6d993d26c74b2828da1c89314f16fda22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d14ec99610d2167b484bd7712af7bb6d993d26c74b2828da1c89314f16fda22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f6f08dd853b25881e2f3c66206aa80d0c493518284217780aa40ac4b69f0c0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f6f08dd853b25881e2f3c66206aa80d0c493518284217780aa40ac4b69f0c0b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://71026d00ec1a3d91b019deb3c9f43ee52fb05b8f2d6374e1a5b762c9e5c1b10d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71026d00ec1a3d91b019deb3c9f43ee52fb05b8f2d6374e1a5b762c9e5c1b10d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b9910d2470acb16ebb436d7030b3161061607cb2a5ef50d482d8d5c4f48a8ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b9910d2470acb16ebb436d7030b3161061607cb2a5ef50d482d8d5c4f48a8ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0a93b61eacc55111995e69c15f89a210b8aa15468ac9ac053b7cf6b007d301e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b0a93b61eacc55111995e69c15f89a210b8aa15468ac9ac053b7cf6b007d301e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cafb3208e7e3bcdd01898a68fc5bbf4cb52ca0800b879b77a60bb102794257f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cafb3208e7e3bcdd01898a68fc5bbf4cb52ca0800b879b77a60bb102794257f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tskt7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:13Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:13 crc kubenswrapper[4754]: I1005 20:55:13.337853 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be95a413-6a59-45b4-84b7-b43cae694a26\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10b211cf6e47dbd1ee42c82e49de0a13a45581b92a0c4f6bbba0326c7149e41d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2ggg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4bb8c5903e8b70dd19eebb111097d96f725e17de44e06c3e74c067376f6c7360\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2ggg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-b2h9k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:13Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:13 crc kubenswrapper[4754]: I1005 20:55:13.358943 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rddkl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"15afc2a1-5cb7-41c6-954b-446c36822f50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqv8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqv8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:55:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rddkl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:13Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:13 crc kubenswrapper[4754]: I1005 20:55:13.364751 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:13 crc kubenswrapper[4754]: I1005 20:55:13.364805 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:13 crc kubenswrapper[4754]: I1005 20:55:13.364824 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:13 crc kubenswrapper[4754]: I1005 20:55:13.364850 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:13 crc kubenswrapper[4754]: I1005 20:55:13.364870 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:13Z","lastTransitionTime":"2025-10-05T20:55:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:13 crc kubenswrapper[4754]: I1005 20:55:13.387775 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e380f9b2a0e319dc325c54d1247a6db98346292548fa688cb41a2ca1b3048419\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:13Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:13 crc kubenswrapper[4754]: I1005 20:55:13.408291 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:13Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:13 crc kubenswrapper[4754]: I1005 20:55:13.442013 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9309179-b92a-4fa4-9fba-747b8f911936\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d969d3d50d6ad280fc6654df9441bab238bb538f277ed6d1273779813ffbaaed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://800a731ed8ef57714b2318af14054ef23f0368936630887a093fc77167b33d12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8cd9b08e83dc174413cc37891c14562816787028471a4201680c86dd620b7a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93b729f3f7f993db0d6bc3ba60c04930aada3cef8272ffcc3c6a8a242a9b1475\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dde1519111fc3af5d14463fdb58d4f5c62a7045e7de98c60981c7e4da37b0e7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ea160a5b37aea7f2970ae4dd2614b09d3dcf5dcbd1fc103b9997f8ecc3a804c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ea160a5b37aea7f2970ae4dd2614b09d3dcf5dcbd1fc103b9997f8ecc3a804c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://17ae6aa6c53e7dd061f70b41428ef50bbc7ffcc65352fad1a4a9d3c69dbd2ba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17ae6aa6c53e7dd061f70b41428ef50bbc7ffcc65352fad1a4a9d3c69dbd2ba9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://ad4cd69df7608c7d87d839fe09d4557c674c7efbc4730f4e4d3daaf9a26f5aee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad4cd69df7608c7d87d839fe09d4557c674c7efbc4730f4e4d3daaf9a26f5aee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:36Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:13Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:13 crc kubenswrapper[4754]: I1005 20:55:13.458108 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:13Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:13 crc kubenswrapper[4754]: I1005 20:55:13.467870 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:13 crc kubenswrapper[4754]: I1005 20:55:13.467914 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:13 crc kubenswrapper[4754]: I1005 20:55:13.467926 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:13 crc kubenswrapper[4754]: I1005 20:55:13.467947 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:13 crc kubenswrapper[4754]: I1005 20:55:13.467961 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:13Z","lastTransitionTime":"2025-10-05T20:55:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:13 crc kubenswrapper[4754]: I1005 20:55:13.477601 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b44790fb-fecb-4ec3-9816-8e23dd72ab33\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a7dc53e7f96e48213d724a2cb4859514df925005644108ae3a0c33b14dcf52d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfdd8b1e542ab676576c311e2923fb02861f92de91fe1f4b3e4b968c549c375d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3a63451ffaa14cacc3b78d11340f4ea81445fa3af96178e9b172d063d9697d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3cfd3f2274a8a46fa36660e94ac6c9c298c791283f12d52330c5594ac131998\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f626cb108716da0b595844ea0251b7a04d49b5908a1d6dd0aef5cb2d402c4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://822223f515b5add6d71d193fd5bc77df029b8d37a1cc07be4145b53c7226bf61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cbed906d679176209043edd5dcc649f5ba7d5848a10cfd09efa9e656dcacfef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cbed906d679176209043edd5dcc649f5ba7d5848a10cfd09efa9e656dcacfef\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-05T20:55:11Z\\\",\\\"message\\\":\\\"gressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1005 20:55:11.136172 6140 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1005 20:55:11.136316 6140 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1005 20:55:11.136624 6140 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1005 20:55:11.136664 6140 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1005 20:55:11.136709 6140 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1005 20:55:11.136749 6140 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1005 20:55:11.138444 6140 factory.go:656] Stopping watch factory\\\\nI1005 20:55:11.148575 6140 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1005 20:55:11.148612 6140 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1005 20:55:11.148701 6140 ovnkube.go:599] Stopped ovnkube\\\\nI1005 20:55:11.148748 6140 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1005 20:55:11.148871 6140 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:10Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-7rhps_openshift-ovn-kubernetes(b44790fb-fecb-4ec3-9816-8e23dd72ab33)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2acf9637fa25e5d3ffeb05224c11218c47defc679568fb928a498869613de4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35ebeb7a87d5d47b6567917d38eca38a3e622c497fba58971296c4609c287423\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35ebeb7a87d5d47b6567917d38eca38a3e622c497fba58971296c4609c287423\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7rhps\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:13Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:13 crc kubenswrapper[4754]: I1005 20:55:13.494764 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"934bcd66-9c03-48b9-82dd-5093338f9c8f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9042df2f43243d3d79ba9a3940f8dd4f6c7a89ef9e3510e81fe60b1944d7e83b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5366bbf46a91667355f43f5df7335a536b395cda788b90b35ef18e2b53bc4c02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://600bc3bb551b747cf2f18e26745883232d7a37ac5877fad0c67100e52d7a43aa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78f74982180048eb50b9be79fa72c04f62f925b54fe511920f37e6407f864631\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:13Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:13 crc kubenswrapper[4754]: I1005 20:55:13.508170 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:13Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:13 crc kubenswrapper[4754]: I1005 20:55:13.522235 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9ad070bb699b643d29b82bcc1126ce3b585f7492f5ae2817fa491b309b1691d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:13Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:13 crc kubenswrapper[4754]: I1005 20:55:13.543237 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-57sr4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e854c7b9-1229-4da4-80e9-fea6479459b8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceb577e1055d84007936a50d0b3d54b9ae0299a1fc1a6618b37163a6d3fbb9cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vw2ps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:58Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-57sr4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:13Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:13 crc kubenswrapper[4754]: I1005 20:55:13.566861 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cn76s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02d5e3f9-73c2-4496-9aca-0787184aef19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7ba2b076839f4a5f586d2e6c4f4d7188254128b782bfd2180e22bf3ea5ffdd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpx5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cn76s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:13Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:13 crc kubenswrapper[4754]: I1005 20:55:13.570980 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:13 crc kubenswrapper[4754]: I1005 20:55:13.571051 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:13 crc kubenswrapper[4754]: I1005 20:55:13.571074 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:13 crc kubenswrapper[4754]: I1005 20:55:13.571100 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:13 crc kubenswrapper[4754]: I1005 20:55:13.571119 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:13Z","lastTransitionTime":"2025-10-05T20:55:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:13 crc kubenswrapper[4754]: I1005 20:55:13.585477 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ss7lm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f9ef817-0622-4185-a7fb-57221690aa75\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ba613b0d7753a003c572c3733ed394707c45b6f8072cb7aa88f3c7312b8882b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wpwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:55:02Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ss7lm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:13Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:13 crc kubenswrapper[4754]: I1005 20:55:13.653964 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-nwrnt"] Oct 05 20:55:13 crc kubenswrapper[4754]: I1005 20:55:13.654758 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nwrnt" Oct 05 20:55:13 crc kubenswrapper[4754]: E1005 20:55:13.654874 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nwrnt" podUID="7dc728ea-2601-44f0-bcab-2913f034007d" Oct 05 20:55:13 crc kubenswrapper[4754]: I1005 20:55:13.674284 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:13 crc kubenswrapper[4754]: I1005 20:55:13.674330 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:13 crc kubenswrapper[4754]: I1005 20:55:13.674351 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:13 crc kubenswrapper[4754]: I1005 20:55:13.674397 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:13 crc kubenswrapper[4754]: I1005 20:55:13.674416 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:13Z","lastTransitionTime":"2025-10-05T20:55:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:13 crc kubenswrapper[4754]: I1005 20:55:13.681993 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e380f9b2a0e319dc325c54d1247a6db98346292548fa688cb41a2ca1b3048419\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:13Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:13 crc kubenswrapper[4754]: I1005 20:55:13.712207 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:13Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:13 crc kubenswrapper[4754]: I1005 20:55:13.736231 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9309179-b92a-4fa4-9fba-747b8f911936\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d969d3d50d6ad280fc6654df9441bab238bb538f277ed6d1273779813ffbaaed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://800a731ed8ef57714b2318af14054ef23f0368936630887a093fc77167b33d12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8cd9b08e83dc174413cc37891c14562816787028471a4201680c86dd620b7a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93b729f3f7f993db0d6bc3ba60c04930aada3cef8272ffcc3c6a8a242a9b1475\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dde1519111fc3af5d14463fdb58d4f5c62a7045e7de98c60981c7e4da37b0e7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ea160a5b37aea7f2970ae4dd2614b09d3dcf5dcbd1fc103b9997f8ecc3a804c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ea160a5b37aea7f2970ae4dd2614b09d3dcf5dcbd1fc103b9997f8ecc3a804c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://17ae6aa6c53e7dd061f70b41428ef50bbc7ffcc65352fad1a4a9d3c69dbd2ba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17ae6aa6c53e7dd061f70b41428ef50bbc7ffcc65352fad1a4a9d3c69dbd2ba9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://ad4cd69df7608c7d87d839fe09d4557c674c7efbc4730f4e4d3daaf9a26f5aee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad4cd69df7608c7d87d839fe09d4557c674c7efbc4730f4e4d3daaf9a26f5aee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:36Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:13Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:13 crc kubenswrapper[4754]: I1005 20:55:13.749954 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:13Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:13 crc kubenswrapper[4754]: I1005 20:55:13.768079 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b44790fb-fecb-4ec3-9816-8e23dd72ab33\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a7dc53e7f96e48213d724a2cb4859514df925005644108ae3a0c33b14dcf52d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfdd8b1e542ab676576c311e2923fb02861f92de91fe1f4b3e4b968c549c375d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3a63451ffaa14cacc3b78d11340f4ea81445fa3af96178e9b172d063d9697d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3cfd3f2274a8a46fa36660e94ac6c9c298c791283f12d52330c5594ac131998\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f626cb108716da0b595844ea0251b7a04d49b5908a1d6dd0aef5cb2d402c4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://822223f515b5add6d71d193fd5bc77df029b8d37a1cc07be4145b53c7226bf61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cbed906d679176209043edd5dcc649f5ba7d5848a10cfd09efa9e656dcacfef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cbed906d679176209043edd5dcc649f5ba7d5848a10cfd09efa9e656dcacfef\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-05T20:55:11Z\\\",\\\"message\\\":\\\"gressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1005 20:55:11.136172 6140 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1005 20:55:11.136316 6140 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1005 20:55:11.136624 6140 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1005 20:55:11.136664 6140 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1005 20:55:11.136709 6140 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1005 20:55:11.136749 6140 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1005 20:55:11.138444 6140 factory.go:656] Stopping watch factory\\\\nI1005 20:55:11.148575 6140 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1005 20:55:11.148612 6140 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1005 20:55:11.148701 6140 ovnkube.go:599] Stopped ovnkube\\\\nI1005 20:55:11.148748 6140 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1005 20:55:11.148871 6140 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:10Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-7rhps_openshift-ovn-kubernetes(b44790fb-fecb-4ec3-9816-8e23dd72ab33)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2acf9637fa25e5d3ffeb05224c11218c47defc679568fb928a498869613de4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35ebeb7a87d5d47b6567917d38eca38a3e622c497fba58971296c4609c287423\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35ebeb7a87d5d47b6567917d38eca38a3e622c497fba58971296c4609c287423\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7rhps\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:13Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:13 crc kubenswrapper[4754]: I1005 20:55:13.777212 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:13 crc kubenswrapper[4754]: I1005 20:55:13.777234 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:13 crc kubenswrapper[4754]: I1005 20:55:13.777241 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:13 crc kubenswrapper[4754]: I1005 20:55:13.777257 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:13 crc kubenswrapper[4754]: I1005 20:55:13.777268 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:13Z","lastTransitionTime":"2025-10-05T20:55:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:13 crc kubenswrapper[4754]: I1005 20:55:13.781917 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7dc728ea-2601-44f0-bcab-2913f034007d-metrics-certs\") pod \"network-metrics-daemon-nwrnt\" (UID: \"7dc728ea-2601-44f0-bcab-2913f034007d\") " pod="openshift-multus/network-metrics-daemon-nwrnt" Oct 05 20:55:13 crc kubenswrapper[4754]: I1005 20:55:13.781946 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-846xq\" (UniqueName: \"kubernetes.io/projected/7dc728ea-2601-44f0-bcab-2913f034007d-kube-api-access-846xq\") pod \"network-metrics-daemon-nwrnt\" (UID: \"7dc728ea-2601-44f0-bcab-2913f034007d\") " pod="openshift-multus/network-metrics-daemon-nwrnt" Oct 05 20:55:13 crc kubenswrapper[4754]: I1005 20:55:13.795640 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cn76s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02d5e3f9-73c2-4496-9aca-0787184aef19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7ba2b076839f4a5f586d2e6c4f4d7188254128b782bfd2180e22bf3ea5ffdd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpx5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cn76s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:13Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:13 crc kubenswrapper[4754]: I1005 20:55:13.807290 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ss7lm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f9ef817-0622-4185-a7fb-57221690aa75\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ba613b0d7753a003c572c3733ed394707c45b6f8072cb7aa88f3c7312b8882b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wpwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:55:02Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ss7lm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:13Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:13 crc kubenswrapper[4754]: I1005 20:55:13.818395 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-nwrnt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7dc728ea-2601-44f0-bcab-2913f034007d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-846xq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-846xq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:55:13Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-nwrnt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:13Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:13 crc kubenswrapper[4754]: I1005 20:55:13.830982 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"934bcd66-9c03-48b9-82dd-5093338f9c8f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9042df2f43243d3d79ba9a3940f8dd4f6c7a89ef9e3510e81fe60b1944d7e83b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5366bbf46a91667355f43f5df7335a536b395cda788b90b35ef18e2b53bc4c02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://600bc3bb551b747cf2f18e26745883232d7a37ac5877fad0c67100e52d7a43aa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78f74982180048eb50b9be79fa72c04f62f925b54fe511920f37e6407f864631\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:13Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:13 crc kubenswrapper[4754]: I1005 20:55:13.844542 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:13Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:13 crc kubenswrapper[4754]: I1005 20:55:13.856487 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9ad070bb699b643d29b82bcc1126ce3b585f7492f5ae2817fa491b309b1691d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:13Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:13 crc kubenswrapper[4754]: I1005 20:55:13.864113 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-57sr4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e854c7b9-1229-4da4-80e9-fea6479459b8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceb577e1055d84007936a50d0b3d54b9ae0299a1fc1a6618b37163a6d3fbb9cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vw2ps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:58Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-57sr4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:13Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:13 crc kubenswrapper[4754]: I1005 20:55:13.872981 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rddkl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"15afc2a1-5cb7-41c6-954b-446c36822f50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqv8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqv8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:55:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rddkl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:13Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:13 crc kubenswrapper[4754]: I1005 20:55:13.879908 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:13 crc kubenswrapper[4754]: I1005 20:55:13.879931 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:13 crc kubenswrapper[4754]: I1005 20:55:13.879939 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:13 crc kubenswrapper[4754]: I1005 20:55:13.879954 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:13 crc kubenswrapper[4754]: I1005 20:55:13.879965 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:13Z","lastTransitionTime":"2025-10-05T20:55:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:13 crc kubenswrapper[4754]: I1005 20:55:13.882785 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-846xq\" (UniqueName: \"kubernetes.io/projected/7dc728ea-2601-44f0-bcab-2913f034007d-kube-api-access-846xq\") pod \"network-metrics-daemon-nwrnt\" (UID: \"7dc728ea-2601-44f0-bcab-2913f034007d\") " pod="openshift-multus/network-metrics-daemon-nwrnt" Oct 05 20:55:13 crc kubenswrapper[4754]: I1005 20:55:13.882816 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7dc728ea-2601-44f0-bcab-2913f034007d-metrics-certs\") pod \"network-metrics-daemon-nwrnt\" (UID: \"7dc728ea-2601-44f0-bcab-2913f034007d\") " pod="openshift-multus/network-metrics-daemon-nwrnt" Oct 05 20:55:13 crc kubenswrapper[4754]: E1005 20:55:13.882975 4754 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 05 20:55:13 crc kubenswrapper[4754]: E1005 20:55:13.883026 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7dc728ea-2601-44f0-bcab-2913f034007d-metrics-certs podName:7dc728ea-2601-44f0-bcab-2913f034007d nodeName:}" failed. No retries permitted until 2025-10-05 20:55:14.383008841 +0000 UTC m=+38.287127551 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/7dc728ea-2601-44f0-bcab-2913f034007d-metrics-certs") pod "network-metrics-daemon-nwrnt" (UID: "7dc728ea-2601-44f0-bcab-2913f034007d") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 05 20:55:13 crc kubenswrapper[4754]: I1005 20:55:13.884510 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7506d496-65ff-41dc-8230-24c8ee653a2a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33da5a6bb3cd3b042c193bc571e6622f2f87f27e560cdd0bbb9be31e2010d707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df6d700a947d19c84b4f65fd2dd2f8d1f4f1f80fa9b4800c2acfa8dae3912639\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35ca8ee77a0e58cf969bd67b277a3392614c4e73b963e6f6002097709370715d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e32c2154f46ba10c509fe40c357304aa8e82d04e9906bbaf2ccecd47b42279e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8907b9b69b249c34ec6db1e460d641b31711b75d6d31d4caafd867339e6571d0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1005 20:54:52.589115 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 20:54:52.590181 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1304624737/tls.crt::/tmp/serving-cert-1304624737/tls.key\\\\\\\"\\\\nI1005 20:54:58.070544 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1005 20:54:58.090352 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1005 20:54:58.090382 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1005 20:54:58.090412 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1005 20:54:58.090418 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1005 20:54:58.144656 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1005 20:54:58.144688 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 20:54:58.144695 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 20:54:58.144699 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1005 20:54:58.144702 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1005 20:54:58.144705 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1005 20:54:58.144708 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1005 20:54:58.145024 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1005 20:54:58.148112 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4efb34f3493cc8854bc4cff6b29f63973c52b754404d7a21daec4e1f612d327d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54c568b637f3737667433047b956c7eb22fdf0f9301acdfd3b49fb84c39de6ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54c568b637f3737667433047b956c7eb22fdf0f9301acdfd3b49fb84c39de6ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:13Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:13 crc kubenswrapper[4754]: I1005 20:55:13.897761 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c76e6e4740035a6ca54191a770c4516e28101aae69afbb89739174c2f076ef5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9eededc0cb1af94fb7b5daba08785f6b077a85e9cb5ddd8b9f397b658194f61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:13Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:13 crc kubenswrapper[4754]: I1005 20:55:13.901488 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-846xq\" (UniqueName: \"kubernetes.io/projected/7dc728ea-2601-44f0-bcab-2913f034007d-kube-api-access-846xq\") pod \"network-metrics-daemon-nwrnt\" (UID: \"7dc728ea-2601-44f0-bcab-2913f034007d\") " pod="openshift-multus/network-metrics-daemon-nwrnt" Oct 05 20:55:13 crc kubenswrapper[4754]: I1005 20:55:13.918144 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tskt7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a68e9c89-465b-4c01-a470-e21a480ee465\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba6edba14cf321b4cd2f7bb44067a3ecf49f1594e76c0197e807ff3daa45e9b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d14ec99610d2167b484bd7712af7bb6d993d26c74b2828da1c89314f16fda22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d14ec99610d2167b484bd7712af7bb6d993d26c74b2828da1c89314f16fda22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f6f08dd853b25881e2f3c66206aa80d0c493518284217780aa40ac4b69f0c0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f6f08dd853b25881e2f3c66206aa80d0c493518284217780aa40ac4b69f0c0b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://71026d00ec1a3d91b019deb3c9f43ee52fb05b8f2d6374e1a5b762c9e5c1b10d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71026d00ec1a3d91b019deb3c9f43ee52fb05b8f2d6374e1a5b762c9e5c1b10d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b9910d2470acb16ebb436d7030b3161061607cb2a5ef50d482d8d5c4f48a8ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b9910d2470acb16ebb436d7030b3161061607cb2a5ef50d482d8d5c4f48a8ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0a93b61eacc55111995e69c15f89a210b8aa15468ac9ac053b7cf6b007d301e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b0a93b61eacc55111995e69c15f89a210b8aa15468ac9ac053b7cf6b007d301e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cafb3208e7e3bcdd01898a68fc5bbf4cb52ca0800b879b77a60bb102794257f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cafb3208e7e3bcdd01898a68fc5bbf4cb52ca0800b879b77a60bb102794257f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tskt7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:13Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:13 crc kubenswrapper[4754]: I1005 20:55:13.929663 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be95a413-6a59-45b4-84b7-b43cae694a26\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10b211cf6e47dbd1ee42c82e49de0a13a45581b92a0c4f6bbba0326c7149e41d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2ggg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4bb8c5903e8b70dd19eebb111097d96f725e17de44e06c3e74c067376f6c7360\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2ggg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-b2h9k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:13Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:13 crc kubenswrapper[4754]: I1005 20:55:13.983459 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:13 crc kubenswrapper[4754]: I1005 20:55:13.983524 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:13 crc kubenswrapper[4754]: I1005 20:55:13.983543 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:13 crc kubenswrapper[4754]: I1005 20:55:13.983593 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:13 crc kubenswrapper[4754]: I1005 20:55:13.983609 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:13Z","lastTransitionTime":"2025-10-05T20:55:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:14 crc kubenswrapper[4754]: I1005 20:55:14.086794 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:14 crc kubenswrapper[4754]: I1005 20:55:14.086842 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:14 crc kubenswrapper[4754]: I1005 20:55:14.086860 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:14 crc kubenswrapper[4754]: I1005 20:55:14.086883 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:14 crc kubenswrapper[4754]: I1005 20:55:14.086900 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:14Z","lastTransitionTime":"2025-10-05T20:55:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:14 crc kubenswrapper[4754]: I1005 20:55:14.190367 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:14 crc kubenswrapper[4754]: I1005 20:55:14.190427 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:14 crc kubenswrapper[4754]: I1005 20:55:14.190439 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:14 crc kubenswrapper[4754]: I1005 20:55:14.190463 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:14 crc kubenswrapper[4754]: I1005 20:55:14.190474 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:14Z","lastTransitionTime":"2025-10-05T20:55:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:14 crc kubenswrapper[4754]: I1005 20:55:14.228924 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rddkl" event={"ID":"15afc2a1-5cb7-41c6-954b-446c36822f50","Type":"ContainerStarted","Data":"2a175565f83cfcba7825006eafb0f83efb70ec3d415c2c039c6499d5f6576e89"} Oct 05 20:55:14 crc kubenswrapper[4754]: I1005 20:55:14.229030 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rddkl" event={"ID":"15afc2a1-5cb7-41c6-954b-446c36822f50","Type":"ContainerStarted","Data":"866d85ee126f8641d12f1a05fabc3de7b9392c324e9fe06ea92214155c05334c"} Oct 05 20:55:14 crc kubenswrapper[4754]: I1005 20:55:14.229062 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rddkl" event={"ID":"15afc2a1-5cb7-41c6-954b-446c36822f50","Type":"ContainerStarted","Data":"417cf2da915ea42018a3552eb889433c02fd4ceba67aa4cb53ee23657ef6230a"} Oct 05 20:55:14 crc kubenswrapper[4754]: I1005 20:55:14.258408 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9309179-b92a-4fa4-9fba-747b8f911936\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d969d3d50d6ad280fc6654df9441bab238bb538f277ed6d1273779813ffbaaed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://800a731ed8ef57714b2318af14054ef23f0368936630887a093fc77167b33d12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8cd9b08e83dc174413cc37891c14562816787028471a4201680c86dd620b7a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93b729f3f7f993db0d6bc3ba60c04930aada3cef8272ffcc3c6a8a242a9b1475\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dde1519111fc3af5d14463fdb58d4f5c62a7045e7de98c60981c7e4da37b0e7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ea160a5b37aea7f2970ae4dd2614b09d3dcf5dcbd1fc103b9997f8ecc3a804c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ea160a5b37aea7f2970ae4dd2614b09d3dcf5dcbd1fc103b9997f8ecc3a804c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://17ae6aa6c53e7dd061f70b41428ef50bbc7ffcc65352fad1a4a9d3c69dbd2ba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17ae6aa6c53e7dd061f70b41428ef50bbc7ffcc65352fad1a4a9d3c69dbd2ba9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://ad4cd69df7608c7d87d839fe09d4557c674c7efbc4730f4e4d3daaf9a26f5aee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad4cd69df7608c7d87d839fe09d4557c674c7efbc4730f4e4d3daaf9a26f5aee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:36Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:14Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:14 crc kubenswrapper[4754]: I1005 20:55:14.276599 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:14Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:14 crc kubenswrapper[4754]: I1005 20:55:14.292693 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:14 crc kubenswrapper[4754]: I1005 20:55:14.292726 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:14 crc kubenswrapper[4754]: I1005 20:55:14.292736 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:14 crc kubenswrapper[4754]: I1005 20:55:14.292751 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:14 crc kubenswrapper[4754]: I1005 20:55:14.292761 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:14Z","lastTransitionTime":"2025-10-05T20:55:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:14 crc kubenswrapper[4754]: I1005 20:55:14.301918 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b44790fb-fecb-4ec3-9816-8e23dd72ab33\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a7dc53e7f96e48213d724a2cb4859514df925005644108ae3a0c33b14dcf52d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfdd8b1e542ab676576c311e2923fb02861f92de91fe1f4b3e4b968c549c375d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3a63451ffaa14cacc3b78d11340f4ea81445fa3af96178e9b172d063d9697d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3cfd3f2274a8a46fa36660e94ac6c9c298c791283f12d52330c5594ac131998\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f626cb108716da0b595844ea0251b7a04d49b5908a1d6dd0aef5cb2d402c4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://822223f515b5add6d71d193fd5bc77df029b8d37a1cc07be4145b53c7226bf61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cbed906d679176209043edd5dcc649f5ba7d5848a10cfd09efa9e656dcacfef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cbed906d679176209043edd5dcc649f5ba7d5848a10cfd09efa9e656dcacfef\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-05T20:55:11Z\\\",\\\"message\\\":\\\"gressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1005 20:55:11.136172 6140 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1005 20:55:11.136316 6140 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1005 20:55:11.136624 6140 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1005 20:55:11.136664 6140 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1005 20:55:11.136709 6140 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1005 20:55:11.136749 6140 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1005 20:55:11.138444 6140 factory.go:656] Stopping watch factory\\\\nI1005 20:55:11.148575 6140 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1005 20:55:11.148612 6140 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1005 20:55:11.148701 6140 ovnkube.go:599] Stopped ovnkube\\\\nI1005 20:55:11.148748 6140 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1005 20:55:11.148871 6140 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:10Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-7rhps_openshift-ovn-kubernetes(b44790fb-fecb-4ec3-9816-8e23dd72ab33)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2acf9637fa25e5d3ffeb05224c11218c47defc679568fb928a498869613de4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35ebeb7a87d5d47b6567917d38eca38a3e622c497fba58971296c4609c287423\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35ebeb7a87d5d47b6567917d38eca38a3e622c497fba58971296c4609c287423\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7rhps\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:14Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:14 crc kubenswrapper[4754]: I1005 20:55:14.321180 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"934bcd66-9c03-48b9-82dd-5093338f9c8f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9042df2f43243d3d79ba9a3940f8dd4f6c7a89ef9e3510e81fe60b1944d7e83b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5366bbf46a91667355f43f5df7335a536b395cda788b90b35ef18e2b53bc4c02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://600bc3bb551b747cf2f18e26745883232d7a37ac5877fad0c67100e52d7a43aa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78f74982180048eb50b9be79fa72c04f62f925b54fe511920f37e6407f864631\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:14Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:14 crc kubenswrapper[4754]: I1005 20:55:14.341952 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:14Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:14 crc kubenswrapper[4754]: I1005 20:55:14.359487 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9ad070bb699b643d29b82bcc1126ce3b585f7492f5ae2817fa491b309b1691d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:14Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:14 crc kubenswrapper[4754]: I1005 20:55:14.372476 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-57sr4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e854c7b9-1229-4da4-80e9-fea6479459b8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceb577e1055d84007936a50d0b3d54b9ae0299a1fc1a6618b37163a6d3fbb9cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vw2ps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:58Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-57sr4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:14Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:14 crc kubenswrapper[4754]: I1005 20:55:14.387914 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7dc728ea-2601-44f0-bcab-2913f034007d-metrics-certs\") pod \"network-metrics-daemon-nwrnt\" (UID: \"7dc728ea-2601-44f0-bcab-2913f034007d\") " pod="openshift-multus/network-metrics-daemon-nwrnt" Oct 05 20:55:14 crc kubenswrapper[4754]: E1005 20:55:14.388794 4754 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 05 20:55:14 crc kubenswrapper[4754]: E1005 20:55:14.388855 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7dc728ea-2601-44f0-bcab-2913f034007d-metrics-certs podName:7dc728ea-2601-44f0-bcab-2913f034007d nodeName:}" failed. No retries permitted until 2025-10-05 20:55:15.388838337 +0000 UTC m=+39.292957047 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/7dc728ea-2601-44f0-bcab-2913f034007d-metrics-certs") pod "network-metrics-daemon-nwrnt" (UID: "7dc728ea-2601-44f0-bcab-2913f034007d") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 05 20:55:14 crc kubenswrapper[4754]: I1005 20:55:14.392475 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cn76s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02d5e3f9-73c2-4496-9aca-0787184aef19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7ba2b076839f4a5f586d2e6c4f4d7188254128b782bfd2180e22bf3ea5ffdd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpx5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cn76s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:14Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:14 crc kubenswrapper[4754]: I1005 20:55:14.395250 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:14 crc kubenswrapper[4754]: I1005 20:55:14.395285 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:14 crc kubenswrapper[4754]: I1005 20:55:14.395297 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:14 crc kubenswrapper[4754]: I1005 20:55:14.395313 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:14 crc kubenswrapper[4754]: I1005 20:55:14.395325 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:14Z","lastTransitionTime":"2025-10-05T20:55:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:14 crc kubenswrapper[4754]: I1005 20:55:14.406484 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ss7lm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f9ef817-0622-4185-a7fb-57221690aa75\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ba613b0d7753a003c572c3733ed394707c45b6f8072cb7aa88f3c7312b8882b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wpwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:55:02Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ss7lm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:14Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:14 crc kubenswrapper[4754]: I1005 20:55:14.416404 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:14 crc kubenswrapper[4754]: I1005 20:55:14.416456 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:14 crc kubenswrapper[4754]: I1005 20:55:14.416467 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:14 crc kubenswrapper[4754]: I1005 20:55:14.416484 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:14 crc kubenswrapper[4754]: I1005 20:55:14.416515 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:14Z","lastTransitionTime":"2025-10-05T20:55:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:14 crc kubenswrapper[4754]: I1005 20:55:14.418370 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-nwrnt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7dc728ea-2601-44f0-bcab-2913f034007d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-846xq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-846xq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:55:13Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-nwrnt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:14Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:14 crc kubenswrapper[4754]: E1005 20:55:14.431150 4754 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T20:55:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T20:55:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T20:55:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T20:55:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ede2b442-76b0-443a-a450-252ee8711c80\\\",\\\"systemUUID\\\":\\\"72d081dd-fe7e-42a6-9cf0-f0ee7f733ac9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:14Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:14 crc kubenswrapper[4754]: I1005 20:55:14.434141 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:14 crc kubenswrapper[4754]: I1005 20:55:14.434197 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:14 crc kubenswrapper[4754]: I1005 20:55:14.434210 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:14 crc kubenswrapper[4754]: I1005 20:55:14.434235 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:14 crc kubenswrapper[4754]: I1005 20:55:14.434249 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:14Z","lastTransitionTime":"2025-10-05T20:55:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:14 crc kubenswrapper[4754]: I1005 20:55:14.438396 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7506d496-65ff-41dc-8230-24c8ee653a2a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33da5a6bb3cd3b042c193bc571e6622f2f87f27e560cdd0bbb9be31e2010d707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df6d700a947d19c84b4f65fd2dd2f8d1f4f1f80fa9b4800c2acfa8dae3912639\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35ca8ee77a0e58cf969bd67b277a3392614c4e73b963e6f6002097709370715d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e32c2154f46ba10c509fe40c357304aa8e82d04e9906bbaf2ccecd47b42279e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8907b9b69b249c34ec6db1e460d641b31711b75d6d31d4caafd867339e6571d0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1005 20:54:52.589115 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 20:54:52.590181 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1304624737/tls.crt::/tmp/serving-cert-1304624737/tls.key\\\\\\\"\\\\nI1005 20:54:58.070544 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1005 20:54:58.090352 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1005 20:54:58.090382 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1005 20:54:58.090412 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1005 20:54:58.090418 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1005 20:54:58.144656 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1005 20:54:58.144688 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 20:54:58.144695 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 20:54:58.144699 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1005 20:54:58.144702 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1005 20:54:58.144705 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1005 20:54:58.144708 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1005 20:54:58.145024 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1005 20:54:58.148112 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4efb34f3493cc8854bc4cff6b29f63973c52b754404d7a21daec4e1f612d327d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54c568b637f3737667433047b956c7eb22fdf0f9301acdfd3b49fb84c39de6ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54c568b637f3737667433047b956c7eb22fdf0f9301acdfd3b49fb84c39de6ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:14Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:14 crc kubenswrapper[4754]: E1005 20:55:14.447113 4754 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T20:55:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T20:55:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T20:55:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T20:55:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ede2b442-76b0-443a-a450-252ee8711c80\\\",\\\"systemUUID\\\":\\\"72d081dd-fe7e-42a6-9cf0-f0ee7f733ac9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:14Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:14 crc kubenswrapper[4754]: I1005 20:55:14.451464 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:14 crc kubenswrapper[4754]: I1005 20:55:14.451560 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:14 crc kubenswrapper[4754]: I1005 20:55:14.451577 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:14 crc kubenswrapper[4754]: I1005 20:55:14.451601 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:14 crc kubenswrapper[4754]: I1005 20:55:14.451617 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:14Z","lastTransitionTime":"2025-10-05T20:55:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:14 crc kubenswrapper[4754]: I1005 20:55:14.455192 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c76e6e4740035a6ca54191a770c4516e28101aae69afbb89739174c2f076ef5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9eededc0cb1af94fb7b5daba08785f6b077a85e9cb5ddd8b9f397b658194f61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:14Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:14 crc kubenswrapper[4754]: E1005 20:55:14.465078 4754 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T20:55:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T20:55:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T20:55:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T20:55:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ede2b442-76b0-443a-a450-252ee8711c80\\\",\\\"systemUUID\\\":\\\"72d081dd-fe7e-42a6-9cf0-f0ee7f733ac9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:14Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:14 crc kubenswrapper[4754]: I1005 20:55:14.469060 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:14 crc kubenswrapper[4754]: I1005 20:55:14.469103 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:14 crc kubenswrapper[4754]: I1005 20:55:14.469116 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:14 crc kubenswrapper[4754]: I1005 20:55:14.469138 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:14 crc kubenswrapper[4754]: I1005 20:55:14.469154 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:14Z","lastTransitionTime":"2025-10-05T20:55:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:14 crc kubenswrapper[4754]: I1005 20:55:14.477349 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tskt7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a68e9c89-465b-4c01-a470-e21a480ee465\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba6edba14cf321b4cd2f7bb44067a3ecf49f1594e76c0197e807ff3daa45e9b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d14ec99610d2167b484bd7712af7bb6d993d26c74b2828da1c89314f16fda22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d14ec99610d2167b484bd7712af7bb6d993d26c74b2828da1c89314f16fda22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f6f08dd853b25881e2f3c66206aa80d0c493518284217780aa40ac4b69f0c0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f6f08dd853b25881e2f3c66206aa80d0c493518284217780aa40ac4b69f0c0b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://71026d00ec1a3d91b019deb3c9f43ee52fb05b8f2d6374e1a5b762c9e5c1b10d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71026d00ec1a3d91b019deb3c9f43ee52fb05b8f2d6374e1a5b762c9e5c1b10d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b9910d2470acb16ebb436d7030b3161061607cb2a5ef50d482d8d5c4f48a8ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b9910d2470acb16ebb436d7030b3161061607cb2a5ef50d482d8d5c4f48a8ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0a93b61eacc55111995e69c15f89a210b8aa15468ac9ac053b7cf6b007d301e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b0a93b61eacc55111995e69c15f89a210b8aa15468ac9ac053b7cf6b007d301e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cafb3208e7e3bcdd01898a68fc5bbf4cb52ca0800b879b77a60bb102794257f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cafb3208e7e3bcdd01898a68fc5bbf4cb52ca0800b879b77a60bb102794257f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tskt7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:14Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:14 crc kubenswrapper[4754]: E1005 20:55:14.483367 4754 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T20:55:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T20:55:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T20:55:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T20:55:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ede2b442-76b0-443a-a450-252ee8711c80\\\",\\\"systemUUID\\\":\\\"72d081dd-fe7e-42a6-9cf0-f0ee7f733ac9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:14Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:14 crc kubenswrapper[4754]: I1005 20:55:14.487915 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:14 crc kubenswrapper[4754]: I1005 20:55:14.487949 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:14 crc kubenswrapper[4754]: I1005 20:55:14.487962 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:14 crc kubenswrapper[4754]: I1005 20:55:14.487983 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:14 crc kubenswrapper[4754]: I1005 20:55:14.487997 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:14Z","lastTransitionTime":"2025-10-05T20:55:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:14 crc kubenswrapper[4754]: I1005 20:55:14.489043 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 20:55:14 crc kubenswrapper[4754]: E1005 20:55:14.489414 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 20:55:30.489386005 +0000 UTC m=+54.393504745 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:55:14 crc kubenswrapper[4754]: I1005 20:55:14.491688 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be95a413-6a59-45b4-84b7-b43cae694a26\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10b211cf6e47dbd1ee42c82e49de0a13a45581b92a0c4f6bbba0326c7149e41d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2ggg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4bb8c5903e8b70dd19eebb111097d96f725e17de44e06c3e74c067376f6c7360\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2ggg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-b2h9k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:14Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:14 crc kubenswrapper[4754]: E1005 20:55:14.501217 4754 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T20:55:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T20:55:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T20:55:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T20:55:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ede2b442-76b0-443a-a450-252ee8711c80\\\",\\\"systemUUID\\\":\\\"72d081dd-fe7e-42a6-9cf0-f0ee7f733ac9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:14Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:14 crc kubenswrapper[4754]: E1005 20:55:14.501416 4754 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 05 20:55:14 crc kubenswrapper[4754]: I1005 20:55:14.503931 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:14 crc kubenswrapper[4754]: I1005 20:55:14.503965 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:14 crc kubenswrapper[4754]: I1005 20:55:14.503978 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:14 crc kubenswrapper[4754]: I1005 20:55:14.503997 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:14 crc kubenswrapper[4754]: I1005 20:55:14.504010 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:14Z","lastTransitionTime":"2025-10-05T20:55:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:14 crc kubenswrapper[4754]: I1005 20:55:14.505649 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rddkl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"15afc2a1-5cb7-41c6-954b-446c36822f50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://866d85ee126f8641d12f1a05fabc3de7b9392c324e9fe06ea92214155c05334c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqv8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a175565f83cfcba7825006eafb0f83efb70ec3d415c2c039c6499d5f6576e89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqv8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:55:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rddkl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:14Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:14 crc kubenswrapper[4754]: I1005 20:55:14.524863 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e380f9b2a0e319dc325c54d1247a6db98346292548fa688cb41a2ca1b3048419\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:14Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:14 crc kubenswrapper[4754]: I1005 20:55:14.543907 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:14Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:14 crc kubenswrapper[4754]: I1005 20:55:14.590700 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 20:55:14 crc kubenswrapper[4754]: I1005 20:55:14.590765 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 20:55:14 crc kubenswrapper[4754]: I1005 20:55:14.590797 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 20:55:14 crc kubenswrapper[4754]: I1005 20:55:14.590825 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 20:55:14 crc kubenswrapper[4754]: E1005 20:55:14.590842 4754 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 05 20:55:14 crc kubenswrapper[4754]: E1005 20:55:14.590921 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-05 20:55:30.590903697 +0000 UTC m=+54.495022407 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 05 20:55:14 crc kubenswrapper[4754]: E1005 20:55:14.590978 4754 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 05 20:55:14 crc kubenswrapper[4754]: E1005 20:55:14.591019 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-05 20:55:30.59100773 +0000 UTC m=+54.495126440 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 05 20:55:14 crc kubenswrapper[4754]: E1005 20:55:14.591067 4754 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 05 20:55:14 crc kubenswrapper[4754]: E1005 20:55:14.591143 4754 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 05 20:55:14 crc kubenswrapper[4754]: E1005 20:55:14.591154 4754 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 05 20:55:14 crc kubenswrapper[4754]: E1005 20:55:14.591168 4754 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 05 20:55:14 crc kubenswrapper[4754]: E1005 20:55:14.591177 4754 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 05 20:55:14 crc kubenswrapper[4754]: E1005 20:55:14.591197 4754 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 05 20:55:14 crc kubenswrapper[4754]: E1005 20:55:14.591230 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-05 20:55:30.591219805 +0000 UTC m=+54.495338505 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 05 20:55:14 crc kubenswrapper[4754]: E1005 20:55:14.591250 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-05 20:55:30.591241056 +0000 UTC m=+54.495359776 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 05 20:55:14 crc kubenswrapper[4754]: I1005 20:55:14.609070 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:14 crc kubenswrapper[4754]: I1005 20:55:14.609117 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:14 crc kubenswrapper[4754]: I1005 20:55:14.609127 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:14 crc kubenswrapper[4754]: I1005 20:55:14.609146 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:14 crc kubenswrapper[4754]: I1005 20:55:14.609157 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:14Z","lastTransitionTime":"2025-10-05T20:55:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:14 crc kubenswrapper[4754]: I1005 20:55:14.712481 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:14 crc kubenswrapper[4754]: I1005 20:55:14.712582 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:14 crc kubenswrapper[4754]: I1005 20:55:14.712600 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:14 crc kubenswrapper[4754]: I1005 20:55:14.712630 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:14 crc kubenswrapper[4754]: I1005 20:55:14.712649 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:14Z","lastTransitionTime":"2025-10-05T20:55:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:14 crc kubenswrapper[4754]: I1005 20:55:14.815685 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:14 crc kubenswrapper[4754]: I1005 20:55:14.815740 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:14 crc kubenswrapper[4754]: I1005 20:55:14.815750 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:14 crc kubenswrapper[4754]: I1005 20:55:14.815766 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:14 crc kubenswrapper[4754]: I1005 20:55:14.815777 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:14Z","lastTransitionTime":"2025-10-05T20:55:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:14 crc kubenswrapper[4754]: I1005 20:55:14.837360 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 20:55:14 crc kubenswrapper[4754]: I1005 20:55:14.837464 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 20:55:14 crc kubenswrapper[4754]: E1005 20:55:14.837554 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 20:55:14 crc kubenswrapper[4754]: I1005 20:55:14.837464 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 20:55:14 crc kubenswrapper[4754]: E1005 20:55:14.837711 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 20:55:14 crc kubenswrapper[4754]: E1005 20:55:14.838143 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 20:55:14 crc kubenswrapper[4754]: I1005 20:55:14.919365 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:14 crc kubenswrapper[4754]: I1005 20:55:14.919436 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:14 crc kubenswrapper[4754]: I1005 20:55:14.919462 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:14 crc kubenswrapper[4754]: I1005 20:55:14.919531 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:14 crc kubenswrapper[4754]: I1005 20:55:14.919553 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:14Z","lastTransitionTime":"2025-10-05T20:55:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:15 crc kubenswrapper[4754]: I1005 20:55:15.022727 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:15 crc kubenswrapper[4754]: I1005 20:55:15.022793 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:15 crc kubenswrapper[4754]: I1005 20:55:15.022813 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:15 crc kubenswrapper[4754]: I1005 20:55:15.022838 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:15 crc kubenswrapper[4754]: I1005 20:55:15.022861 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:15Z","lastTransitionTime":"2025-10-05T20:55:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:15 crc kubenswrapper[4754]: I1005 20:55:15.126480 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:15 crc kubenswrapper[4754]: I1005 20:55:15.126612 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:15 crc kubenswrapper[4754]: I1005 20:55:15.126640 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:15 crc kubenswrapper[4754]: I1005 20:55:15.126682 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:15 crc kubenswrapper[4754]: I1005 20:55:15.126714 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:15Z","lastTransitionTime":"2025-10-05T20:55:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:15 crc kubenswrapper[4754]: I1005 20:55:15.230651 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:15 crc kubenswrapper[4754]: I1005 20:55:15.230721 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:15 crc kubenswrapper[4754]: I1005 20:55:15.230735 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:15 crc kubenswrapper[4754]: I1005 20:55:15.230760 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:15 crc kubenswrapper[4754]: I1005 20:55:15.230776 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:15Z","lastTransitionTime":"2025-10-05T20:55:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:15 crc kubenswrapper[4754]: I1005 20:55:15.334322 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:15 crc kubenswrapper[4754]: I1005 20:55:15.334380 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:15 crc kubenswrapper[4754]: I1005 20:55:15.334397 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:15 crc kubenswrapper[4754]: I1005 20:55:15.334422 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:15 crc kubenswrapper[4754]: I1005 20:55:15.334442 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:15Z","lastTransitionTime":"2025-10-05T20:55:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:15 crc kubenswrapper[4754]: I1005 20:55:15.402142 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7dc728ea-2601-44f0-bcab-2913f034007d-metrics-certs\") pod \"network-metrics-daemon-nwrnt\" (UID: \"7dc728ea-2601-44f0-bcab-2913f034007d\") " pod="openshift-multus/network-metrics-daemon-nwrnt" Oct 05 20:55:15 crc kubenswrapper[4754]: E1005 20:55:15.402536 4754 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 05 20:55:15 crc kubenswrapper[4754]: E1005 20:55:15.402813 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7dc728ea-2601-44f0-bcab-2913f034007d-metrics-certs podName:7dc728ea-2601-44f0-bcab-2913f034007d nodeName:}" failed. No retries permitted until 2025-10-05 20:55:17.40278566 +0000 UTC m=+41.306904410 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/7dc728ea-2601-44f0-bcab-2913f034007d-metrics-certs") pod "network-metrics-daemon-nwrnt" (UID: "7dc728ea-2601-44f0-bcab-2913f034007d") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 05 20:55:15 crc kubenswrapper[4754]: I1005 20:55:15.438582 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:15 crc kubenswrapper[4754]: I1005 20:55:15.438653 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:15 crc kubenswrapper[4754]: I1005 20:55:15.438672 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:15 crc kubenswrapper[4754]: I1005 20:55:15.438700 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:15 crc kubenswrapper[4754]: I1005 20:55:15.438719 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:15Z","lastTransitionTime":"2025-10-05T20:55:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:15 crc kubenswrapper[4754]: I1005 20:55:15.541331 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:15 crc kubenswrapper[4754]: I1005 20:55:15.541382 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:15 crc kubenswrapper[4754]: I1005 20:55:15.541393 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:15 crc kubenswrapper[4754]: I1005 20:55:15.541413 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:15 crc kubenswrapper[4754]: I1005 20:55:15.541425 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:15Z","lastTransitionTime":"2025-10-05T20:55:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:15 crc kubenswrapper[4754]: I1005 20:55:15.644656 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:15 crc kubenswrapper[4754]: I1005 20:55:15.644716 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:15 crc kubenswrapper[4754]: I1005 20:55:15.644733 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:15 crc kubenswrapper[4754]: I1005 20:55:15.644762 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:15 crc kubenswrapper[4754]: I1005 20:55:15.644782 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:15Z","lastTransitionTime":"2025-10-05T20:55:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:15 crc kubenswrapper[4754]: I1005 20:55:15.748488 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:15 crc kubenswrapper[4754]: I1005 20:55:15.748580 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:15 crc kubenswrapper[4754]: I1005 20:55:15.748598 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:15 crc kubenswrapper[4754]: I1005 20:55:15.748627 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:15 crc kubenswrapper[4754]: I1005 20:55:15.748647 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:15Z","lastTransitionTime":"2025-10-05T20:55:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:15 crc kubenswrapper[4754]: I1005 20:55:15.837044 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nwrnt" Oct 05 20:55:15 crc kubenswrapper[4754]: E1005 20:55:15.837334 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nwrnt" podUID="7dc728ea-2601-44f0-bcab-2913f034007d" Oct 05 20:55:15 crc kubenswrapper[4754]: I1005 20:55:15.852856 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:15 crc kubenswrapper[4754]: I1005 20:55:15.852936 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:15 crc kubenswrapper[4754]: I1005 20:55:15.852957 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:15 crc kubenswrapper[4754]: I1005 20:55:15.852986 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:15 crc kubenswrapper[4754]: I1005 20:55:15.853006 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:15Z","lastTransitionTime":"2025-10-05T20:55:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:15 crc kubenswrapper[4754]: I1005 20:55:15.955861 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:15 crc kubenswrapper[4754]: I1005 20:55:15.955934 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:15 crc kubenswrapper[4754]: I1005 20:55:15.955953 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:15 crc kubenswrapper[4754]: I1005 20:55:15.955980 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:15 crc kubenswrapper[4754]: I1005 20:55:15.955999 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:15Z","lastTransitionTime":"2025-10-05T20:55:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:16 crc kubenswrapper[4754]: I1005 20:55:16.059553 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:16 crc kubenswrapper[4754]: I1005 20:55:16.059637 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:16 crc kubenswrapper[4754]: I1005 20:55:16.059656 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:16 crc kubenswrapper[4754]: I1005 20:55:16.059699 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:16 crc kubenswrapper[4754]: I1005 20:55:16.059720 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:16Z","lastTransitionTime":"2025-10-05T20:55:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:16 crc kubenswrapper[4754]: I1005 20:55:16.163138 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:16 crc kubenswrapper[4754]: I1005 20:55:16.163209 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:16 crc kubenswrapper[4754]: I1005 20:55:16.163229 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:16 crc kubenswrapper[4754]: I1005 20:55:16.163256 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:16 crc kubenswrapper[4754]: I1005 20:55:16.163276 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:16Z","lastTransitionTime":"2025-10-05T20:55:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:16 crc kubenswrapper[4754]: I1005 20:55:16.266415 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:16 crc kubenswrapper[4754]: I1005 20:55:16.266516 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:16 crc kubenswrapper[4754]: I1005 20:55:16.266536 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:16 crc kubenswrapper[4754]: I1005 20:55:16.266565 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:16 crc kubenswrapper[4754]: I1005 20:55:16.266583 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:16Z","lastTransitionTime":"2025-10-05T20:55:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:16 crc kubenswrapper[4754]: I1005 20:55:16.369776 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:16 crc kubenswrapper[4754]: I1005 20:55:16.369852 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:16 crc kubenswrapper[4754]: I1005 20:55:16.369871 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:16 crc kubenswrapper[4754]: I1005 20:55:16.369907 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:16 crc kubenswrapper[4754]: I1005 20:55:16.369926 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:16Z","lastTransitionTime":"2025-10-05T20:55:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:16 crc kubenswrapper[4754]: I1005 20:55:16.474759 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:16 crc kubenswrapper[4754]: I1005 20:55:16.474821 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:16 crc kubenswrapper[4754]: I1005 20:55:16.474836 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:16 crc kubenswrapper[4754]: I1005 20:55:16.474856 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:16 crc kubenswrapper[4754]: I1005 20:55:16.474871 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:16Z","lastTransitionTime":"2025-10-05T20:55:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:16 crc kubenswrapper[4754]: I1005 20:55:16.577864 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:16 crc kubenswrapper[4754]: I1005 20:55:16.577933 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:16 crc kubenswrapper[4754]: I1005 20:55:16.577952 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:16 crc kubenswrapper[4754]: I1005 20:55:16.577979 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:16 crc kubenswrapper[4754]: I1005 20:55:16.577999 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:16Z","lastTransitionTime":"2025-10-05T20:55:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:16 crc kubenswrapper[4754]: I1005 20:55:16.681317 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:16 crc kubenswrapper[4754]: I1005 20:55:16.681401 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:16 crc kubenswrapper[4754]: I1005 20:55:16.681424 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:16 crc kubenswrapper[4754]: I1005 20:55:16.681455 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:16 crc kubenswrapper[4754]: I1005 20:55:16.681481 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:16Z","lastTransitionTime":"2025-10-05T20:55:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:16 crc kubenswrapper[4754]: I1005 20:55:16.786195 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:16 crc kubenswrapper[4754]: I1005 20:55:16.786278 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:16 crc kubenswrapper[4754]: I1005 20:55:16.786298 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:16 crc kubenswrapper[4754]: I1005 20:55:16.786327 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:16 crc kubenswrapper[4754]: I1005 20:55:16.786350 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:16Z","lastTransitionTime":"2025-10-05T20:55:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:16 crc kubenswrapper[4754]: I1005 20:55:16.837148 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 20:55:16 crc kubenswrapper[4754]: E1005 20:55:16.837346 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 20:55:16 crc kubenswrapper[4754]: I1005 20:55:16.837735 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 20:55:16 crc kubenswrapper[4754]: E1005 20:55:16.838290 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 20:55:16 crc kubenswrapper[4754]: I1005 20:55:16.838381 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 20:55:16 crc kubenswrapper[4754]: E1005 20:55:16.838649 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 20:55:16 crc kubenswrapper[4754]: I1005 20:55:16.875696 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b44790fb-fecb-4ec3-9816-8e23dd72ab33\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a7dc53e7f96e48213d724a2cb4859514df925005644108ae3a0c33b14dcf52d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfdd8b1e542ab676576c311e2923fb02861f92de91fe1f4b3e4b968c549c375d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3a63451ffaa14cacc3b78d11340f4ea81445fa3af96178e9b172d063d9697d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3cfd3f2274a8a46fa36660e94ac6c9c298c791283f12d52330c5594ac131998\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f626cb108716da0b595844ea0251b7a04d49b5908a1d6dd0aef5cb2d402c4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://822223f515b5add6d71d193fd5bc77df029b8d37a1cc07be4145b53c7226bf61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cbed906d679176209043edd5dcc649f5ba7d5848a10cfd09efa9e656dcacfef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cbed906d679176209043edd5dcc649f5ba7d5848a10cfd09efa9e656dcacfef\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-05T20:55:11Z\\\",\\\"message\\\":\\\"gressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1005 20:55:11.136172 6140 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1005 20:55:11.136316 6140 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1005 20:55:11.136624 6140 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1005 20:55:11.136664 6140 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1005 20:55:11.136709 6140 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1005 20:55:11.136749 6140 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1005 20:55:11.138444 6140 factory.go:656] Stopping watch factory\\\\nI1005 20:55:11.148575 6140 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1005 20:55:11.148612 6140 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1005 20:55:11.148701 6140 ovnkube.go:599] Stopped ovnkube\\\\nI1005 20:55:11.148748 6140 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1005 20:55:11.148871 6140 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:10Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-7rhps_openshift-ovn-kubernetes(b44790fb-fecb-4ec3-9816-8e23dd72ab33)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2acf9637fa25e5d3ffeb05224c11218c47defc679568fb928a498869613de4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35ebeb7a87d5d47b6567917d38eca38a3e622c497fba58971296c4609c287423\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35ebeb7a87d5d47b6567917d38eca38a3e622c497fba58971296c4609c287423\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7rhps\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:16Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:16 crc kubenswrapper[4754]: I1005 20:55:16.891076 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:16 crc kubenswrapper[4754]: I1005 20:55:16.891140 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:16 crc kubenswrapper[4754]: I1005 20:55:16.891158 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:16 crc kubenswrapper[4754]: I1005 20:55:16.891185 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:16 crc kubenswrapper[4754]: I1005 20:55:16.891204 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:16Z","lastTransitionTime":"2025-10-05T20:55:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:16 crc kubenswrapper[4754]: I1005 20:55:16.921147 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9309179-b92a-4fa4-9fba-747b8f911936\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d969d3d50d6ad280fc6654df9441bab238bb538f277ed6d1273779813ffbaaed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://800a731ed8ef57714b2318af14054ef23f0368936630887a093fc77167b33d12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8cd9b08e83dc174413cc37891c14562816787028471a4201680c86dd620b7a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93b729f3f7f993db0d6bc3ba60c04930aada3cef8272ffcc3c6a8a242a9b1475\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dde1519111fc3af5d14463fdb58d4f5c62a7045e7de98c60981c7e4da37b0e7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ea160a5b37aea7f2970ae4dd2614b09d3dcf5dcbd1fc103b9997f8ecc3a804c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ea160a5b37aea7f2970ae4dd2614b09d3dcf5dcbd1fc103b9997f8ecc3a804c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://17ae6aa6c53e7dd061f70b41428ef50bbc7ffcc65352fad1a4a9d3c69dbd2ba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17ae6aa6c53e7dd061f70b41428ef50bbc7ffcc65352fad1a4a9d3c69dbd2ba9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://ad4cd69df7608c7d87d839fe09d4557c674c7efbc4730f4e4d3daaf9a26f5aee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad4cd69df7608c7d87d839fe09d4557c674c7efbc4730f4e4d3daaf9a26f5aee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:36Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:16Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:16 crc kubenswrapper[4754]: I1005 20:55:16.942739 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:16Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:16 crc kubenswrapper[4754]: I1005 20:55:16.963979 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9ad070bb699b643d29b82bcc1126ce3b585f7492f5ae2817fa491b309b1691d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:16Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:16 crc kubenswrapper[4754]: I1005 20:55:16.983364 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-57sr4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e854c7b9-1229-4da4-80e9-fea6479459b8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceb577e1055d84007936a50d0b3d54b9ae0299a1fc1a6618b37163a6d3fbb9cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vw2ps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:58Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-57sr4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:16Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:16 crc kubenswrapper[4754]: I1005 20:55:16.997274 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:16 crc kubenswrapper[4754]: I1005 20:55:16.997346 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:16 crc kubenswrapper[4754]: I1005 20:55:16.997367 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:16 crc kubenswrapper[4754]: I1005 20:55:16.997402 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:16 crc kubenswrapper[4754]: I1005 20:55:16.997423 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:16Z","lastTransitionTime":"2025-10-05T20:55:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:17 crc kubenswrapper[4754]: I1005 20:55:17.005442 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cn76s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02d5e3f9-73c2-4496-9aca-0787184aef19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7ba2b076839f4a5f586d2e6c4f4d7188254128b782bfd2180e22bf3ea5ffdd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpx5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cn76s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:17Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:17 crc kubenswrapper[4754]: I1005 20:55:17.023103 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ss7lm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f9ef817-0622-4185-a7fb-57221690aa75\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ba613b0d7753a003c572c3733ed394707c45b6f8072cb7aa88f3c7312b8882b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wpwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:55:02Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ss7lm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:17Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:17 crc kubenswrapper[4754]: I1005 20:55:17.042309 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-nwrnt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7dc728ea-2601-44f0-bcab-2913f034007d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-846xq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-846xq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:55:13Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-nwrnt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:17Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:17 crc kubenswrapper[4754]: I1005 20:55:17.065355 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"934bcd66-9c03-48b9-82dd-5093338f9c8f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9042df2f43243d3d79ba9a3940f8dd4f6c7a89ef9e3510e81fe60b1944d7e83b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5366bbf46a91667355f43f5df7335a536b395cda788b90b35ef18e2b53bc4c02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://600bc3bb551b747cf2f18e26745883232d7a37ac5877fad0c67100e52d7a43aa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78f74982180048eb50b9be79fa72c04f62f925b54fe511920f37e6407f864631\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:17Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:17 crc kubenswrapper[4754]: I1005 20:55:17.087149 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:17Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:17 crc kubenswrapper[4754]: I1005 20:55:17.101062 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:17 crc kubenswrapper[4754]: I1005 20:55:17.101124 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:17 crc kubenswrapper[4754]: I1005 20:55:17.101137 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:17 crc kubenswrapper[4754]: I1005 20:55:17.101178 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:17 crc kubenswrapper[4754]: I1005 20:55:17.101197 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:17Z","lastTransitionTime":"2025-10-05T20:55:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:17 crc kubenswrapper[4754]: I1005 20:55:17.107826 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tskt7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a68e9c89-465b-4c01-a470-e21a480ee465\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba6edba14cf321b4cd2f7bb44067a3ecf49f1594e76c0197e807ff3daa45e9b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d14ec99610d2167b484bd7712af7bb6d993d26c74b2828da1c89314f16fda22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d14ec99610d2167b484bd7712af7bb6d993d26c74b2828da1c89314f16fda22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f6f08dd853b25881e2f3c66206aa80d0c493518284217780aa40ac4b69f0c0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f6f08dd853b25881e2f3c66206aa80d0c493518284217780aa40ac4b69f0c0b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://71026d00ec1a3d91b019deb3c9f43ee52fb05b8f2d6374e1a5b762c9e5c1b10d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71026d00ec1a3d91b019deb3c9f43ee52fb05b8f2d6374e1a5b762c9e5c1b10d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b9910d2470acb16ebb436d7030b3161061607cb2a5ef50d482d8d5c4f48a8ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b9910d2470acb16ebb436d7030b3161061607cb2a5ef50d482d8d5c4f48a8ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0a93b61eacc55111995e69c15f89a210b8aa15468ac9ac053b7cf6b007d301e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b0a93b61eacc55111995e69c15f89a210b8aa15468ac9ac053b7cf6b007d301e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cafb3208e7e3bcdd01898a68fc5bbf4cb52ca0800b879b77a60bb102794257f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cafb3208e7e3bcdd01898a68fc5bbf4cb52ca0800b879b77a60bb102794257f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tskt7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:17Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:17 crc kubenswrapper[4754]: I1005 20:55:17.126058 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be95a413-6a59-45b4-84b7-b43cae694a26\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10b211cf6e47dbd1ee42c82e49de0a13a45581b92a0c4f6bbba0326c7149e41d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2ggg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4bb8c5903e8b70dd19eebb111097d96f725e17de44e06c3e74c067376f6c7360\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2ggg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-b2h9k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:17Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:17 crc kubenswrapper[4754]: I1005 20:55:17.150464 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rddkl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"15afc2a1-5cb7-41c6-954b-446c36822f50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://866d85ee126f8641d12f1a05fabc3de7b9392c324e9fe06ea92214155c05334c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqv8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a175565f83cfcba7825006eafb0f83efb70ec3d415c2c039c6499d5f6576e89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqv8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:55:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rddkl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:17Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:17 crc kubenswrapper[4754]: I1005 20:55:17.170270 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7506d496-65ff-41dc-8230-24c8ee653a2a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33da5a6bb3cd3b042c193bc571e6622f2f87f27e560cdd0bbb9be31e2010d707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df6d700a947d19c84b4f65fd2dd2f8d1f4f1f80fa9b4800c2acfa8dae3912639\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35ca8ee77a0e58cf969bd67b277a3392614c4e73b963e6f6002097709370715d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e32c2154f46ba10c509fe40c357304aa8e82d04e9906bbaf2ccecd47b42279e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8907b9b69b249c34ec6db1e460d641b31711b75d6d31d4caafd867339e6571d0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1005 20:54:52.589115 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 20:54:52.590181 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1304624737/tls.crt::/tmp/serving-cert-1304624737/tls.key\\\\\\\"\\\\nI1005 20:54:58.070544 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1005 20:54:58.090352 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1005 20:54:58.090382 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1005 20:54:58.090412 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1005 20:54:58.090418 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1005 20:54:58.144656 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1005 20:54:58.144688 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 20:54:58.144695 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 20:54:58.144699 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1005 20:54:58.144702 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1005 20:54:58.144705 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1005 20:54:58.144708 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1005 20:54:58.145024 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1005 20:54:58.148112 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4efb34f3493cc8854bc4cff6b29f63973c52b754404d7a21daec4e1f612d327d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54c568b637f3737667433047b956c7eb22fdf0f9301acdfd3b49fb84c39de6ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54c568b637f3737667433047b956c7eb22fdf0f9301acdfd3b49fb84c39de6ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:17Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:17 crc kubenswrapper[4754]: I1005 20:55:17.187920 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c76e6e4740035a6ca54191a770c4516e28101aae69afbb89739174c2f076ef5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9eededc0cb1af94fb7b5daba08785f6b077a85e9cb5ddd8b9f397b658194f61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:17Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:17 crc kubenswrapper[4754]: I1005 20:55:17.204253 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:17 crc kubenswrapper[4754]: I1005 20:55:17.204315 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:17 crc kubenswrapper[4754]: I1005 20:55:17.204336 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:17 crc kubenswrapper[4754]: I1005 20:55:17.204367 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:17 crc kubenswrapper[4754]: I1005 20:55:17.204387 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:17Z","lastTransitionTime":"2025-10-05T20:55:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:17 crc kubenswrapper[4754]: I1005 20:55:17.208882 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e380f9b2a0e319dc325c54d1247a6db98346292548fa688cb41a2ca1b3048419\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:17Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:17 crc kubenswrapper[4754]: I1005 20:55:17.231153 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:17Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:17 crc kubenswrapper[4754]: I1005 20:55:17.308764 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:17 crc kubenswrapper[4754]: I1005 20:55:17.308848 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:17 crc kubenswrapper[4754]: I1005 20:55:17.308874 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:17 crc kubenswrapper[4754]: I1005 20:55:17.308906 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:17 crc kubenswrapper[4754]: I1005 20:55:17.308932 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:17Z","lastTransitionTime":"2025-10-05T20:55:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:17 crc kubenswrapper[4754]: I1005 20:55:17.413253 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:17 crc kubenswrapper[4754]: I1005 20:55:17.413314 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:17 crc kubenswrapper[4754]: I1005 20:55:17.413332 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:17 crc kubenswrapper[4754]: I1005 20:55:17.413361 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:17 crc kubenswrapper[4754]: I1005 20:55:17.413381 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:17Z","lastTransitionTime":"2025-10-05T20:55:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:17 crc kubenswrapper[4754]: I1005 20:55:17.432288 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7dc728ea-2601-44f0-bcab-2913f034007d-metrics-certs\") pod \"network-metrics-daemon-nwrnt\" (UID: \"7dc728ea-2601-44f0-bcab-2913f034007d\") " pod="openshift-multus/network-metrics-daemon-nwrnt" Oct 05 20:55:17 crc kubenswrapper[4754]: E1005 20:55:17.432685 4754 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 05 20:55:17 crc kubenswrapper[4754]: E1005 20:55:17.432796 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7dc728ea-2601-44f0-bcab-2913f034007d-metrics-certs podName:7dc728ea-2601-44f0-bcab-2913f034007d nodeName:}" failed. No retries permitted until 2025-10-05 20:55:21.432771569 +0000 UTC m=+45.336890319 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/7dc728ea-2601-44f0-bcab-2913f034007d-metrics-certs") pod "network-metrics-daemon-nwrnt" (UID: "7dc728ea-2601-44f0-bcab-2913f034007d") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 05 20:55:17 crc kubenswrapper[4754]: I1005 20:55:17.517055 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:17 crc kubenswrapper[4754]: I1005 20:55:17.517128 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:17 crc kubenswrapper[4754]: I1005 20:55:17.517146 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:17 crc kubenswrapper[4754]: I1005 20:55:17.517179 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:17 crc kubenswrapper[4754]: I1005 20:55:17.517203 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:17Z","lastTransitionTime":"2025-10-05T20:55:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:17 crc kubenswrapper[4754]: I1005 20:55:17.620249 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:17 crc kubenswrapper[4754]: I1005 20:55:17.620358 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:17 crc kubenswrapper[4754]: I1005 20:55:17.620857 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:17 crc kubenswrapper[4754]: I1005 20:55:17.620893 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:17 crc kubenswrapper[4754]: I1005 20:55:17.620915 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:17Z","lastTransitionTime":"2025-10-05T20:55:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:17 crc kubenswrapper[4754]: I1005 20:55:17.724990 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:17 crc kubenswrapper[4754]: I1005 20:55:17.725066 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:17 crc kubenswrapper[4754]: I1005 20:55:17.725087 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:17 crc kubenswrapper[4754]: I1005 20:55:17.725118 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:17 crc kubenswrapper[4754]: I1005 20:55:17.725140 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:17Z","lastTransitionTime":"2025-10-05T20:55:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:17 crc kubenswrapper[4754]: I1005 20:55:17.828367 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:17 crc kubenswrapper[4754]: I1005 20:55:17.828472 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:17 crc kubenswrapper[4754]: I1005 20:55:17.828522 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:17 crc kubenswrapper[4754]: I1005 20:55:17.828552 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:17 crc kubenswrapper[4754]: I1005 20:55:17.828577 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:17Z","lastTransitionTime":"2025-10-05T20:55:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:17 crc kubenswrapper[4754]: I1005 20:55:17.836978 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nwrnt" Oct 05 20:55:17 crc kubenswrapper[4754]: E1005 20:55:17.837273 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nwrnt" podUID="7dc728ea-2601-44f0-bcab-2913f034007d" Oct 05 20:55:17 crc kubenswrapper[4754]: I1005 20:55:17.933145 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:17 crc kubenswrapper[4754]: I1005 20:55:17.933215 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:17 crc kubenswrapper[4754]: I1005 20:55:17.933238 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:17 crc kubenswrapper[4754]: I1005 20:55:17.933266 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:17 crc kubenswrapper[4754]: I1005 20:55:17.933288 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:17Z","lastTransitionTime":"2025-10-05T20:55:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:18 crc kubenswrapper[4754]: I1005 20:55:18.037226 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:18 crc kubenswrapper[4754]: I1005 20:55:18.037301 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:18 crc kubenswrapper[4754]: I1005 20:55:18.037321 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:18 crc kubenswrapper[4754]: I1005 20:55:18.037366 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:18 crc kubenswrapper[4754]: I1005 20:55:18.037391 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:18Z","lastTransitionTime":"2025-10-05T20:55:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:18 crc kubenswrapper[4754]: I1005 20:55:18.140793 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:18 crc kubenswrapper[4754]: I1005 20:55:18.140867 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:18 crc kubenswrapper[4754]: I1005 20:55:18.140893 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:18 crc kubenswrapper[4754]: I1005 20:55:18.140945 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:18 crc kubenswrapper[4754]: I1005 20:55:18.140975 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:18Z","lastTransitionTime":"2025-10-05T20:55:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:18 crc kubenswrapper[4754]: I1005 20:55:18.244772 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:18 crc kubenswrapper[4754]: I1005 20:55:18.245236 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:18 crc kubenswrapper[4754]: I1005 20:55:18.245369 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:18 crc kubenswrapper[4754]: I1005 20:55:18.245591 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:18 crc kubenswrapper[4754]: I1005 20:55:18.245739 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:18Z","lastTransitionTime":"2025-10-05T20:55:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:18 crc kubenswrapper[4754]: I1005 20:55:18.350300 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:18 crc kubenswrapper[4754]: I1005 20:55:18.350380 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:18 crc kubenswrapper[4754]: I1005 20:55:18.350400 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:18 crc kubenswrapper[4754]: I1005 20:55:18.350431 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:18 crc kubenswrapper[4754]: I1005 20:55:18.350452 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:18Z","lastTransitionTime":"2025-10-05T20:55:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:18 crc kubenswrapper[4754]: I1005 20:55:18.454071 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:18 crc kubenswrapper[4754]: I1005 20:55:18.454157 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:18 crc kubenswrapper[4754]: I1005 20:55:18.454184 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:18 crc kubenswrapper[4754]: I1005 20:55:18.454218 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:18 crc kubenswrapper[4754]: I1005 20:55:18.454246 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:18Z","lastTransitionTime":"2025-10-05T20:55:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:18 crc kubenswrapper[4754]: I1005 20:55:18.557340 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:18 crc kubenswrapper[4754]: I1005 20:55:18.557429 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:18 crc kubenswrapper[4754]: I1005 20:55:18.557444 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:18 crc kubenswrapper[4754]: I1005 20:55:18.557463 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:18 crc kubenswrapper[4754]: I1005 20:55:18.557477 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:18Z","lastTransitionTime":"2025-10-05T20:55:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:18 crc kubenswrapper[4754]: I1005 20:55:18.659809 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:18 crc kubenswrapper[4754]: I1005 20:55:18.659860 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:18 crc kubenswrapper[4754]: I1005 20:55:18.659873 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:18 crc kubenswrapper[4754]: I1005 20:55:18.659892 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:18 crc kubenswrapper[4754]: I1005 20:55:18.659908 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:18Z","lastTransitionTime":"2025-10-05T20:55:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:18 crc kubenswrapper[4754]: I1005 20:55:18.763933 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:18 crc kubenswrapper[4754]: I1005 20:55:18.764036 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:18 crc kubenswrapper[4754]: I1005 20:55:18.764080 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:18 crc kubenswrapper[4754]: I1005 20:55:18.764113 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:18 crc kubenswrapper[4754]: I1005 20:55:18.764134 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:18Z","lastTransitionTime":"2025-10-05T20:55:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:18 crc kubenswrapper[4754]: I1005 20:55:18.837651 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 20:55:18 crc kubenswrapper[4754]: E1005 20:55:18.838017 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 20:55:18 crc kubenswrapper[4754]: I1005 20:55:18.838099 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 20:55:18 crc kubenswrapper[4754]: I1005 20:55:18.838149 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 20:55:18 crc kubenswrapper[4754]: E1005 20:55:18.838361 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 20:55:18 crc kubenswrapper[4754]: E1005 20:55:18.838488 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 20:55:18 crc kubenswrapper[4754]: I1005 20:55:18.867376 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:18 crc kubenswrapper[4754]: I1005 20:55:18.867470 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:18 crc kubenswrapper[4754]: I1005 20:55:18.867488 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:18 crc kubenswrapper[4754]: I1005 20:55:18.867561 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:18 crc kubenswrapper[4754]: I1005 20:55:18.867577 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:18Z","lastTransitionTime":"2025-10-05T20:55:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:18 crc kubenswrapper[4754]: I1005 20:55:18.971364 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:18 crc kubenswrapper[4754]: I1005 20:55:18.971416 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:18 crc kubenswrapper[4754]: I1005 20:55:18.971434 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:18 crc kubenswrapper[4754]: I1005 20:55:18.971462 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:18 crc kubenswrapper[4754]: I1005 20:55:18.971480 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:18Z","lastTransitionTime":"2025-10-05T20:55:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:19 crc kubenswrapper[4754]: I1005 20:55:19.075033 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:19 crc kubenswrapper[4754]: I1005 20:55:19.075098 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:19 crc kubenswrapper[4754]: I1005 20:55:19.075115 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:19 crc kubenswrapper[4754]: I1005 20:55:19.075143 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:19 crc kubenswrapper[4754]: I1005 20:55:19.075161 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:19Z","lastTransitionTime":"2025-10-05T20:55:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:19 crc kubenswrapper[4754]: I1005 20:55:19.178407 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:19 crc kubenswrapper[4754]: I1005 20:55:19.178479 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:19 crc kubenswrapper[4754]: I1005 20:55:19.178531 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:19 crc kubenswrapper[4754]: I1005 20:55:19.178562 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:19 crc kubenswrapper[4754]: I1005 20:55:19.178584 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:19Z","lastTransitionTime":"2025-10-05T20:55:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:19 crc kubenswrapper[4754]: I1005 20:55:19.288574 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:19 crc kubenswrapper[4754]: I1005 20:55:19.288675 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:19 crc kubenswrapper[4754]: I1005 20:55:19.288698 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:19 crc kubenswrapper[4754]: I1005 20:55:19.288729 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:19 crc kubenswrapper[4754]: I1005 20:55:19.288761 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:19Z","lastTransitionTime":"2025-10-05T20:55:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:19 crc kubenswrapper[4754]: I1005 20:55:19.393164 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:19 crc kubenswrapper[4754]: I1005 20:55:19.393609 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:19 crc kubenswrapper[4754]: I1005 20:55:19.393775 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:19 crc kubenswrapper[4754]: I1005 20:55:19.393917 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:19 crc kubenswrapper[4754]: I1005 20:55:19.394044 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:19Z","lastTransitionTime":"2025-10-05T20:55:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:19 crc kubenswrapper[4754]: I1005 20:55:19.498259 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:19 crc kubenswrapper[4754]: I1005 20:55:19.498328 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:19 crc kubenswrapper[4754]: I1005 20:55:19.498346 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:19 crc kubenswrapper[4754]: I1005 20:55:19.498377 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:19 crc kubenswrapper[4754]: I1005 20:55:19.498399 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:19Z","lastTransitionTime":"2025-10-05T20:55:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:19 crc kubenswrapper[4754]: I1005 20:55:19.601615 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:19 crc kubenswrapper[4754]: I1005 20:55:19.601681 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:19 crc kubenswrapper[4754]: I1005 20:55:19.601701 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:19 crc kubenswrapper[4754]: I1005 20:55:19.601730 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:19 crc kubenswrapper[4754]: I1005 20:55:19.601750 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:19Z","lastTransitionTime":"2025-10-05T20:55:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:19 crc kubenswrapper[4754]: I1005 20:55:19.705237 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:19 crc kubenswrapper[4754]: I1005 20:55:19.705305 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:19 crc kubenswrapper[4754]: I1005 20:55:19.705323 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:19 crc kubenswrapper[4754]: I1005 20:55:19.705349 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:19 crc kubenswrapper[4754]: I1005 20:55:19.705370 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:19Z","lastTransitionTime":"2025-10-05T20:55:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:19 crc kubenswrapper[4754]: I1005 20:55:19.809158 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:19 crc kubenswrapper[4754]: I1005 20:55:19.809234 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:19 crc kubenswrapper[4754]: I1005 20:55:19.809256 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:19 crc kubenswrapper[4754]: I1005 20:55:19.809290 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:19 crc kubenswrapper[4754]: I1005 20:55:19.809315 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:19Z","lastTransitionTime":"2025-10-05T20:55:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:19 crc kubenswrapper[4754]: I1005 20:55:19.836303 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nwrnt" Oct 05 20:55:19 crc kubenswrapper[4754]: E1005 20:55:19.836532 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nwrnt" podUID="7dc728ea-2601-44f0-bcab-2913f034007d" Oct 05 20:55:19 crc kubenswrapper[4754]: I1005 20:55:19.913188 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:19 crc kubenswrapper[4754]: I1005 20:55:19.913656 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:19 crc kubenswrapper[4754]: I1005 20:55:19.913954 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:19 crc kubenswrapper[4754]: I1005 20:55:19.914157 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:19 crc kubenswrapper[4754]: I1005 20:55:19.914340 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:19Z","lastTransitionTime":"2025-10-05T20:55:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:20 crc kubenswrapper[4754]: I1005 20:55:20.017753 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:20 crc kubenswrapper[4754]: I1005 20:55:20.018237 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:20 crc kubenswrapper[4754]: I1005 20:55:20.018715 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:20 crc kubenswrapper[4754]: I1005 20:55:20.019068 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:20 crc kubenswrapper[4754]: I1005 20:55:20.019401 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:20Z","lastTransitionTime":"2025-10-05T20:55:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:20 crc kubenswrapper[4754]: I1005 20:55:20.122572 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:20 crc kubenswrapper[4754]: I1005 20:55:20.122639 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:20 crc kubenswrapper[4754]: I1005 20:55:20.122658 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:20 crc kubenswrapper[4754]: I1005 20:55:20.123074 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:20 crc kubenswrapper[4754]: I1005 20:55:20.123122 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:20Z","lastTransitionTime":"2025-10-05T20:55:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:20 crc kubenswrapper[4754]: I1005 20:55:20.226014 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:20 crc kubenswrapper[4754]: I1005 20:55:20.226070 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:20 crc kubenswrapper[4754]: I1005 20:55:20.226087 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:20 crc kubenswrapper[4754]: I1005 20:55:20.226111 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:20 crc kubenswrapper[4754]: I1005 20:55:20.226134 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:20Z","lastTransitionTime":"2025-10-05T20:55:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:20 crc kubenswrapper[4754]: I1005 20:55:20.330708 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:20 crc kubenswrapper[4754]: I1005 20:55:20.330800 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:20 crc kubenswrapper[4754]: I1005 20:55:20.330834 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:20 crc kubenswrapper[4754]: I1005 20:55:20.330868 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:20 crc kubenswrapper[4754]: I1005 20:55:20.330896 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:20Z","lastTransitionTime":"2025-10-05T20:55:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:20 crc kubenswrapper[4754]: I1005 20:55:20.438777 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:20 crc kubenswrapper[4754]: I1005 20:55:20.438831 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:20 crc kubenswrapper[4754]: I1005 20:55:20.438849 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:20 crc kubenswrapper[4754]: I1005 20:55:20.438873 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:20 crc kubenswrapper[4754]: I1005 20:55:20.438894 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:20Z","lastTransitionTime":"2025-10-05T20:55:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:20 crc kubenswrapper[4754]: I1005 20:55:20.542696 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:20 crc kubenswrapper[4754]: I1005 20:55:20.542763 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:20 crc kubenswrapper[4754]: I1005 20:55:20.542783 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:20 crc kubenswrapper[4754]: I1005 20:55:20.542811 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:20 crc kubenswrapper[4754]: I1005 20:55:20.542830 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:20Z","lastTransitionTime":"2025-10-05T20:55:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:20 crc kubenswrapper[4754]: I1005 20:55:20.646604 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:20 crc kubenswrapper[4754]: I1005 20:55:20.646672 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:20 crc kubenswrapper[4754]: I1005 20:55:20.646690 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:20 crc kubenswrapper[4754]: I1005 20:55:20.646737 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:20 crc kubenswrapper[4754]: I1005 20:55:20.646756 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:20Z","lastTransitionTime":"2025-10-05T20:55:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:20 crc kubenswrapper[4754]: I1005 20:55:20.749641 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:20 crc kubenswrapper[4754]: I1005 20:55:20.749726 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:20 crc kubenswrapper[4754]: I1005 20:55:20.749749 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:20 crc kubenswrapper[4754]: I1005 20:55:20.749780 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:20 crc kubenswrapper[4754]: I1005 20:55:20.749803 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:20Z","lastTransitionTime":"2025-10-05T20:55:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:20 crc kubenswrapper[4754]: I1005 20:55:20.837269 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 20:55:20 crc kubenswrapper[4754]: I1005 20:55:20.837269 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 20:55:20 crc kubenswrapper[4754]: E1005 20:55:20.837550 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 20:55:20 crc kubenswrapper[4754]: I1005 20:55:20.837315 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 20:55:20 crc kubenswrapper[4754]: E1005 20:55:20.837789 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 20:55:20 crc kubenswrapper[4754]: E1005 20:55:20.837891 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 20:55:20 crc kubenswrapper[4754]: I1005 20:55:20.853228 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:20 crc kubenswrapper[4754]: I1005 20:55:20.853329 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:20 crc kubenswrapper[4754]: I1005 20:55:20.853353 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:20 crc kubenswrapper[4754]: I1005 20:55:20.853381 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:20 crc kubenswrapper[4754]: I1005 20:55:20.853405 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:20Z","lastTransitionTime":"2025-10-05T20:55:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:20 crc kubenswrapper[4754]: I1005 20:55:20.957185 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:20 crc kubenswrapper[4754]: I1005 20:55:20.957244 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:20 crc kubenswrapper[4754]: I1005 20:55:20.957257 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:20 crc kubenswrapper[4754]: I1005 20:55:20.957279 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:20 crc kubenswrapper[4754]: I1005 20:55:20.957302 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:20Z","lastTransitionTime":"2025-10-05T20:55:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:21 crc kubenswrapper[4754]: I1005 20:55:21.061227 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:21 crc kubenswrapper[4754]: I1005 20:55:21.061302 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:21 crc kubenswrapper[4754]: I1005 20:55:21.061323 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:21 crc kubenswrapper[4754]: I1005 20:55:21.061357 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:21 crc kubenswrapper[4754]: I1005 20:55:21.061392 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:21Z","lastTransitionTime":"2025-10-05T20:55:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:21 crc kubenswrapper[4754]: I1005 20:55:21.165094 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:21 crc kubenswrapper[4754]: I1005 20:55:21.165468 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:21 crc kubenswrapper[4754]: I1005 20:55:21.165757 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:21 crc kubenswrapper[4754]: I1005 20:55:21.165967 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:21 crc kubenswrapper[4754]: I1005 20:55:21.166173 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:21Z","lastTransitionTime":"2025-10-05T20:55:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:21 crc kubenswrapper[4754]: I1005 20:55:21.269808 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:21 crc kubenswrapper[4754]: I1005 20:55:21.270345 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:21 crc kubenswrapper[4754]: I1005 20:55:21.270541 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:21 crc kubenswrapper[4754]: I1005 20:55:21.270697 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:21 crc kubenswrapper[4754]: I1005 20:55:21.270917 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:21Z","lastTransitionTime":"2025-10-05T20:55:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:21 crc kubenswrapper[4754]: I1005 20:55:21.374649 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:21 crc kubenswrapper[4754]: I1005 20:55:21.374724 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:21 crc kubenswrapper[4754]: I1005 20:55:21.374805 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:21 crc kubenswrapper[4754]: I1005 20:55:21.374835 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:21 crc kubenswrapper[4754]: I1005 20:55:21.374857 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:21Z","lastTransitionTime":"2025-10-05T20:55:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:21 crc kubenswrapper[4754]: I1005 20:55:21.478834 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:21 crc kubenswrapper[4754]: I1005 20:55:21.478885 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:21 crc kubenswrapper[4754]: I1005 20:55:21.478903 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:21 crc kubenswrapper[4754]: I1005 20:55:21.478930 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:21 crc kubenswrapper[4754]: I1005 20:55:21.478949 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:21Z","lastTransitionTime":"2025-10-05T20:55:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:21 crc kubenswrapper[4754]: I1005 20:55:21.486053 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7dc728ea-2601-44f0-bcab-2913f034007d-metrics-certs\") pod \"network-metrics-daemon-nwrnt\" (UID: \"7dc728ea-2601-44f0-bcab-2913f034007d\") " pod="openshift-multus/network-metrics-daemon-nwrnt" Oct 05 20:55:21 crc kubenswrapper[4754]: E1005 20:55:21.486369 4754 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 05 20:55:21 crc kubenswrapper[4754]: E1005 20:55:21.486561 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7dc728ea-2601-44f0-bcab-2913f034007d-metrics-certs podName:7dc728ea-2601-44f0-bcab-2913f034007d nodeName:}" failed. No retries permitted until 2025-10-05 20:55:29.486523509 +0000 UTC m=+53.390642259 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/7dc728ea-2601-44f0-bcab-2913f034007d-metrics-certs") pod "network-metrics-daemon-nwrnt" (UID: "7dc728ea-2601-44f0-bcab-2913f034007d") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 05 20:55:21 crc kubenswrapper[4754]: I1005 20:55:21.582796 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:21 crc kubenswrapper[4754]: I1005 20:55:21.582859 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:21 crc kubenswrapper[4754]: I1005 20:55:21.582876 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:21 crc kubenswrapper[4754]: I1005 20:55:21.582902 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:21 crc kubenswrapper[4754]: I1005 20:55:21.582922 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:21Z","lastTransitionTime":"2025-10-05T20:55:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:21 crc kubenswrapper[4754]: I1005 20:55:21.686821 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:21 crc kubenswrapper[4754]: I1005 20:55:21.686887 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:21 crc kubenswrapper[4754]: I1005 20:55:21.686906 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:21 crc kubenswrapper[4754]: I1005 20:55:21.686935 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:21 crc kubenswrapper[4754]: I1005 20:55:21.686955 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:21Z","lastTransitionTime":"2025-10-05T20:55:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:21 crc kubenswrapper[4754]: I1005 20:55:21.790917 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:21 crc kubenswrapper[4754]: I1005 20:55:21.790961 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:21 crc kubenswrapper[4754]: I1005 20:55:21.790978 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:21 crc kubenswrapper[4754]: I1005 20:55:21.791000 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:21 crc kubenswrapper[4754]: I1005 20:55:21.791017 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:21Z","lastTransitionTime":"2025-10-05T20:55:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:21 crc kubenswrapper[4754]: I1005 20:55:21.837368 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nwrnt" Oct 05 20:55:21 crc kubenswrapper[4754]: E1005 20:55:21.837656 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nwrnt" podUID="7dc728ea-2601-44f0-bcab-2913f034007d" Oct 05 20:55:21 crc kubenswrapper[4754]: I1005 20:55:21.894753 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:21 crc kubenswrapper[4754]: I1005 20:55:21.894808 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:21 crc kubenswrapper[4754]: I1005 20:55:21.894826 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:21 crc kubenswrapper[4754]: I1005 20:55:21.894852 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:21 crc kubenswrapper[4754]: I1005 20:55:21.894872 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:21Z","lastTransitionTime":"2025-10-05T20:55:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:21 crc kubenswrapper[4754]: I1005 20:55:21.998647 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:21 crc kubenswrapper[4754]: I1005 20:55:21.998696 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:21 crc kubenswrapper[4754]: I1005 20:55:21.998708 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:21 crc kubenswrapper[4754]: I1005 20:55:21.998727 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:21 crc kubenswrapper[4754]: I1005 20:55:21.998738 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:21Z","lastTransitionTime":"2025-10-05T20:55:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:22 crc kubenswrapper[4754]: I1005 20:55:22.102796 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:22 crc kubenswrapper[4754]: I1005 20:55:22.102861 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:22 crc kubenswrapper[4754]: I1005 20:55:22.102878 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:22 crc kubenswrapper[4754]: I1005 20:55:22.102907 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:22 crc kubenswrapper[4754]: I1005 20:55:22.102932 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:22Z","lastTransitionTime":"2025-10-05T20:55:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:22 crc kubenswrapper[4754]: I1005 20:55:22.207141 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:22 crc kubenswrapper[4754]: I1005 20:55:22.207214 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:22 crc kubenswrapper[4754]: I1005 20:55:22.207236 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:22 crc kubenswrapper[4754]: I1005 20:55:22.207269 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:22 crc kubenswrapper[4754]: I1005 20:55:22.207289 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:22Z","lastTransitionTime":"2025-10-05T20:55:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:22 crc kubenswrapper[4754]: I1005 20:55:22.311267 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:22 crc kubenswrapper[4754]: I1005 20:55:22.311329 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:22 crc kubenswrapper[4754]: I1005 20:55:22.311347 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:22 crc kubenswrapper[4754]: I1005 20:55:22.311373 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:22 crc kubenswrapper[4754]: I1005 20:55:22.311394 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:22Z","lastTransitionTime":"2025-10-05T20:55:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:22 crc kubenswrapper[4754]: I1005 20:55:22.415038 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:22 crc kubenswrapper[4754]: I1005 20:55:22.415107 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:22 crc kubenswrapper[4754]: I1005 20:55:22.415134 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:22 crc kubenswrapper[4754]: I1005 20:55:22.415166 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:22 crc kubenswrapper[4754]: I1005 20:55:22.415189 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:22Z","lastTransitionTime":"2025-10-05T20:55:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:22 crc kubenswrapper[4754]: I1005 20:55:22.519115 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:22 crc kubenswrapper[4754]: I1005 20:55:22.519179 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:22 crc kubenswrapper[4754]: I1005 20:55:22.519189 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:22 crc kubenswrapper[4754]: I1005 20:55:22.519207 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:22 crc kubenswrapper[4754]: I1005 20:55:22.519216 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:22Z","lastTransitionTime":"2025-10-05T20:55:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:22 crc kubenswrapper[4754]: I1005 20:55:22.623395 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:22 crc kubenswrapper[4754]: I1005 20:55:22.623461 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:22 crc kubenswrapper[4754]: I1005 20:55:22.623483 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:22 crc kubenswrapper[4754]: I1005 20:55:22.623556 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:22 crc kubenswrapper[4754]: I1005 20:55:22.623580 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:22Z","lastTransitionTime":"2025-10-05T20:55:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:22 crc kubenswrapper[4754]: I1005 20:55:22.727666 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:22 crc kubenswrapper[4754]: I1005 20:55:22.727746 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:22 crc kubenswrapper[4754]: I1005 20:55:22.727771 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:22 crc kubenswrapper[4754]: I1005 20:55:22.727806 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:22 crc kubenswrapper[4754]: I1005 20:55:22.727834 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:22Z","lastTransitionTime":"2025-10-05T20:55:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:22 crc kubenswrapper[4754]: I1005 20:55:22.832062 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:22 crc kubenswrapper[4754]: I1005 20:55:22.832127 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:22 crc kubenswrapper[4754]: I1005 20:55:22.832146 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:22 crc kubenswrapper[4754]: I1005 20:55:22.832174 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:22 crc kubenswrapper[4754]: I1005 20:55:22.832209 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:22Z","lastTransitionTime":"2025-10-05T20:55:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:22 crc kubenswrapper[4754]: I1005 20:55:22.836577 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 20:55:22 crc kubenswrapper[4754]: I1005 20:55:22.836673 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 20:55:22 crc kubenswrapper[4754]: E1005 20:55:22.836758 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 20:55:22 crc kubenswrapper[4754]: I1005 20:55:22.836592 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 20:55:22 crc kubenswrapper[4754]: E1005 20:55:22.836868 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 20:55:22 crc kubenswrapper[4754]: E1005 20:55:22.842998 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 20:55:22 crc kubenswrapper[4754]: I1005 20:55:22.935073 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:22 crc kubenswrapper[4754]: I1005 20:55:22.935138 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:22 crc kubenswrapper[4754]: I1005 20:55:22.935154 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:22 crc kubenswrapper[4754]: I1005 20:55:22.935179 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:22 crc kubenswrapper[4754]: I1005 20:55:22.935197 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:22Z","lastTransitionTime":"2025-10-05T20:55:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:23 crc kubenswrapper[4754]: I1005 20:55:23.039936 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:23 crc kubenswrapper[4754]: I1005 20:55:23.040002 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:23 crc kubenswrapper[4754]: I1005 20:55:23.040064 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:23 crc kubenswrapper[4754]: I1005 20:55:23.040091 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:23 crc kubenswrapper[4754]: I1005 20:55:23.040190 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:23Z","lastTransitionTime":"2025-10-05T20:55:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:23 crc kubenswrapper[4754]: I1005 20:55:23.143025 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:23 crc kubenswrapper[4754]: I1005 20:55:23.143092 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:23 crc kubenswrapper[4754]: I1005 20:55:23.143111 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:23 crc kubenswrapper[4754]: I1005 20:55:23.143138 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:23 crc kubenswrapper[4754]: I1005 20:55:23.143161 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:23Z","lastTransitionTime":"2025-10-05T20:55:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:23 crc kubenswrapper[4754]: I1005 20:55:23.246177 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:23 crc kubenswrapper[4754]: I1005 20:55:23.246239 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:23 crc kubenswrapper[4754]: I1005 20:55:23.246260 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:23 crc kubenswrapper[4754]: I1005 20:55:23.246292 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:23 crc kubenswrapper[4754]: I1005 20:55:23.246311 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:23Z","lastTransitionTime":"2025-10-05T20:55:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:23 crc kubenswrapper[4754]: I1005 20:55:23.349317 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:23 crc kubenswrapper[4754]: I1005 20:55:23.349384 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:23 crc kubenswrapper[4754]: I1005 20:55:23.349401 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:23 crc kubenswrapper[4754]: I1005 20:55:23.349429 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:23 crc kubenswrapper[4754]: I1005 20:55:23.349448 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:23Z","lastTransitionTime":"2025-10-05T20:55:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:23 crc kubenswrapper[4754]: I1005 20:55:23.454900 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:23 crc kubenswrapper[4754]: I1005 20:55:23.454964 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:23 crc kubenswrapper[4754]: I1005 20:55:23.454983 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:23 crc kubenswrapper[4754]: I1005 20:55:23.455012 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:23 crc kubenswrapper[4754]: I1005 20:55:23.455043 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:23Z","lastTransitionTime":"2025-10-05T20:55:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:23 crc kubenswrapper[4754]: I1005 20:55:23.558988 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:23 crc kubenswrapper[4754]: I1005 20:55:23.559073 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:23 crc kubenswrapper[4754]: I1005 20:55:23.559091 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:23 crc kubenswrapper[4754]: I1005 20:55:23.559124 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:23 crc kubenswrapper[4754]: I1005 20:55:23.559143 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:23Z","lastTransitionTime":"2025-10-05T20:55:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:23 crc kubenswrapper[4754]: I1005 20:55:23.663259 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:23 crc kubenswrapper[4754]: I1005 20:55:23.663337 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:23 crc kubenswrapper[4754]: I1005 20:55:23.663359 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:23 crc kubenswrapper[4754]: I1005 20:55:23.663388 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:23 crc kubenswrapper[4754]: I1005 20:55:23.663411 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:23Z","lastTransitionTime":"2025-10-05T20:55:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:23 crc kubenswrapper[4754]: I1005 20:55:23.767844 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:23 crc kubenswrapper[4754]: I1005 20:55:23.767959 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:23 crc kubenswrapper[4754]: I1005 20:55:23.768018 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:23 crc kubenswrapper[4754]: I1005 20:55:23.768050 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:23 crc kubenswrapper[4754]: I1005 20:55:23.768068 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:23Z","lastTransitionTime":"2025-10-05T20:55:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:23 crc kubenswrapper[4754]: I1005 20:55:23.836409 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nwrnt" Oct 05 20:55:23 crc kubenswrapper[4754]: E1005 20:55:23.836671 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nwrnt" podUID="7dc728ea-2601-44f0-bcab-2913f034007d" Oct 05 20:55:23 crc kubenswrapper[4754]: I1005 20:55:23.871037 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:23 crc kubenswrapper[4754]: I1005 20:55:23.871124 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:23 crc kubenswrapper[4754]: I1005 20:55:23.871145 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:23 crc kubenswrapper[4754]: I1005 20:55:23.871739 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:23 crc kubenswrapper[4754]: I1005 20:55:23.871800 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:23Z","lastTransitionTime":"2025-10-05T20:55:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:23 crc kubenswrapper[4754]: I1005 20:55:23.976141 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:23 crc kubenswrapper[4754]: I1005 20:55:23.976201 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:23 crc kubenswrapper[4754]: I1005 20:55:23.976222 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:23 crc kubenswrapper[4754]: I1005 20:55:23.976251 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:23 crc kubenswrapper[4754]: I1005 20:55:23.976269 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:23Z","lastTransitionTime":"2025-10-05T20:55:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:24 crc kubenswrapper[4754]: I1005 20:55:24.080435 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:24 crc kubenswrapper[4754]: I1005 20:55:24.080527 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:24 crc kubenswrapper[4754]: I1005 20:55:24.080547 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:24 crc kubenswrapper[4754]: I1005 20:55:24.080575 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:24 crc kubenswrapper[4754]: I1005 20:55:24.080597 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:24Z","lastTransitionTime":"2025-10-05T20:55:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:24 crc kubenswrapper[4754]: I1005 20:55:24.183870 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:24 crc kubenswrapper[4754]: I1005 20:55:24.183925 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:24 crc kubenswrapper[4754]: I1005 20:55:24.183938 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:24 crc kubenswrapper[4754]: I1005 20:55:24.183958 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:24 crc kubenswrapper[4754]: I1005 20:55:24.183970 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:24Z","lastTransitionTime":"2025-10-05T20:55:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:24 crc kubenswrapper[4754]: I1005 20:55:24.287114 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:24 crc kubenswrapper[4754]: I1005 20:55:24.287179 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:24 crc kubenswrapper[4754]: I1005 20:55:24.287197 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:24 crc kubenswrapper[4754]: I1005 20:55:24.287226 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:24 crc kubenswrapper[4754]: I1005 20:55:24.287245 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:24Z","lastTransitionTime":"2025-10-05T20:55:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:24 crc kubenswrapper[4754]: I1005 20:55:24.390152 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:24 crc kubenswrapper[4754]: I1005 20:55:24.390227 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:24 crc kubenswrapper[4754]: I1005 20:55:24.390245 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:24 crc kubenswrapper[4754]: I1005 20:55:24.390273 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:24 crc kubenswrapper[4754]: I1005 20:55:24.390292 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:24Z","lastTransitionTime":"2025-10-05T20:55:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:24 crc kubenswrapper[4754]: I1005 20:55:24.494586 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:24 crc kubenswrapper[4754]: I1005 20:55:24.494655 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:24 crc kubenswrapper[4754]: I1005 20:55:24.494674 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:24 crc kubenswrapper[4754]: I1005 20:55:24.494701 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:24 crc kubenswrapper[4754]: I1005 20:55:24.494720 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:24Z","lastTransitionTime":"2025-10-05T20:55:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:24 crc kubenswrapper[4754]: I1005 20:55:24.550564 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:24 crc kubenswrapper[4754]: I1005 20:55:24.550631 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:24 crc kubenswrapper[4754]: I1005 20:55:24.550649 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:24 crc kubenswrapper[4754]: I1005 20:55:24.550678 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:24 crc kubenswrapper[4754]: I1005 20:55:24.550698 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:24Z","lastTransitionTime":"2025-10-05T20:55:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:24 crc kubenswrapper[4754]: E1005 20:55:24.573714 4754 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T20:55:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T20:55:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T20:55:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T20:55:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ede2b442-76b0-443a-a450-252ee8711c80\\\",\\\"systemUUID\\\":\\\"72d081dd-fe7e-42a6-9cf0-f0ee7f733ac9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:24Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:24 crc kubenswrapper[4754]: I1005 20:55:24.581382 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:24 crc kubenswrapper[4754]: I1005 20:55:24.581453 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:24 crc kubenswrapper[4754]: I1005 20:55:24.581471 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:24 crc kubenswrapper[4754]: I1005 20:55:24.581524 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:24 crc kubenswrapper[4754]: I1005 20:55:24.581546 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:24Z","lastTransitionTime":"2025-10-05T20:55:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:24 crc kubenswrapper[4754]: E1005 20:55:24.604199 4754 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T20:55:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T20:55:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T20:55:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T20:55:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ede2b442-76b0-443a-a450-252ee8711c80\\\",\\\"systemUUID\\\":\\\"72d081dd-fe7e-42a6-9cf0-f0ee7f733ac9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:24Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:24 crc kubenswrapper[4754]: I1005 20:55:24.610249 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:24 crc kubenswrapper[4754]: I1005 20:55:24.610314 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:24 crc kubenswrapper[4754]: I1005 20:55:24.610333 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:24 crc kubenswrapper[4754]: I1005 20:55:24.610366 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:24 crc kubenswrapper[4754]: I1005 20:55:24.610386 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:24Z","lastTransitionTime":"2025-10-05T20:55:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:24 crc kubenswrapper[4754]: E1005 20:55:24.631597 4754 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T20:55:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T20:55:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T20:55:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T20:55:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ede2b442-76b0-443a-a450-252ee8711c80\\\",\\\"systemUUID\\\":\\\"72d081dd-fe7e-42a6-9cf0-f0ee7f733ac9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:24Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:24 crc kubenswrapper[4754]: I1005 20:55:24.640228 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:24 crc kubenswrapper[4754]: I1005 20:55:24.640309 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:24 crc kubenswrapper[4754]: I1005 20:55:24.640333 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:24 crc kubenswrapper[4754]: I1005 20:55:24.640372 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:24 crc kubenswrapper[4754]: I1005 20:55:24.640397 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:24Z","lastTransitionTime":"2025-10-05T20:55:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:24 crc kubenswrapper[4754]: E1005 20:55:24.661154 4754 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T20:55:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T20:55:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T20:55:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T20:55:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ede2b442-76b0-443a-a450-252ee8711c80\\\",\\\"systemUUID\\\":\\\"72d081dd-fe7e-42a6-9cf0-f0ee7f733ac9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:24Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:24 crc kubenswrapper[4754]: I1005 20:55:24.667190 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:24 crc kubenswrapper[4754]: I1005 20:55:24.667266 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:24 crc kubenswrapper[4754]: I1005 20:55:24.667290 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:24 crc kubenswrapper[4754]: I1005 20:55:24.667320 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:24 crc kubenswrapper[4754]: I1005 20:55:24.667339 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:24Z","lastTransitionTime":"2025-10-05T20:55:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:24 crc kubenswrapper[4754]: E1005 20:55:24.692567 4754 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T20:55:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T20:55:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T20:55:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T20:55:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ede2b442-76b0-443a-a450-252ee8711c80\\\",\\\"systemUUID\\\":\\\"72d081dd-fe7e-42a6-9cf0-f0ee7f733ac9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:24Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:24 crc kubenswrapper[4754]: E1005 20:55:24.692801 4754 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 05 20:55:24 crc kubenswrapper[4754]: I1005 20:55:24.695922 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:24 crc kubenswrapper[4754]: I1005 20:55:24.696039 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:24 crc kubenswrapper[4754]: I1005 20:55:24.696097 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:24 crc kubenswrapper[4754]: I1005 20:55:24.696133 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:24 crc kubenswrapper[4754]: I1005 20:55:24.696193 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:24Z","lastTransitionTime":"2025-10-05T20:55:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:24 crc kubenswrapper[4754]: I1005 20:55:24.799395 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:24 crc kubenswrapper[4754]: I1005 20:55:24.800158 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:24 crc kubenswrapper[4754]: I1005 20:55:24.800320 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:24 crc kubenswrapper[4754]: I1005 20:55:24.801247 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:24 crc kubenswrapper[4754]: I1005 20:55:24.801360 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:24Z","lastTransitionTime":"2025-10-05T20:55:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:24 crc kubenswrapper[4754]: I1005 20:55:24.837565 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 20:55:24 crc kubenswrapper[4754]: I1005 20:55:24.837665 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 20:55:24 crc kubenswrapper[4754]: I1005 20:55:24.837704 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 20:55:24 crc kubenswrapper[4754]: E1005 20:55:24.837757 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 20:55:24 crc kubenswrapper[4754]: E1005 20:55:24.837912 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 20:55:24 crc kubenswrapper[4754]: E1005 20:55:24.838046 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 20:55:24 crc kubenswrapper[4754]: I1005 20:55:24.905287 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:24 crc kubenswrapper[4754]: I1005 20:55:24.905360 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:24 crc kubenswrapper[4754]: I1005 20:55:24.905379 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:24 crc kubenswrapper[4754]: I1005 20:55:24.905409 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:24 crc kubenswrapper[4754]: I1005 20:55:24.905433 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:24Z","lastTransitionTime":"2025-10-05T20:55:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:25 crc kubenswrapper[4754]: I1005 20:55:25.009278 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:25 crc kubenswrapper[4754]: I1005 20:55:25.009345 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:25 crc kubenswrapper[4754]: I1005 20:55:25.009364 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:25 crc kubenswrapper[4754]: I1005 20:55:25.009402 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:25 crc kubenswrapper[4754]: I1005 20:55:25.009419 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:25Z","lastTransitionTime":"2025-10-05T20:55:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:25 crc kubenswrapper[4754]: I1005 20:55:25.112930 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:25 crc kubenswrapper[4754]: I1005 20:55:25.113036 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:25 crc kubenswrapper[4754]: I1005 20:55:25.113056 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:25 crc kubenswrapper[4754]: I1005 20:55:25.113087 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:25 crc kubenswrapper[4754]: I1005 20:55:25.113104 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:25Z","lastTransitionTime":"2025-10-05T20:55:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:25 crc kubenswrapper[4754]: I1005 20:55:25.217480 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:25 crc kubenswrapper[4754]: I1005 20:55:25.217586 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:25 crc kubenswrapper[4754]: I1005 20:55:25.217608 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:25 crc kubenswrapper[4754]: I1005 20:55:25.217636 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:25 crc kubenswrapper[4754]: I1005 20:55:25.217658 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:25Z","lastTransitionTime":"2025-10-05T20:55:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:25 crc kubenswrapper[4754]: I1005 20:55:25.322891 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:25 crc kubenswrapper[4754]: I1005 20:55:25.322951 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:25 crc kubenswrapper[4754]: I1005 20:55:25.322969 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:25 crc kubenswrapper[4754]: I1005 20:55:25.323001 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:25 crc kubenswrapper[4754]: I1005 20:55:25.323020 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:25Z","lastTransitionTime":"2025-10-05T20:55:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:25 crc kubenswrapper[4754]: I1005 20:55:25.426934 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:25 crc kubenswrapper[4754]: I1005 20:55:25.427001 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:25 crc kubenswrapper[4754]: I1005 20:55:25.427019 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:25 crc kubenswrapper[4754]: I1005 20:55:25.427045 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:25 crc kubenswrapper[4754]: I1005 20:55:25.427063 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:25Z","lastTransitionTime":"2025-10-05T20:55:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:25 crc kubenswrapper[4754]: I1005 20:55:25.533771 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:25 crc kubenswrapper[4754]: I1005 20:55:25.533848 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:25 crc kubenswrapper[4754]: I1005 20:55:25.533860 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:25 crc kubenswrapper[4754]: I1005 20:55:25.533881 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:25 crc kubenswrapper[4754]: I1005 20:55:25.533893 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:25Z","lastTransitionTime":"2025-10-05T20:55:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:25 crc kubenswrapper[4754]: I1005 20:55:25.637744 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:25 crc kubenswrapper[4754]: I1005 20:55:25.637798 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:25 crc kubenswrapper[4754]: I1005 20:55:25.637812 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:25 crc kubenswrapper[4754]: I1005 20:55:25.637832 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:25 crc kubenswrapper[4754]: I1005 20:55:25.637844 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:25Z","lastTransitionTime":"2025-10-05T20:55:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:25 crc kubenswrapper[4754]: I1005 20:55:25.741114 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:25 crc kubenswrapper[4754]: I1005 20:55:25.741187 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:25 crc kubenswrapper[4754]: I1005 20:55:25.741209 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:25 crc kubenswrapper[4754]: I1005 20:55:25.741243 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:25 crc kubenswrapper[4754]: I1005 20:55:25.741261 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:25Z","lastTransitionTime":"2025-10-05T20:55:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:25 crc kubenswrapper[4754]: I1005 20:55:25.836957 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nwrnt" Oct 05 20:55:25 crc kubenswrapper[4754]: E1005 20:55:25.838469 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nwrnt" podUID="7dc728ea-2601-44f0-bcab-2913f034007d" Oct 05 20:55:25 crc kubenswrapper[4754]: I1005 20:55:25.839272 4754 scope.go:117] "RemoveContainer" containerID="2cbed906d679176209043edd5dcc649f5ba7d5848a10cfd09efa9e656dcacfef" Oct 05 20:55:25 crc kubenswrapper[4754]: I1005 20:55:25.845736 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:25 crc kubenswrapper[4754]: I1005 20:55:25.845779 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:25 crc kubenswrapper[4754]: I1005 20:55:25.845793 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:25 crc kubenswrapper[4754]: I1005 20:55:25.845812 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:25 crc kubenswrapper[4754]: I1005 20:55:25.845827 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:25Z","lastTransitionTime":"2025-10-05T20:55:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:25 crc kubenswrapper[4754]: I1005 20:55:25.949326 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:25 crc kubenswrapper[4754]: I1005 20:55:25.949406 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:25 crc kubenswrapper[4754]: I1005 20:55:25.949430 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:25 crc kubenswrapper[4754]: I1005 20:55:25.949464 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:25 crc kubenswrapper[4754]: I1005 20:55:25.949486 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:25Z","lastTransitionTime":"2025-10-05T20:55:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:26 crc kubenswrapper[4754]: I1005 20:55:26.052946 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:26 crc kubenswrapper[4754]: I1005 20:55:26.052992 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:26 crc kubenswrapper[4754]: I1005 20:55:26.053008 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:26 crc kubenswrapper[4754]: I1005 20:55:26.053028 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:26 crc kubenswrapper[4754]: I1005 20:55:26.053039 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:26Z","lastTransitionTime":"2025-10-05T20:55:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:26 crc kubenswrapper[4754]: I1005 20:55:26.156103 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:26 crc kubenswrapper[4754]: I1005 20:55:26.156175 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:26 crc kubenswrapper[4754]: I1005 20:55:26.156192 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:26 crc kubenswrapper[4754]: I1005 20:55:26.156220 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:26 crc kubenswrapper[4754]: I1005 20:55:26.156237 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:26Z","lastTransitionTime":"2025-10-05T20:55:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:26 crc kubenswrapper[4754]: I1005 20:55:26.259212 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:26 crc kubenswrapper[4754]: I1005 20:55:26.259268 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:26 crc kubenswrapper[4754]: I1005 20:55:26.259285 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:26 crc kubenswrapper[4754]: I1005 20:55:26.259307 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:26 crc kubenswrapper[4754]: I1005 20:55:26.259324 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:26Z","lastTransitionTime":"2025-10-05T20:55:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:26 crc kubenswrapper[4754]: I1005 20:55:26.286663 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7rhps_b44790fb-fecb-4ec3-9816-8e23dd72ab33/ovnkube-controller/1.log" Oct 05 20:55:26 crc kubenswrapper[4754]: I1005 20:55:26.291380 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" event={"ID":"b44790fb-fecb-4ec3-9816-8e23dd72ab33","Type":"ContainerStarted","Data":"af1c313e0dc0fe3b48feac3cce1ecd4452c9efadf3d19f257c07cc7fcdec875f"} Oct 05 20:55:26 crc kubenswrapper[4754]: I1005 20:55:26.292101 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" Oct 05 20:55:26 crc kubenswrapper[4754]: I1005 20:55:26.339673 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9309179-b92a-4fa4-9fba-747b8f911936\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d969d3d50d6ad280fc6654df9441bab238bb538f277ed6d1273779813ffbaaed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://800a731ed8ef57714b2318af14054ef23f0368936630887a093fc77167b33d12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8cd9b08e83dc174413cc37891c14562816787028471a4201680c86dd620b7a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93b729f3f7f993db0d6bc3ba60c04930aada3cef8272ffcc3c6a8a242a9b1475\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dde1519111fc3af5d14463fdb58d4f5c62a7045e7de98c60981c7e4da37b0e7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ea160a5b37aea7f2970ae4dd2614b09d3dcf5dcbd1fc103b9997f8ecc3a804c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ea160a5b37aea7f2970ae4dd2614b09d3dcf5dcbd1fc103b9997f8ecc3a804c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://17ae6aa6c53e7dd061f70b41428ef50bbc7ffcc65352fad1a4a9d3c69dbd2ba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17ae6aa6c53e7dd061f70b41428ef50bbc7ffcc65352fad1a4a9d3c69dbd2ba9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://ad4cd69df7608c7d87d839fe09d4557c674c7efbc4730f4e4d3daaf9a26f5aee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad4cd69df7608c7d87d839fe09d4557c674c7efbc4730f4e4d3daaf9a26f5aee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:36Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:26Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:26 crc kubenswrapper[4754]: I1005 20:55:26.362560 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:26 crc kubenswrapper[4754]: I1005 20:55:26.362603 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:26 crc kubenswrapper[4754]: I1005 20:55:26.362614 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:26 crc kubenswrapper[4754]: I1005 20:55:26.362630 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:26 crc kubenswrapper[4754]: I1005 20:55:26.362642 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:26Z","lastTransitionTime":"2025-10-05T20:55:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:26 crc kubenswrapper[4754]: I1005 20:55:26.363385 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:26Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:26 crc kubenswrapper[4754]: I1005 20:55:26.395447 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b44790fb-fecb-4ec3-9816-8e23dd72ab33\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a7dc53e7f96e48213d724a2cb4859514df925005644108ae3a0c33b14dcf52d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfdd8b1e542ab676576c311e2923fb02861f92de91fe1f4b3e4b968c549c375d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3a63451ffaa14cacc3b78d11340f4ea81445fa3af96178e9b172d063d9697d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3cfd3f2274a8a46fa36660e94ac6c9c298c791283f12d52330c5594ac131998\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f626cb108716da0b595844ea0251b7a04d49b5908a1d6dd0aef5cb2d402c4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://822223f515b5add6d71d193fd5bc77df029b8d37a1cc07be4145b53c7226bf61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af1c313e0dc0fe3b48feac3cce1ecd4452c9efadf3d19f257c07cc7fcdec875f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cbed906d679176209043edd5dcc649f5ba7d5848a10cfd09efa9e656dcacfef\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-05T20:55:11Z\\\",\\\"message\\\":\\\"gressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1005 20:55:11.136172 6140 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1005 20:55:11.136316 6140 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1005 20:55:11.136624 6140 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1005 20:55:11.136664 6140 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1005 20:55:11.136709 6140 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1005 20:55:11.136749 6140 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1005 20:55:11.138444 6140 factory.go:656] Stopping watch factory\\\\nI1005 20:55:11.148575 6140 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1005 20:55:11.148612 6140 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1005 20:55:11.148701 6140 ovnkube.go:599] Stopped ovnkube\\\\nI1005 20:55:11.148748 6140 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1005 20:55:11.148871 6140 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:10Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2acf9637fa25e5d3ffeb05224c11218c47defc679568fb928a498869613de4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35ebeb7a87d5d47b6567917d38eca38a3e622c497fba58971296c4609c287423\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35ebeb7a87d5d47b6567917d38eca38a3e622c497fba58971296c4609c287423\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7rhps\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:26Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:26 crc kubenswrapper[4754]: I1005 20:55:26.408334 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-nwrnt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7dc728ea-2601-44f0-bcab-2913f034007d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-846xq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-846xq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:55:13Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-nwrnt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:26Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:26 crc kubenswrapper[4754]: I1005 20:55:26.421576 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"934bcd66-9c03-48b9-82dd-5093338f9c8f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9042df2f43243d3d79ba9a3940f8dd4f6c7a89ef9e3510e81fe60b1944d7e83b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5366bbf46a91667355f43f5df7335a536b395cda788b90b35ef18e2b53bc4c02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://600bc3bb551b747cf2f18e26745883232d7a37ac5877fad0c67100e52d7a43aa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78f74982180048eb50b9be79fa72c04f62f925b54fe511920f37e6407f864631\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:26Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:26 crc kubenswrapper[4754]: I1005 20:55:26.437716 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:26Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:26 crc kubenswrapper[4754]: I1005 20:55:26.449998 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9ad070bb699b643d29b82bcc1126ce3b585f7492f5ae2817fa491b309b1691d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:26Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:26 crc kubenswrapper[4754]: I1005 20:55:26.462016 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-57sr4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e854c7b9-1229-4da4-80e9-fea6479459b8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceb577e1055d84007936a50d0b3d54b9ae0299a1fc1a6618b37163a6d3fbb9cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vw2ps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:58Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-57sr4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:26Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:26 crc kubenswrapper[4754]: I1005 20:55:26.465338 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:26 crc kubenswrapper[4754]: I1005 20:55:26.465380 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:26 crc kubenswrapper[4754]: I1005 20:55:26.465389 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:26 crc kubenswrapper[4754]: I1005 20:55:26.465408 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:26 crc kubenswrapper[4754]: I1005 20:55:26.465420 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:26Z","lastTransitionTime":"2025-10-05T20:55:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:26 crc kubenswrapper[4754]: I1005 20:55:26.478795 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cn76s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02d5e3f9-73c2-4496-9aca-0787184aef19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7ba2b076839f4a5f586d2e6c4f4d7188254128b782bfd2180e22bf3ea5ffdd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpx5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cn76s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:26Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:26 crc kubenswrapper[4754]: I1005 20:55:26.490555 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ss7lm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f9ef817-0622-4185-a7fb-57221690aa75\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ba613b0d7753a003c572c3733ed394707c45b6f8072cb7aa88f3c7312b8882b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wpwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:55:02Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ss7lm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:26Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:26 crc kubenswrapper[4754]: I1005 20:55:26.507714 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7506d496-65ff-41dc-8230-24c8ee653a2a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33da5a6bb3cd3b042c193bc571e6622f2f87f27e560cdd0bbb9be31e2010d707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df6d700a947d19c84b4f65fd2dd2f8d1f4f1f80fa9b4800c2acfa8dae3912639\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35ca8ee77a0e58cf969bd67b277a3392614c4e73b963e6f6002097709370715d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e32c2154f46ba10c509fe40c357304aa8e82d04e9906bbaf2ccecd47b42279e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8907b9b69b249c34ec6db1e460d641b31711b75d6d31d4caafd867339e6571d0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1005 20:54:52.589115 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 20:54:52.590181 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1304624737/tls.crt::/tmp/serving-cert-1304624737/tls.key\\\\\\\"\\\\nI1005 20:54:58.070544 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1005 20:54:58.090352 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1005 20:54:58.090382 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1005 20:54:58.090412 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1005 20:54:58.090418 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1005 20:54:58.144656 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1005 20:54:58.144688 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 20:54:58.144695 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 20:54:58.144699 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1005 20:54:58.144702 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1005 20:54:58.144705 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1005 20:54:58.144708 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1005 20:54:58.145024 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1005 20:54:58.148112 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4efb34f3493cc8854bc4cff6b29f63973c52b754404d7a21daec4e1f612d327d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54c568b637f3737667433047b956c7eb22fdf0f9301acdfd3b49fb84c39de6ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54c568b637f3737667433047b956c7eb22fdf0f9301acdfd3b49fb84c39de6ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:26Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:26 crc kubenswrapper[4754]: I1005 20:55:26.525057 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c76e6e4740035a6ca54191a770c4516e28101aae69afbb89739174c2f076ef5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9eededc0cb1af94fb7b5daba08785f6b077a85e9cb5ddd8b9f397b658194f61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:26Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:26 crc kubenswrapper[4754]: I1005 20:55:26.550995 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tskt7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a68e9c89-465b-4c01-a470-e21a480ee465\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba6edba14cf321b4cd2f7bb44067a3ecf49f1594e76c0197e807ff3daa45e9b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d14ec99610d2167b484bd7712af7bb6d993d26c74b2828da1c89314f16fda22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d14ec99610d2167b484bd7712af7bb6d993d26c74b2828da1c89314f16fda22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f6f08dd853b25881e2f3c66206aa80d0c493518284217780aa40ac4b69f0c0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f6f08dd853b25881e2f3c66206aa80d0c493518284217780aa40ac4b69f0c0b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://71026d00ec1a3d91b019deb3c9f43ee52fb05b8f2d6374e1a5b762c9e5c1b10d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71026d00ec1a3d91b019deb3c9f43ee52fb05b8f2d6374e1a5b762c9e5c1b10d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b9910d2470acb16ebb436d7030b3161061607cb2a5ef50d482d8d5c4f48a8ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b9910d2470acb16ebb436d7030b3161061607cb2a5ef50d482d8d5c4f48a8ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0a93b61eacc55111995e69c15f89a210b8aa15468ac9ac053b7cf6b007d301e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b0a93b61eacc55111995e69c15f89a210b8aa15468ac9ac053b7cf6b007d301e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cafb3208e7e3bcdd01898a68fc5bbf4cb52ca0800b879b77a60bb102794257f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cafb3208e7e3bcdd01898a68fc5bbf4cb52ca0800b879b77a60bb102794257f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tskt7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:26Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:26 crc kubenswrapper[4754]: I1005 20:55:26.565007 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be95a413-6a59-45b4-84b7-b43cae694a26\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10b211cf6e47dbd1ee42c82e49de0a13a45581b92a0c4f6bbba0326c7149e41d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2ggg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4bb8c5903e8b70dd19eebb111097d96f725e17de44e06c3e74c067376f6c7360\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2ggg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-b2h9k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:26Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:26 crc kubenswrapper[4754]: I1005 20:55:26.568323 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:26 crc kubenswrapper[4754]: I1005 20:55:26.568357 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:26 crc kubenswrapper[4754]: I1005 20:55:26.568369 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:26 crc kubenswrapper[4754]: I1005 20:55:26.568396 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:26 crc kubenswrapper[4754]: I1005 20:55:26.568407 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:26Z","lastTransitionTime":"2025-10-05T20:55:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:26 crc kubenswrapper[4754]: I1005 20:55:26.579039 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rddkl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"15afc2a1-5cb7-41c6-954b-446c36822f50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://866d85ee126f8641d12f1a05fabc3de7b9392c324e9fe06ea92214155c05334c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqv8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a175565f83cfcba7825006eafb0f83efb70ec3d415c2c039c6499d5f6576e89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqv8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:55:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rddkl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:26Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:26 crc kubenswrapper[4754]: I1005 20:55:26.593756 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e380f9b2a0e319dc325c54d1247a6db98346292548fa688cb41a2ca1b3048419\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:26Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:26 crc kubenswrapper[4754]: I1005 20:55:26.607764 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:26Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:26 crc kubenswrapper[4754]: I1005 20:55:26.671546 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:26 crc kubenswrapper[4754]: I1005 20:55:26.671603 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:26 crc kubenswrapper[4754]: I1005 20:55:26.671622 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:26 crc kubenswrapper[4754]: I1005 20:55:26.671654 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:26 crc kubenswrapper[4754]: I1005 20:55:26.671673 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:26Z","lastTransitionTime":"2025-10-05T20:55:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:26 crc kubenswrapper[4754]: I1005 20:55:26.775280 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:26 crc kubenswrapper[4754]: I1005 20:55:26.775354 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:26 crc kubenswrapper[4754]: I1005 20:55:26.775375 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:26 crc kubenswrapper[4754]: I1005 20:55:26.775404 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:26 crc kubenswrapper[4754]: I1005 20:55:26.775428 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:26Z","lastTransitionTime":"2025-10-05T20:55:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:26 crc kubenswrapper[4754]: I1005 20:55:26.836638 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 20:55:26 crc kubenswrapper[4754]: E1005 20:55:26.836859 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 20:55:26 crc kubenswrapper[4754]: I1005 20:55:26.837334 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 20:55:26 crc kubenswrapper[4754]: I1005 20:55:26.837459 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 20:55:26 crc kubenswrapper[4754]: E1005 20:55:26.837609 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 20:55:26 crc kubenswrapper[4754]: E1005 20:55:26.837697 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 20:55:26 crc kubenswrapper[4754]: I1005 20:55:26.863870 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7506d496-65ff-41dc-8230-24c8ee653a2a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33da5a6bb3cd3b042c193bc571e6622f2f87f27e560cdd0bbb9be31e2010d707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df6d700a947d19c84b4f65fd2dd2f8d1f4f1f80fa9b4800c2acfa8dae3912639\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35ca8ee77a0e58cf969bd67b277a3392614c4e73b963e6f6002097709370715d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e32c2154f46ba10c509fe40c357304aa8e82d04e9906bbaf2ccecd47b42279e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8907b9b69b249c34ec6db1e460d641b31711b75d6d31d4caafd867339e6571d0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1005 20:54:52.589115 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 20:54:52.590181 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1304624737/tls.crt::/tmp/serving-cert-1304624737/tls.key\\\\\\\"\\\\nI1005 20:54:58.070544 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1005 20:54:58.090352 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1005 20:54:58.090382 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1005 20:54:58.090412 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1005 20:54:58.090418 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1005 20:54:58.144656 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1005 20:54:58.144688 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 20:54:58.144695 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 20:54:58.144699 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1005 20:54:58.144702 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1005 20:54:58.144705 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1005 20:54:58.144708 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1005 20:54:58.145024 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1005 20:54:58.148112 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4efb34f3493cc8854bc4cff6b29f63973c52b754404d7a21daec4e1f612d327d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54c568b637f3737667433047b956c7eb22fdf0f9301acdfd3b49fb84c39de6ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54c568b637f3737667433047b956c7eb22fdf0f9301acdfd3b49fb84c39de6ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:26Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:26 crc kubenswrapper[4754]: I1005 20:55:26.878992 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:26 crc kubenswrapper[4754]: I1005 20:55:26.879046 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:26 crc kubenswrapper[4754]: I1005 20:55:26.879066 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:26 crc kubenswrapper[4754]: I1005 20:55:26.879101 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:26 crc kubenswrapper[4754]: I1005 20:55:26.879122 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:26Z","lastTransitionTime":"2025-10-05T20:55:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:26 crc kubenswrapper[4754]: I1005 20:55:26.886197 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c76e6e4740035a6ca54191a770c4516e28101aae69afbb89739174c2f076ef5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9eededc0cb1af94fb7b5daba08785f6b077a85e9cb5ddd8b9f397b658194f61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:26Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:26 crc kubenswrapper[4754]: I1005 20:55:26.912900 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tskt7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a68e9c89-465b-4c01-a470-e21a480ee465\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba6edba14cf321b4cd2f7bb44067a3ecf49f1594e76c0197e807ff3daa45e9b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d14ec99610d2167b484bd7712af7bb6d993d26c74b2828da1c89314f16fda22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d14ec99610d2167b484bd7712af7bb6d993d26c74b2828da1c89314f16fda22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f6f08dd853b25881e2f3c66206aa80d0c493518284217780aa40ac4b69f0c0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f6f08dd853b25881e2f3c66206aa80d0c493518284217780aa40ac4b69f0c0b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://71026d00ec1a3d91b019deb3c9f43ee52fb05b8f2d6374e1a5b762c9e5c1b10d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71026d00ec1a3d91b019deb3c9f43ee52fb05b8f2d6374e1a5b762c9e5c1b10d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b9910d2470acb16ebb436d7030b3161061607cb2a5ef50d482d8d5c4f48a8ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b9910d2470acb16ebb436d7030b3161061607cb2a5ef50d482d8d5c4f48a8ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0a93b61eacc55111995e69c15f89a210b8aa15468ac9ac053b7cf6b007d301e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b0a93b61eacc55111995e69c15f89a210b8aa15468ac9ac053b7cf6b007d301e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cafb3208e7e3bcdd01898a68fc5bbf4cb52ca0800b879b77a60bb102794257f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cafb3208e7e3bcdd01898a68fc5bbf4cb52ca0800b879b77a60bb102794257f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tskt7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:26Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:26 crc kubenswrapper[4754]: I1005 20:55:26.935576 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be95a413-6a59-45b4-84b7-b43cae694a26\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10b211cf6e47dbd1ee42c82e49de0a13a45581b92a0c4f6bbba0326c7149e41d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2ggg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4bb8c5903e8b70dd19eebb111097d96f725e17de44e06c3e74c067376f6c7360\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2ggg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-b2h9k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:26Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:26 crc kubenswrapper[4754]: I1005 20:55:26.954404 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rddkl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"15afc2a1-5cb7-41c6-954b-446c36822f50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://866d85ee126f8641d12f1a05fabc3de7b9392c324e9fe06ea92214155c05334c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqv8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a175565f83cfcba7825006eafb0f83efb70ec3d415c2c039c6499d5f6576e89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqv8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:55:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rddkl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:26Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:26 crc kubenswrapper[4754]: I1005 20:55:26.976352 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e380f9b2a0e319dc325c54d1247a6db98346292548fa688cb41a2ca1b3048419\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:26Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:26 crc kubenswrapper[4754]: I1005 20:55:26.983916 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:26 crc kubenswrapper[4754]: I1005 20:55:26.983970 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:26 crc kubenswrapper[4754]: I1005 20:55:26.983990 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:26 crc kubenswrapper[4754]: I1005 20:55:26.984022 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:26 crc kubenswrapper[4754]: I1005 20:55:26.984045 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:26Z","lastTransitionTime":"2025-10-05T20:55:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:26 crc kubenswrapper[4754]: I1005 20:55:26.993686 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:26Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:27 crc kubenswrapper[4754]: I1005 20:55:27.021349 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9309179-b92a-4fa4-9fba-747b8f911936\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d969d3d50d6ad280fc6654df9441bab238bb538f277ed6d1273779813ffbaaed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://800a731ed8ef57714b2318af14054ef23f0368936630887a093fc77167b33d12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8cd9b08e83dc174413cc37891c14562816787028471a4201680c86dd620b7a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93b729f3f7f993db0d6bc3ba60c04930aada3cef8272ffcc3c6a8a242a9b1475\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dde1519111fc3af5d14463fdb58d4f5c62a7045e7de98c60981c7e4da37b0e7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ea160a5b37aea7f2970ae4dd2614b09d3dcf5dcbd1fc103b9997f8ecc3a804c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ea160a5b37aea7f2970ae4dd2614b09d3dcf5dcbd1fc103b9997f8ecc3a804c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://17ae6aa6c53e7dd061f70b41428ef50bbc7ffcc65352fad1a4a9d3c69dbd2ba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17ae6aa6c53e7dd061f70b41428ef50bbc7ffcc65352fad1a4a9d3c69dbd2ba9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://ad4cd69df7608c7d87d839fe09d4557c674c7efbc4730f4e4d3daaf9a26f5aee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad4cd69df7608c7d87d839fe09d4557c674c7efbc4730f4e4d3daaf9a26f5aee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:36Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:27Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:27 crc kubenswrapper[4754]: I1005 20:55:27.039621 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:27Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:27 crc kubenswrapper[4754]: I1005 20:55:27.074187 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b44790fb-fecb-4ec3-9816-8e23dd72ab33\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a7dc53e7f96e48213d724a2cb4859514df925005644108ae3a0c33b14dcf52d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfdd8b1e542ab676576c311e2923fb02861f92de91fe1f4b3e4b968c549c375d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3a63451ffaa14cacc3b78d11340f4ea81445fa3af96178e9b172d063d9697d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3cfd3f2274a8a46fa36660e94ac6c9c298c791283f12d52330c5594ac131998\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f626cb108716da0b595844ea0251b7a04d49b5908a1d6dd0aef5cb2d402c4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://822223f515b5add6d71d193fd5bc77df029b8d37a1cc07be4145b53c7226bf61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af1c313e0dc0fe3b48feac3cce1ecd4452c9efadf3d19f257c07cc7fcdec875f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cbed906d679176209043edd5dcc649f5ba7d5848a10cfd09efa9e656dcacfef\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-05T20:55:11Z\\\",\\\"message\\\":\\\"gressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1005 20:55:11.136172 6140 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1005 20:55:11.136316 6140 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1005 20:55:11.136624 6140 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1005 20:55:11.136664 6140 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1005 20:55:11.136709 6140 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1005 20:55:11.136749 6140 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1005 20:55:11.138444 6140 factory.go:656] Stopping watch factory\\\\nI1005 20:55:11.148575 6140 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1005 20:55:11.148612 6140 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1005 20:55:11.148701 6140 ovnkube.go:599] Stopped ovnkube\\\\nI1005 20:55:11.148748 6140 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1005 20:55:11.148871 6140 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:10Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2acf9637fa25e5d3ffeb05224c11218c47defc679568fb928a498869613de4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35ebeb7a87d5d47b6567917d38eca38a3e622c497fba58971296c4609c287423\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35ebeb7a87d5d47b6567917d38eca38a3e622c497fba58971296c4609c287423\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7rhps\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:27Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:27 crc kubenswrapper[4754]: I1005 20:55:27.086897 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:27 crc kubenswrapper[4754]: I1005 20:55:27.087284 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:27 crc kubenswrapper[4754]: I1005 20:55:27.087368 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:27 crc kubenswrapper[4754]: I1005 20:55:27.087514 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:27 crc kubenswrapper[4754]: I1005 20:55:27.087626 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:27Z","lastTransitionTime":"2025-10-05T20:55:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:27 crc kubenswrapper[4754]: I1005 20:55:27.098656 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"934bcd66-9c03-48b9-82dd-5093338f9c8f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9042df2f43243d3d79ba9a3940f8dd4f6c7a89ef9e3510e81fe60b1944d7e83b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5366bbf46a91667355f43f5df7335a536b395cda788b90b35ef18e2b53bc4c02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://600bc3bb551b747cf2f18e26745883232d7a37ac5877fad0c67100e52d7a43aa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78f74982180048eb50b9be79fa72c04f62f925b54fe511920f37e6407f864631\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:27Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:27 crc kubenswrapper[4754]: I1005 20:55:27.116780 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:27Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:27 crc kubenswrapper[4754]: I1005 20:55:27.137383 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9ad070bb699b643d29b82bcc1126ce3b585f7492f5ae2817fa491b309b1691d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:27Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:27 crc kubenswrapper[4754]: I1005 20:55:27.154752 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-57sr4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e854c7b9-1229-4da4-80e9-fea6479459b8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceb577e1055d84007936a50d0b3d54b9ae0299a1fc1a6618b37163a6d3fbb9cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vw2ps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:58Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-57sr4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:27Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:27 crc kubenswrapper[4754]: I1005 20:55:27.180588 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cn76s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02d5e3f9-73c2-4496-9aca-0787184aef19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7ba2b076839f4a5f586d2e6c4f4d7188254128b782bfd2180e22bf3ea5ffdd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpx5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cn76s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:27Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:27 crc kubenswrapper[4754]: I1005 20:55:27.194684 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:27 crc kubenswrapper[4754]: I1005 20:55:27.194760 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:27 crc kubenswrapper[4754]: I1005 20:55:27.194782 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:27 crc kubenswrapper[4754]: I1005 20:55:27.194812 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:27 crc kubenswrapper[4754]: I1005 20:55:27.194833 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:27Z","lastTransitionTime":"2025-10-05T20:55:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:27 crc kubenswrapper[4754]: I1005 20:55:27.198089 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ss7lm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f9ef817-0622-4185-a7fb-57221690aa75\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ba613b0d7753a003c572c3733ed394707c45b6f8072cb7aa88f3c7312b8882b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wpwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:55:02Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ss7lm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:27Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:27 crc kubenswrapper[4754]: I1005 20:55:27.214314 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-nwrnt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7dc728ea-2601-44f0-bcab-2913f034007d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-846xq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-846xq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:55:13Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-nwrnt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:27Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:27 crc kubenswrapper[4754]: I1005 20:55:27.298593 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:27 crc kubenswrapper[4754]: I1005 20:55:27.298665 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:27 crc kubenswrapper[4754]: I1005 20:55:27.298684 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:27 crc kubenswrapper[4754]: I1005 20:55:27.298711 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:27 crc kubenswrapper[4754]: I1005 20:55:27.298731 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:27Z","lastTransitionTime":"2025-10-05T20:55:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:27 crc kubenswrapper[4754]: I1005 20:55:27.300325 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7rhps_b44790fb-fecb-4ec3-9816-8e23dd72ab33/ovnkube-controller/2.log" Oct 05 20:55:27 crc kubenswrapper[4754]: I1005 20:55:27.301693 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7rhps_b44790fb-fecb-4ec3-9816-8e23dd72ab33/ovnkube-controller/1.log" Oct 05 20:55:27 crc kubenswrapper[4754]: I1005 20:55:27.307126 4754 generic.go:334] "Generic (PLEG): container finished" podID="b44790fb-fecb-4ec3-9816-8e23dd72ab33" containerID="af1c313e0dc0fe3b48feac3cce1ecd4452c9efadf3d19f257c07cc7fcdec875f" exitCode=1 Oct 05 20:55:27 crc kubenswrapper[4754]: I1005 20:55:27.307220 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" event={"ID":"b44790fb-fecb-4ec3-9816-8e23dd72ab33","Type":"ContainerDied","Data":"af1c313e0dc0fe3b48feac3cce1ecd4452c9efadf3d19f257c07cc7fcdec875f"} Oct 05 20:55:27 crc kubenswrapper[4754]: I1005 20:55:27.307294 4754 scope.go:117] "RemoveContainer" containerID="2cbed906d679176209043edd5dcc649f5ba7d5848a10cfd09efa9e656dcacfef" Oct 05 20:55:27 crc kubenswrapper[4754]: I1005 20:55:27.311731 4754 scope.go:117] "RemoveContainer" containerID="af1c313e0dc0fe3b48feac3cce1ecd4452c9efadf3d19f257c07cc7fcdec875f" Oct 05 20:55:27 crc kubenswrapper[4754]: E1005 20:55:27.316049 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-7rhps_openshift-ovn-kubernetes(b44790fb-fecb-4ec3-9816-8e23dd72ab33)\"" pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" podUID="b44790fb-fecb-4ec3-9816-8e23dd72ab33" Oct 05 20:55:27 crc kubenswrapper[4754]: I1005 20:55:27.331689 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ss7lm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f9ef817-0622-4185-a7fb-57221690aa75\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ba613b0d7753a003c572c3733ed394707c45b6f8072cb7aa88f3c7312b8882b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wpwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:55:02Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ss7lm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:27Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:27 crc kubenswrapper[4754]: I1005 20:55:27.354222 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-nwrnt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7dc728ea-2601-44f0-bcab-2913f034007d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-846xq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-846xq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:55:13Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-nwrnt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:27Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:27 crc kubenswrapper[4754]: I1005 20:55:27.376613 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"934bcd66-9c03-48b9-82dd-5093338f9c8f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9042df2f43243d3d79ba9a3940f8dd4f6c7a89ef9e3510e81fe60b1944d7e83b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5366bbf46a91667355f43f5df7335a536b395cda788b90b35ef18e2b53bc4c02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://600bc3bb551b747cf2f18e26745883232d7a37ac5877fad0c67100e52d7a43aa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78f74982180048eb50b9be79fa72c04f62f925b54fe511920f37e6407f864631\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:27Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:27 crc kubenswrapper[4754]: I1005 20:55:27.398379 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:27Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:27 crc kubenswrapper[4754]: I1005 20:55:27.402960 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:27 crc kubenswrapper[4754]: I1005 20:55:27.403008 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:27 crc kubenswrapper[4754]: I1005 20:55:27.403027 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:27 crc kubenswrapper[4754]: I1005 20:55:27.403059 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:27 crc kubenswrapper[4754]: I1005 20:55:27.403077 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:27Z","lastTransitionTime":"2025-10-05T20:55:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:27 crc kubenswrapper[4754]: I1005 20:55:27.421110 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9ad070bb699b643d29b82bcc1126ce3b585f7492f5ae2817fa491b309b1691d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:27Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:27 crc kubenswrapper[4754]: I1005 20:55:27.439034 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-57sr4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e854c7b9-1229-4da4-80e9-fea6479459b8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceb577e1055d84007936a50d0b3d54b9ae0299a1fc1a6618b37163a6d3fbb9cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vw2ps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:58Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-57sr4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:27Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:27 crc kubenswrapper[4754]: I1005 20:55:27.461254 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cn76s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02d5e3f9-73c2-4496-9aca-0787184aef19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7ba2b076839f4a5f586d2e6c4f4d7188254128b782bfd2180e22bf3ea5ffdd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpx5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cn76s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:27Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:27 crc kubenswrapper[4754]: I1005 20:55:27.470166 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 05 20:55:27 crc kubenswrapper[4754]: I1005 20:55:27.485479 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7506d496-65ff-41dc-8230-24c8ee653a2a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33da5a6bb3cd3b042c193bc571e6622f2f87f27e560cdd0bbb9be31e2010d707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df6d700a947d19c84b4f65fd2dd2f8d1f4f1f80fa9b4800c2acfa8dae3912639\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35ca8ee77a0e58cf969bd67b277a3392614c4e73b963e6f6002097709370715d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e32c2154f46ba10c509fe40c357304aa8e82d04e9906bbaf2ccecd47b42279e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8907b9b69b249c34ec6db1e460d641b31711b75d6d31d4caafd867339e6571d0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1005 20:54:52.589115 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 20:54:52.590181 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1304624737/tls.crt::/tmp/serving-cert-1304624737/tls.key\\\\\\\"\\\\nI1005 20:54:58.070544 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1005 20:54:58.090352 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1005 20:54:58.090382 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1005 20:54:58.090412 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1005 20:54:58.090418 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1005 20:54:58.144656 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1005 20:54:58.144688 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 20:54:58.144695 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 20:54:58.144699 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1005 20:54:58.144702 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1005 20:54:58.144705 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1005 20:54:58.144708 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1005 20:54:58.145024 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1005 20:54:58.148112 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4efb34f3493cc8854bc4cff6b29f63973c52b754404d7a21daec4e1f612d327d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54c568b637f3737667433047b956c7eb22fdf0f9301acdfd3b49fb84c39de6ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54c568b637f3737667433047b956c7eb22fdf0f9301acdfd3b49fb84c39de6ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:27Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:27 crc kubenswrapper[4754]: I1005 20:55:27.506207 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:27 crc kubenswrapper[4754]: I1005 20:55:27.506259 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:27 crc kubenswrapper[4754]: I1005 20:55:27.506271 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:27 crc kubenswrapper[4754]: I1005 20:55:27.506291 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:27 crc kubenswrapper[4754]: I1005 20:55:27.506304 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:27Z","lastTransitionTime":"2025-10-05T20:55:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:27 crc kubenswrapper[4754]: I1005 20:55:27.512716 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c76e6e4740035a6ca54191a770c4516e28101aae69afbb89739174c2f076ef5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9eededc0cb1af94fb7b5daba08785f6b077a85e9cb5ddd8b9f397b658194f61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:27Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:27 crc kubenswrapper[4754]: I1005 20:55:27.542606 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tskt7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a68e9c89-465b-4c01-a470-e21a480ee465\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba6edba14cf321b4cd2f7bb44067a3ecf49f1594e76c0197e807ff3daa45e9b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d14ec99610d2167b484bd7712af7bb6d993d26c74b2828da1c89314f16fda22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d14ec99610d2167b484bd7712af7bb6d993d26c74b2828da1c89314f16fda22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f6f08dd853b25881e2f3c66206aa80d0c493518284217780aa40ac4b69f0c0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f6f08dd853b25881e2f3c66206aa80d0c493518284217780aa40ac4b69f0c0b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://71026d00ec1a3d91b019deb3c9f43ee52fb05b8f2d6374e1a5b762c9e5c1b10d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71026d00ec1a3d91b019deb3c9f43ee52fb05b8f2d6374e1a5b762c9e5c1b10d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b9910d2470acb16ebb436d7030b3161061607cb2a5ef50d482d8d5c4f48a8ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b9910d2470acb16ebb436d7030b3161061607cb2a5ef50d482d8d5c4f48a8ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0a93b61eacc55111995e69c15f89a210b8aa15468ac9ac053b7cf6b007d301e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b0a93b61eacc55111995e69c15f89a210b8aa15468ac9ac053b7cf6b007d301e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cafb3208e7e3bcdd01898a68fc5bbf4cb52ca0800b879b77a60bb102794257f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cafb3208e7e3bcdd01898a68fc5bbf4cb52ca0800b879b77a60bb102794257f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tskt7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:27Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:27 crc kubenswrapper[4754]: I1005 20:55:27.560827 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be95a413-6a59-45b4-84b7-b43cae694a26\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10b211cf6e47dbd1ee42c82e49de0a13a45581b92a0c4f6bbba0326c7149e41d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2ggg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4bb8c5903e8b70dd19eebb111097d96f725e17de44e06c3e74c067376f6c7360\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2ggg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-b2h9k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:27Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:27 crc kubenswrapper[4754]: I1005 20:55:27.579884 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rddkl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"15afc2a1-5cb7-41c6-954b-446c36822f50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://866d85ee126f8641d12f1a05fabc3de7b9392c324e9fe06ea92214155c05334c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqv8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a175565f83cfcba7825006eafb0f83efb70ec3d415c2c039c6499d5f6576e89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqv8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:55:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rddkl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:27Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:27 crc kubenswrapper[4754]: I1005 20:55:27.598336 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e380f9b2a0e319dc325c54d1247a6db98346292548fa688cb41a2ca1b3048419\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:27Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:27 crc kubenswrapper[4754]: I1005 20:55:27.609098 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:27 crc kubenswrapper[4754]: I1005 20:55:27.609137 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:27 crc kubenswrapper[4754]: I1005 20:55:27.609146 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:27 crc kubenswrapper[4754]: I1005 20:55:27.609163 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:27 crc kubenswrapper[4754]: I1005 20:55:27.609173 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:27Z","lastTransitionTime":"2025-10-05T20:55:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:27 crc kubenswrapper[4754]: I1005 20:55:27.615595 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:27Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:27 crc kubenswrapper[4754]: I1005 20:55:27.649235 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9309179-b92a-4fa4-9fba-747b8f911936\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d969d3d50d6ad280fc6654df9441bab238bb538f277ed6d1273779813ffbaaed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://800a731ed8ef57714b2318af14054ef23f0368936630887a093fc77167b33d12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8cd9b08e83dc174413cc37891c14562816787028471a4201680c86dd620b7a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93b729f3f7f993db0d6bc3ba60c04930aada3cef8272ffcc3c6a8a242a9b1475\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dde1519111fc3af5d14463fdb58d4f5c62a7045e7de98c60981c7e4da37b0e7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ea160a5b37aea7f2970ae4dd2614b09d3dcf5dcbd1fc103b9997f8ecc3a804c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ea160a5b37aea7f2970ae4dd2614b09d3dcf5dcbd1fc103b9997f8ecc3a804c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://17ae6aa6c53e7dd061f70b41428ef50bbc7ffcc65352fad1a4a9d3c69dbd2ba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17ae6aa6c53e7dd061f70b41428ef50bbc7ffcc65352fad1a4a9d3c69dbd2ba9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://ad4cd69df7608c7d87d839fe09d4557c674c7efbc4730f4e4d3daaf9a26f5aee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad4cd69df7608c7d87d839fe09d4557c674c7efbc4730f4e4d3daaf9a26f5aee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:36Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:27Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:27 crc kubenswrapper[4754]: I1005 20:55:27.662734 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:27Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:27 crc kubenswrapper[4754]: I1005 20:55:27.683889 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b44790fb-fecb-4ec3-9816-8e23dd72ab33\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a7dc53e7f96e48213d724a2cb4859514df925005644108ae3a0c33b14dcf52d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfdd8b1e542ab676576c311e2923fb02861f92de91fe1f4b3e4b968c549c375d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3a63451ffaa14cacc3b78d11340f4ea81445fa3af96178e9b172d063d9697d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3cfd3f2274a8a46fa36660e94ac6c9c298c791283f12d52330c5594ac131998\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f626cb108716da0b595844ea0251b7a04d49b5908a1d6dd0aef5cb2d402c4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://822223f515b5add6d71d193fd5bc77df029b8d37a1cc07be4145b53c7226bf61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af1c313e0dc0fe3b48feac3cce1ecd4452c9efadf3d19f257c07cc7fcdec875f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cbed906d679176209043edd5dcc649f5ba7d5848a10cfd09efa9e656dcacfef\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-05T20:55:11Z\\\",\\\"message\\\":\\\"gressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1005 20:55:11.136172 6140 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1005 20:55:11.136316 6140 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1005 20:55:11.136624 6140 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1005 20:55:11.136664 6140 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1005 20:55:11.136709 6140 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1005 20:55:11.136749 6140 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1005 20:55:11.138444 6140 factory.go:656] Stopping watch factory\\\\nI1005 20:55:11.148575 6140 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1005 20:55:11.148612 6140 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1005 20:55:11.148701 6140 ovnkube.go:599] Stopped ovnkube\\\\nI1005 20:55:11.148748 6140 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1005 20:55:11.148871 6140 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:10Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af1c313e0dc0fe3b48feac3cce1ecd4452c9efadf3d19f257c07cc7fcdec875f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-05T20:55:26Z\\\",\\\"message\\\":\\\"5:26.770640 6339 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/iptables-alerter-4ln5h in node crc\\\\nI1005 20:55:26.770478 6339 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-apiserver-operator/metrics]} name:Service_openshift-apiserver-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.38:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {8b82f026-5975-4a1b-bb18-08d5d51147ec}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1005 20:55:26.770652 6339 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/iptables-alerter-4ln5h after 0 failed attempt(s)\\\\nI1005 20:55:26.770662 6339 default_network_controller.go:776] Recording success event on pod openshift-network-operator/iptables-alerter-4ln5h\\\\nF1005 20:55:26.770677 6339 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controlle\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2acf9637fa25e5d3ffeb05224c11218c47defc679568fb928a498869613de4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35ebeb7a87d5d47b6567917d38eca38a3e622c497fba58971296c4609c287423\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35ebeb7a87d5d47b6567917d38eca38a3e622c497fba58971296c4609c287423\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7rhps\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:27Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:27 crc kubenswrapper[4754]: I1005 20:55:27.695173 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-nwrnt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7dc728ea-2601-44f0-bcab-2913f034007d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-846xq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-846xq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:55:13Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-nwrnt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:27Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:27 crc kubenswrapper[4754]: I1005 20:55:27.707286 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"934bcd66-9c03-48b9-82dd-5093338f9c8f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9042df2f43243d3d79ba9a3940f8dd4f6c7a89ef9e3510e81fe60b1944d7e83b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5366bbf46a91667355f43f5df7335a536b395cda788b90b35ef18e2b53bc4c02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://600bc3bb551b747cf2f18e26745883232d7a37ac5877fad0c67100e52d7a43aa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78f74982180048eb50b9be79fa72c04f62f925b54fe511920f37e6407f864631\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:27Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:27 crc kubenswrapper[4754]: I1005 20:55:27.715076 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:27 crc kubenswrapper[4754]: I1005 20:55:27.715113 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:27 crc kubenswrapper[4754]: I1005 20:55:27.715126 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:27 crc kubenswrapper[4754]: I1005 20:55:27.715147 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:27 crc kubenswrapper[4754]: I1005 20:55:27.715162 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:27Z","lastTransitionTime":"2025-10-05T20:55:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:27 crc kubenswrapper[4754]: I1005 20:55:27.720419 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:27Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:27 crc kubenswrapper[4754]: I1005 20:55:27.731544 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9ad070bb699b643d29b82bcc1126ce3b585f7492f5ae2817fa491b309b1691d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:27Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:27 crc kubenswrapper[4754]: I1005 20:55:27.743436 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-57sr4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e854c7b9-1229-4da4-80e9-fea6479459b8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceb577e1055d84007936a50d0b3d54b9ae0299a1fc1a6618b37163a6d3fbb9cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vw2ps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:58Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-57sr4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:27Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:27 crc kubenswrapper[4754]: I1005 20:55:27.757435 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cn76s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02d5e3f9-73c2-4496-9aca-0787184aef19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7ba2b076839f4a5f586d2e6c4f4d7188254128b782bfd2180e22bf3ea5ffdd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpx5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cn76s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:27Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:27 crc kubenswrapper[4754]: I1005 20:55:27.770733 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ss7lm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f9ef817-0622-4185-a7fb-57221690aa75\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ba613b0d7753a003c572c3733ed394707c45b6f8072cb7aa88f3c7312b8882b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wpwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:55:02Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ss7lm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:27Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:27 crc kubenswrapper[4754]: I1005 20:55:27.786693 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7506d496-65ff-41dc-8230-24c8ee653a2a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33da5a6bb3cd3b042c193bc571e6622f2f87f27e560cdd0bbb9be31e2010d707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df6d700a947d19c84b4f65fd2dd2f8d1f4f1f80fa9b4800c2acfa8dae3912639\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35ca8ee77a0e58cf969bd67b277a3392614c4e73b963e6f6002097709370715d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e32c2154f46ba10c509fe40c357304aa8e82d04e9906bbaf2ccecd47b42279e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8907b9b69b249c34ec6db1e460d641b31711b75d6d31d4caafd867339e6571d0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1005 20:54:52.589115 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 20:54:52.590181 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1304624737/tls.crt::/tmp/serving-cert-1304624737/tls.key\\\\\\\"\\\\nI1005 20:54:58.070544 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1005 20:54:58.090352 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1005 20:54:58.090382 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1005 20:54:58.090412 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1005 20:54:58.090418 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1005 20:54:58.144656 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1005 20:54:58.144688 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 20:54:58.144695 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 20:54:58.144699 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1005 20:54:58.144702 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1005 20:54:58.144705 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1005 20:54:58.144708 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1005 20:54:58.145024 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1005 20:54:58.148112 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4efb34f3493cc8854bc4cff6b29f63973c52b754404d7a21daec4e1f612d327d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54c568b637f3737667433047b956c7eb22fdf0f9301acdfd3b49fb84c39de6ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54c568b637f3737667433047b956c7eb22fdf0f9301acdfd3b49fb84c39de6ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:27Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:27 crc kubenswrapper[4754]: I1005 20:55:27.803792 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c76e6e4740035a6ca54191a770c4516e28101aae69afbb89739174c2f076ef5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9eededc0cb1af94fb7b5daba08785f6b077a85e9cb5ddd8b9f397b658194f61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:27Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:27 crc kubenswrapper[4754]: I1005 20:55:27.817613 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:27 crc kubenswrapper[4754]: I1005 20:55:27.817669 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:27 crc kubenswrapper[4754]: I1005 20:55:27.817692 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:27 crc kubenswrapper[4754]: I1005 20:55:27.817719 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:27 crc kubenswrapper[4754]: I1005 20:55:27.817738 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:27Z","lastTransitionTime":"2025-10-05T20:55:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:27 crc kubenswrapper[4754]: I1005 20:55:27.819005 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tskt7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a68e9c89-465b-4c01-a470-e21a480ee465\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba6edba14cf321b4cd2f7bb44067a3ecf49f1594e76c0197e807ff3daa45e9b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d14ec99610d2167b484bd7712af7bb6d993d26c74b2828da1c89314f16fda22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d14ec99610d2167b484bd7712af7bb6d993d26c74b2828da1c89314f16fda22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f6f08dd853b25881e2f3c66206aa80d0c493518284217780aa40ac4b69f0c0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f6f08dd853b25881e2f3c66206aa80d0c493518284217780aa40ac4b69f0c0b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://71026d00ec1a3d91b019deb3c9f43ee52fb05b8f2d6374e1a5b762c9e5c1b10d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71026d00ec1a3d91b019deb3c9f43ee52fb05b8f2d6374e1a5b762c9e5c1b10d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b9910d2470acb16ebb436d7030b3161061607cb2a5ef50d482d8d5c4f48a8ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b9910d2470acb16ebb436d7030b3161061607cb2a5ef50d482d8d5c4f48a8ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0a93b61eacc55111995e69c15f89a210b8aa15468ac9ac053b7cf6b007d301e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b0a93b61eacc55111995e69c15f89a210b8aa15468ac9ac053b7cf6b007d301e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cafb3208e7e3bcdd01898a68fc5bbf4cb52ca0800b879b77a60bb102794257f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cafb3208e7e3bcdd01898a68fc5bbf4cb52ca0800b879b77a60bb102794257f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tskt7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:27Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:27 crc kubenswrapper[4754]: I1005 20:55:27.835448 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be95a413-6a59-45b4-84b7-b43cae694a26\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10b211cf6e47dbd1ee42c82e49de0a13a45581b92a0c4f6bbba0326c7149e41d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2ggg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4bb8c5903e8b70dd19eebb111097d96f725e17de44e06c3e74c067376f6c7360\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2ggg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-b2h9k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:27Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:27 crc kubenswrapper[4754]: I1005 20:55:27.836648 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nwrnt" Oct 05 20:55:27 crc kubenswrapper[4754]: E1005 20:55:27.836843 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nwrnt" podUID="7dc728ea-2601-44f0-bcab-2913f034007d" Oct 05 20:55:27 crc kubenswrapper[4754]: I1005 20:55:27.854168 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rddkl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"15afc2a1-5cb7-41c6-954b-446c36822f50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://866d85ee126f8641d12f1a05fabc3de7b9392c324e9fe06ea92214155c05334c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqv8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a175565f83cfcba7825006eafb0f83efb70ec3d415c2c039c6499d5f6576e89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqv8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:55:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rddkl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:27Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:27 crc kubenswrapper[4754]: I1005 20:55:27.870773 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e380f9b2a0e319dc325c54d1247a6db98346292548fa688cb41a2ca1b3048419\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:27Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:27 crc kubenswrapper[4754]: I1005 20:55:27.887698 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:27Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:27 crc kubenswrapper[4754]: I1005 20:55:27.916762 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9309179-b92a-4fa4-9fba-747b8f911936\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d969d3d50d6ad280fc6654df9441bab238bb538f277ed6d1273779813ffbaaed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://800a731ed8ef57714b2318af14054ef23f0368936630887a093fc77167b33d12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8cd9b08e83dc174413cc37891c14562816787028471a4201680c86dd620b7a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93b729f3f7f993db0d6bc3ba60c04930aada3cef8272ffcc3c6a8a242a9b1475\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dde1519111fc3af5d14463fdb58d4f5c62a7045e7de98c60981c7e4da37b0e7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ea160a5b37aea7f2970ae4dd2614b09d3dcf5dcbd1fc103b9997f8ecc3a804c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ea160a5b37aea7f2970ae4dd2614b09d3dcf5dcbd1fc103b9997f8ecc3a804c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://17ae6aa6c53e7dd061f70b41428ef50bbc7ffcc65352fad1a4a9d3c69dbd2ba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17ae6aa6c53e7dd061f70b41428ef50bbc7ffcc65352fad1a4a9d3c69dbd2ba9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://ad4cd69df7608c7d87d839fe09d4557c674c7efbc4730f4e4d3daaf9a26f5aee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad4cd69df7608c7d87d839fe09d4557c674c7efbc4730f4e4d3daaf9a26f5aee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:36Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:27Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:27 crc kubenswrapper[4754]: I1005 20:55:27.921151 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:27 crc kubenswrapper[4754]: I1005 20:55:27.921229 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:27 crc kubenswrapper[4754]: I1005 20:55:27.921251 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:27 crc kubenswrapper[4754]: I1005 20:55:27.921276 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:27 crc kubenswrapper[4754]: I1005 20:55:27.921294 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:27Z","lastTransitionTime":"2025-10-05T20:55:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:27 crc kubenswrapper[4754]: I1005 20:55:27.930960 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:27Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:27 crc kubenswrapper[4754]: I1005 20:55:27.964307 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b44790fb-fecb-4ec3-9816-8e23dd72ab33\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a7dc53e7f96e48213d724a2cb4859514df925005644108ae3a0c33b14dcf52d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfdd8b1e542ab676576c311e2923fb02861f92de91fe1f4b3e4b968c549c375d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3a63451ffaa14cacc3b78d11340f4ea81445fa3af96178e9b172d063d9697d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3cfd3f2274a8a46fa36660e94ac6c9c298c791283f12d52330c5594ac131998\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f626cb108716da0b595844ea0251b7a04d49b5908a1d6dd0aef5cb2d402c4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://822223f515b5add6d71d193fd5bc77df029b8d37a1cc07be4145b53c7226bf61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af1c313e0dc0fe3b48feac3cce1ecd4452c9efadf3d19f257c07cc7fcdec875f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cbed906d679176209043edd5dcc649f5ba7d5848a10cfd09efa9e656dcacfef\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-05T20:55:11Z\\\",\\\"message\\\":\\\"gressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1005 20:55:11.136172 6140 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1005 20:55:11.136316 6140 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1005 20:55:11.136624 6140 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1005 20:55:11.136664 6140 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1005 20:55:11.136709 6140 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1005 20:55:11.136749 6140 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1005 20:55:11.138444 6140 factory.go:656] Stopping watch factory\\\\nI1005 20:55:11.148575 6140 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1005 20:55:11.148612 6140 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1005 20:55:11.148701 6140 ovnkube.go:599] Stopped ovnkube\\\\nI1005 20:55:11.148748 6140 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1005 20:55:11.148871 6140 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:10Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af1c313e0dc0fe3b48feac3cce1ecd4452c9efadf3d19f257c07cc7fcdec875f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-05T20:55:26Z\\\",\\\"message\\\":\\\"5:26.770640 6339 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/iptables-alerter-4ln5h in node crc\\\\nI1005 20:55:26.770478 6339 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-apiserver-operator/metrics]} name:Service_openshift-apiserver-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.38:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {8b82f026-5975-4a1b-bb18-08d5d51147ec}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1005 20:55:26.770652 6339 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/iptables-alerter-4ln5h after 0 failed attempt(s)\\\\nI1005 20:55:26.770662 6339 default_network_controller.go:776] Recording success event on pod openshift-network-operator/iptables-alerter-4ln5h\\\\nF1005 20:55:26.770677 6339 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controlle\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2acf9637fa25e5d3ffeb05224c11218c47defc679568fb928a498869613de4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35ebeb7a87d5d47b6567917d38eca38a3e622c497fba58971296c4609c287423\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35ebeb7a87d5d47b6567917d38eca38a3e622c497fba58971296c4609c287423\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7rhps\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:27Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:28 crc kubenswrapper[4754]: I1005 20:55:28.024249 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:28 crc kubenswrapper[4754]: I1005 20:55:28.024588 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:28 crc kubenswrapper[4754]: I1005 20:55:28.024723 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:28 crc kubenswrapper[4754]: I1005 20:55:28.024906 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:28 crc kubenswrapper[4754]: I1005 20:55:28.025041 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:28Z","lastTransitionTime":"2025-10-05T20:55:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:28 crc kubenswrapper[4754]: I1005 20:55:28.129699 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:28 crc kubenswrapper[4754]: I1005 20:55:28.129773 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:28 crc kubenswrapper[4754]: I1005 20:55:28.129793 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:28 crc kubenswrapper[4754]: I1005 20:55:28.129822 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:28 crc kubenswrapper[4754]: I1005 20:55:28.129840 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:28Z","lastTransitionTime":"2025-10-05T20:55:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:28 crc kubenswrapper[4754]: I1005 20:55:28.234105 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:28 crc kubenswrapper[4754]: I1005 20:55:28.234161 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:28 crc kubenswrapper[4754]: I1005 20:55:28.234180 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:28 crc kubenswrapper[4754]: I1005 20:55:28.234207 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:28 crc kubenswrapper[4754]: I1005 20:55:28.234224 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:28Z","lastTransitionTime":"2025-10-05T20:55:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:28 crc kubenswrapper[4754]: I1005 20:55:28.315188 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7rhps_b44790fb-fecb-4ec3-9816-8e23dd72ab33/ovnkube-controller/2.log" Oct 05 20:55:28 crc kubenswrapper[4754]: I1005 20:55:28.321354 4754 scope.go:117] "RemoveContainer" containerID="af1c313e0dc0fe3b48feac3cce1ecd4452c9efadf3d19f257c07cc7fcdec875f" Oct 05 20:55:28 crc kubenswrapper[4754]: E1005 20:55:28.321586 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-7rhps_openshift-ovn-kubernetes(b44790fb-fecb-4ec3-9816-8e23dd72ab33)\"" pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" podUID="b44790fb-fecb-4ec3-9816-8e23dd72ab33" Oct 05 20:55:28 crc kubenswrapper[4754]: I1005 20:55:28.337909 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:28 crc kubenswrapper[4754]: I1005 20:55:28.338111 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:28 crc kubenswrapper[4754]: I1005 20:55:28.338288 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:28 crc kubenswrapper[4754]: I1005 20:55:28.338420 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:28 crc kubenswrapper[4754]: I1005 20:55:28.338573 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:28Z","lastTransitionTime":"2025-10-05T20:55:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:28 crc kubenswrapper[4754]: I1005 20:55:28.342576 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:28Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:28 crc kubenswrapper[4754]: I1005 20:55:28.362994 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9ad070bb699b643d29b82bcc1126ce3b585f7492f5ae2817fa491b309b1691d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:28Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:28 crc kubenswrapper[4754]: I1005 20:55:28.379255 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-57sr4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e854c7b9-1229-4da4-80e9-fea6479459b8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceb577e1055d84007936a50d0b3d54b9ae0299a1fc1a6618b37163a6d3fbb9cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vw2ps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:58Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-57sr4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:28Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:28 crc kubenswrapper[4754]: I1005 20:55:28.399681 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cn76s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02d5e3f9-73c2-4496-9aca-0787184aef19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7ba2b076839f4a5f586d2e6c4f4d7188254128b782bfd2180e22bf3ea5ffdd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpx5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cn76s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:28Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:28 crc kubenswrapper[4754]: I1005 20:55:28.415970 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ss7lm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f9ef817-0622-4185-a7fb-57221690aa75\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ba613b0d7753a003c572c3733ed394707c45b6f8072cb7aa88f3c7312b8882b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wpwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:55:02Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ss7lm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:28Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:28 crc kubenswrapper[4754]: I1005 20:55:28.433656 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-nwrnt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7dc728ea-2601-44f0-bcab-2913f034007d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-846xq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-846xq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:55:13Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-nwrnt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:28Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:28 crc kubenswrapper[4754]: I1005 20:55:28.443028 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:28 crc kubenswrapper[4754]: I1005 20:55:28.443085 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:28 crc kubenswrapper[4754]: I1005 20:55:28.443101 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:28 crc kubenswrapper[4754]: I1005 20:55:28.443129 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:28 crc kubenswrapper[4754]: I1005 20:55:28.443156 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:28Z","lastTransitionTime":"2025-10-05T20:55:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:28 crc kubenswrapper[4754]: I1005 20:55:28.455328 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"934bcd66-9c03-48b9-82dd-5093338f9c8f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9042df2f43243d3d79ba9a3940f8dd4f6c7a89ef9e3510e81fe60b1944d7e83b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5366bbf46a91667355f43f5df7335a536b395cda788b90b35ef18e2b53bc4c02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://600bc3bb551b747cf2f18e26745883232d7a37ac5877fad0c67100e52d7a43aa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78f74982180048eb50b9be79fa72c04f62f925b54fe511920f37e6407f864631\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:28Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:28 crc kubenswrapper[4754]: I1005 20:55:28.477159 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c76e6e4740035a6ca54191a770c4516e28101aae69afbb89739174c2f076ef5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9eededc0cb1af94fb7b5daba08785f6b077a85e9cb5ddd8b9f397b658194f61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:28Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:28 crc kubenswrapper[4754]: I1005 20:55:28.494412 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tskt7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a68e9c89-465b-4c01-a470-e21a480ee465\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba6edba14cf321b4cd2f7bb44067a3ecf49f1594e76c0197e807ff3daa45e9b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d14ec99610d2167b484bd7712af7bb6d993d26c74b2828da1c89314f16fda22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d14ec99610d2167b484bd7712af7bb6d993d26c74b2828da1c89314f16fda22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f6f08dd853b25881e2f3c66206aa80d0c493518284217780aa40ac4b69f0c0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f6f08dd853b25881e2f3c66206aa80d0c493518284217780aa40ac4b69f0c0b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://71026d00ec1a3d91b019deb3c9f43ee52fb05b8f2d6374e1a5b762c9e5c1b10d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71026d00ec1a3d91b019deb3c9f43ee52fb05b8f2d6374e1a5b762c9e5c1b10d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b9910d2470acb16ebb436d7030b3161061607cb2a5ef50d482d8d5c4f48a8ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b9910d2470acb16ebb436d7030b3161061607cb2a5ef50d482d8d5c4f48a8ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0a93b61eacc55111995e69c15f89a210b8aa15468ac9ac053b7cf6b007d301e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b0a93b61eacc55111995e69c15f89a210b8aa15468ac9ac053b7cf6b007d301e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cafb3208e7e3bcdd01898a68fc5bbf4cb52ca0800b879b77a60bb102794257f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cafb3208e7e3bcdd01898a68fc5bbf4cb52ca0800b879b77a60bb102794257f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tskt7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:28Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:28 crc kubenswrapper[4754]: I1005 20:55:28.512043 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be95a413-6a59-45b4-84b7-b43cae694a26\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10b211cf6e47dbd1ee42c82e49de0a13a45581b92a0c4f6bbba0326c7149e41d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2ggg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4bb8c5903e8b70dd19eebb111097d96f725e17de44e06c3e74c067376f6c7360\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2ggg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-b2h9k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:28Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:28 crc kubenswrapper[4754]: I1005 20:55:28.526685 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rddkl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"15afc2a1-5cb7-41c6-954b-446c36822f50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://866d85ee126f8641d12f1a05fabc3de7b9392c324e9fe06ea92214155c05334c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqv8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a175565f83cfcba7825006eafb0f83efb70ec3d415c2c039c6499d5f6576e89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqv8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:55:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rddkl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:28Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:28 crc kubenswrapper[4754]: I1005 20:55:28.546470 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:28 crc kubenswrapper[4754]: I1005 20:55:28.546566 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:28 crc kubenswrapper[4754]: I1005 20:55:28.546588 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:28 crc kubenswrapper[4754]: I1005 20:55:28.546621 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:28 crc kubenswrapper[4754]: I1005 20:55:28.546641 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:28Z","lastTransitionTime":"2025-10-05T20:55:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:28 crc kubenswrapper[4754]: I1005 20:55:28.549960 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7506d496-65ff-41dc-8230-24c8ee653a2a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33da5a6bb3cd3b042c193bc571e6622f2f87f27e560cdd0bbb9be31e2010d707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df6d700a947d19c84b4f65fd2dd2f8d1f4f1f80fa9b4800c2acfa8dae3912639\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35ca8ee77a0e58cf969bd67b277a3392614c4e73b963e6f6002097709370715d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e32c2154f46ba10c509fe40c357304aa8e82d04e9906bbaf2ccecd47b42279e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8907b9b69b249c34ec6db1e460d641b31711b75d6d31d4caafd867339e6571d0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1005 20:54:52.589115 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 20:54:52.590181 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1304624737/tls.crt::/tmp/serving-cert-1304624737/tls.key\\\\\\\"\\\\nI1005 20:54:58.070544 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1005 20:54:58.090352 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1005 20:54:58.090382 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1005 20:54:58.090412 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1005 20:54:58.090418 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1005 20:54:58.144656 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1005 20:54:58.144688 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 20:54:58.144695 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 20:54:58.144699 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1005 20:54:58.144702 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1005 20:54:58.144705 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1005 20:54:58.144708 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1005 20:54:58.145024 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1005 20:54:58.148112 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4efb34f3493cc8854bc4cff6b29f63973c52b754404d7a21daec4e1f612d327d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54c568b637f3737667433047b956c7eb22fdf0f9301acdfd3b49fb84c39de6ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54c568b637f3737667433047b956c7eb22fdf0f9301acdfd3b49fb84c39de6ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:28Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:28 crc kubenswrapper[4754]: I1005 20:55:28.573338 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e380f9b2a0e319dc325c54d1247a6db98346292548fa688cb41a2ca1b3048419\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:28Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:28 crc kubenswrapper[4754]: I1005 20:55:28.588906 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:28Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:28 crc kubenswrapper[4754]: I1005 20:55:28.611637 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:28Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:28 crc kubenswrapper[4754]: I1005 20:55:28.641462 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b44790fb-fecb-4ec3-9816-8e23dd72ab33\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a7dc53e7f96e48213d724a2cb4859514df925005644108ae3a0c33b14dcf52d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfdd8b1e542ab676576c311e2923fb02861f92de91fe1f4b3e4b968c549c375d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3a63451ffaa14cacc3b78d11340f4ea81445fa3af96178e9b172d063d9697d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3cfd3f2274a8a46fa36660e94ac6c9c298c791283f12d52330c5594ac131998\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f626cb108716da0b595844ea0251b7a04d49b5908a1d6dd0aef5cb2d402c4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://822223f515b5add6d71d193fd5bc77df029b8d37a1cc07be4145b53c7226bf61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af1c313e0dc0fe3b48feac3cce1ecd4452c9efadf3d19f257c07cc7fcdec875f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af1c313e0dc0fe3b48feac3cce1ecd4452c9efadf3d19f257c07cc7fcdec875f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-05T20:55:26Z\\\",\\\"message\\\":\\\"5:26.770640 6339 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/iptables-alerter-4ln5h in node crc\\\\nI1005 20:55:26.770478 6339 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-apiserver-operator/metrics]} name:Service_openshift-apiserver-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.38:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {8b82f026-5975-4a1b-bb18-08d5d51147ec}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1005 20:55:26.770652 6339 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/iptables-alerter-4ln5h after 0 failed attempt(s)\\\\nI1005 20:55:26.770662 6339 default_network_controller.go:776] Recording success event on pod openshift-network-operator/iptables-alerter-4ln5h\\\\nF1005 20:55:26.770677 6339 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controlle\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:25Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-7rhps_openshift-ovn-kubernetes(b44790fb-fecb-4ec3-9816-8e23dd72ab33)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2acf9637fa25e5d3ffeb05224c11218c47defc679568fb928a498869613de4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35ebeb7a87d5d47b6567917d38eca38a3e622c497fba58971296c4609c287423\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35ebeb7a87d5d47b6567917d38eca38a3e622c497fba58971296c4609c287423\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7rhps\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:28Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:28 crc kubenswrapper[4754]: I1005 20:55:28.649926 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:28 crc kubenswrapper[4754]: I1005 20:55:28.650008 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:28 crc kubenswrapper[4754]: I1005 20:55:28.650028 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:28 crc kubenswrapper[4754]: I1005 20:55:28.650060 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:28 crc kubenswrapper[4754]: I1005 20:55:28.650081 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:28Z","lastTransitionTime":"2025-10-05T20:55:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:28 crc kubenswrapper[4754]: I1005 20:55:28.676291 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9309179-b92a-4fa4-9fba-747b8f911936\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d969d3d50d6ad280fc6654df9441bab238bb538f277ed6d1273779813ffbaaed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://800a731ed8ef57714b2318af14054ef23f0368936630887a093fc77167b33d12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8cd9b08e83dc174413cc37891c14562816787028471a4201680c86dd620b7a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93b729f3f7f993db0d6bc3ba60c04930aada3cef8272ffcc3c6a8a242a9b1475\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dde1519111fc3af5d14463fdb58d4f5c62a7045e7de98c60981c7e4da37b0e7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ea160a5b37aea7f2970ae4dd2614b09d3dcf5dcbd1fc103b9997f8ecc3a804c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ea160a5b37aea7f2970ae4dd2614b09d3dcf5dcbd1fc103b9997f8ecc3a804c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://17ae6aa6c53e7dd061f70b41428ef50bbc7ffcc65352fad1a4a9d3c69dbd2ba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17ae6aa6c53e7dd061f70b41428ef50bbc7ffcc65352fad1a4a9d3c69dbd2ba9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://ad4cd69df7608c7d87d839fe09d4557c674c7efbc4730f4e4d3daaf9a26f5aee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad4cd69df7608c7d87d839fe09d4557c674c7efbc4730f4e4d3daaf9a26f5aee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:36Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:28Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:28 crc kubenswrapper[4754]: I1005 20:55:28.754027 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:28 crc kubenswrapper[4754]: I1005 20:55:28.754086 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:28 crc kubenswrapper[4754]: I1005 20:55:28.754100 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:28 crc kubenswrapper[4754]: I1005 20:55:28.754130 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:28 crc kubenswrapper[4754]: I1005 20:55:28.754150 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:28Z","lastTransitionTime":"2025-10-05T20:55:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:28 crc kubenswrapper[4754]: I1005 20:55:28.836900 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 20:55:28 crc kubenswrapper[4754]: I1005 20:55:28.836911 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 20:55:28 crc kubenswrapper[4754]: E1005 20:55:28.837741 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 20:55:28 crc kubenswrapper[4754]: I1005 20:55:28.836902 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 20:55:28 crc kubenswrapper[4754]: E1005 20:55:28.838067 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 20:55:28 crc kubenswrapper[4754]: E1005 20:55:28.838167 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 20:55:28 crc kubenswrapper[4754]: I1005 20:55:28.857050 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:28 crc kubenswrapper[4754]: I1005 20:55:28.857111 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:28 crc kubenswrapper[4754]: I1005 20:55:28.857120 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:28 crc kubenswrapper[4754]: I1005 20:55:28.857153 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:28 crc kubenswrapper[4754]: I1005 20:55:28.857168 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:28Z","lastTransitionTime":"2025-10-05T20:55:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:28 crc kubenswrapper[4754]: I1005 20:55:28.961001 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:28 crc kubenswrapper[4754]: I1005 20:55:28.961280 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:28 crc kubenswrapper[4754]: I1005 20:55:28.961423 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:28 crc kubenswrapper[4754]: I1005 20:55:28.961603 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:28 crc kubenswrapper[4754]: I1005 20:55:28.961730 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:28Z","lastTransitionTime":"2025-10-05T20:55:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:29 crc kubenswrapper[4754]: I1005 20:55:29.065203 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:29 crc kubenswrapper[4754]: I1005 20:55:29.065289 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:29 crc kubenswrapper[4754]: I1005 20:55:29.065311 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:29 crc kubenswrapper[4754]: I1005 20:55:29.065343 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:29 crc kubenswrapper[4754]: I1005 20:55:29.065363 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:29Z","lastTransitionTime":"2025-10-05T20:55:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:29 crc kubenswrapper[4754]: I1005 20:55:29.168899 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:29 crc kubenswrapper[4754]: I1005 20:55:29.168966 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:29 crc kubenswrapper[4754]: I1005 20:55:29.168984 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:29 crc kubenswrapper[4754]: I1005 20:55:29.169018 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:29 crc kubenswrapper[4754]: I1005 20:55:29.169039 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:29Z","lastTransitionTime":"2025-10-05T20:55:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:29 crc kubenswrapper[4754]: I1005 20:55:29.271821 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:29 crc kubenswrapper[4754]: I1005 20:55:29.271862 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:29 crc kubenswrapper[4754]: I1005 20:55:29.271873 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:29 crc kubenswrapper[4754]: I1005 20:55:29.271895 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:29 crc kubenswrapper[4754]: I1005 20:55:29.271906 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:29Z","lastTransitionTime":"2025-10-05T20:55:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:29 crc kubenswrapper[4754]: I1005 20:55:29.375804 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:29 crc kubenswrapper[4754]: I1005 20:55:29.375863 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:29 crc kubenswrapper[4754]: I1005 20:55:29.375878 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:29 crc kubenswrapper[4754]: I1005 20:55:29.375903 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:29 crc kubenswrapper[4754]: I1005 20:55:29.375920 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:29Z","lastTransitionTime":"2025-10-05T20:55:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:29 crc kubenswrapper[4754]: I1005 20:55:29.479154 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:29 crc kubenswrapper[4754]: I1005 20:55:29.479231 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:29 crc kubenswrapper[4754]: I1005 20:55:29.479254 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:29 crc kubenswrapper[4754]: I1005 20:55:29.479286 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:29 crc kubenswrapper[4754]: I1005 20:55:29.479309 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:29Z","lastTransitionTime":"2025-10-05T20:55:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:29 crc kubenswrapper[4754]: I1005 20:55:29.582381 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:29 crc kubenswrapper[4754]: I1005 20:55:29.582452 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:29 crc kubenswrapper[4754]: I1005 20:55:29.582465 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:29 crc kubenswrapper[4754]: I1005 20:55:29.582486 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:29 crc kubenswrapper[4754]: I1005 20:55:29.582519 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:29Z","lastTransitionTime":"2025-10-05T20:55:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:29 crc kubenswrapper[4754]: I1005 20:55:29.587155 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7dc728ea-2601-44f0-bcab-2913f034007d-metrics-certs\") pod \"network-metrics-daemon-nwrnt\" (UID: \"7dc728ea-2601-44f0-bcab-2913f034007d\") " pod="openshift-multus/network-metrics-daemon-nwrnt" Oct 05 20:55:29 crc kubenswrapper[4754]: E1005 20:55:29.587448 4754 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 05 20:55:29 crc kubenswrapper[4754]: E1005 20:55:29.587582 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7dc728ea-2601-44f0-bcab-2913f034007d-metrics-certs podName:7dc728ea-2601-44f0-bcab-2913f034007d nodeName:}" failed. No retries permitted until 2025-10-05 20:55:45.587545818 +0000 UTC m=+69.491664568 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/7dc728ea-2601-44f0-bcab-2913f034007d-metrics-certs") pod "network-metrics-daemon-nwrnt" (UID: "7dc728ea-2601-44f0-bcab-2913f034007d") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 05 20:55:29 crc kubenswrapper[4754]: I1005 20:55:29.686127 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:29 crc kubenswrapper[4754]: I1005 20:55:29.686181 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:29 crc kubenswrapper[4754]: I1005 20:55:29.686202 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:29 crc kubenswrapper[4754]: I1005 20:55:29.686233 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:29 crc kubenswrapper[4754]: I1005 20:55:29.686251 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:29Z","lastTransitionTime":"2025-10-05T20:55:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:29 crc kubenswrapper[4754]: I1005 20:55:29.791216 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:29 crc kubenswrapper[4754]: I1005 20:55:29.791304 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:29 crc kubenswrapper[4754]: I1005 20:55:29.791326 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:29 crc kubenswrapper[4754]: I1005 20:55:29.791356 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:29 crc kubenswrapper[4754]: I1005 20:55:29.791378 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:29Z","lastTransitionTime":"2025-10-05T20:55:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:29 crc kubenswrapper[4754]: I1005 20:55:29.836638 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nwrnt" Oct 05 20:55:29 crc kubenswrapper[4754]: E1005 20:55:29.836912 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nwrnt" podUID="7dc728ea-2601-44f0-bcab-2913f034007d" Oct 05 20:55:29 crc kubenswrapper[4754]: I1005 20:55:29.894914 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:29 crc kubenswrapper[4754]: I1005 20:55:29.894989 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:29 crc kubenswrapper[4754]: I1005 20:55:29.895007 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:29 crc kubenswrapper[4754]: I1005 20:55:29.895034 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:29 crc kubenswrapper[4754]: I1005 20:55:29.895054 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:29Z","lastTransitionTime":"2025-10-05T20:55:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:29 crc kubenswrapper[4754]: I1005 20:55:29.998301 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:29 crc kubenswrapper[4754]: I1005 20:55:29.998381 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:29 crc kubenswrapper[4754]: I1005 20:55:29.998401 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:29 crc kubenswrapper[4754]: I1005 20:55:29.998437 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:29 crc kubenswrapper[4754]: I1005 20:55:29.998460 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:29Z","lastTransitionTime":"2025-10-05T20:55:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:30 crc kubenswrapper[4754]: I1005 20:55:30.103048 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:30 crc kubenswrapper[4754]: I1005 20:55:30.103124 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:30 crc kubenswrapper[4754]: I1005 20:55:30.103143 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:30 crc kubenswrapper[4754]: I1005 20:55:30.103172 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:30 crc kubenswrapper[4754]: I1005 20:55:30.103190 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:30Z","lastTransitionTime":"2025-10-05T20:55:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:30 crc kubenswrapper[4754]: I1005 20:55:30.205976 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:30 crc kubenswrapper[4754]: I1005 20:55:30.206046 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:30 crc kubenswrapper[4754]: I1005 20:55:30.206065 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:30 crc kubenswrapper[4754]: I1005 20:55:30.206094 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:30 crc kubenswrapper[4754]: I1005 20:55:30.206112 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:30Z","lastTransitionTime":"2025-10-05T20:55:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:30 crc kubenswrapper[4754]: I1005 20:55:30.311370 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:30 crc kubenswrapper[4754]: I1005 20:55:30.311448 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:30 crc kubenswrapper[4754]: I1005 20:55:30.311465 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:30 crc kubenswrapper[4754]: I1005 20:55:30.311514 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:30 crc kubenswrapper[4754]: I1005 20:55:30.311532 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:30Z","lastTransitionTime":"2025-10-05T20:55:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:30 crc kubenswrapper[4754]: I1005 20:55:30.414535 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:30 crc kubenswrapper[4754]: I1005 20:55:30.414885 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:30 crc kubenswrapper[4754]: I1005 20:55:30.415021 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:30 crc kubenswrapper[4754]: I1005 20:55:30.415206 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:30 crc kubenswrapper[4754]: I1005 20:55:30.415346 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:30Z","lastTransitionTime":"2025-10-05T20:55:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:30 crc kubenswrapper[4754]: I1005 20:55:30.498423 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 20:55:30 crc kubenswrapper[4754]: E1005 20:55:30.498692 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 20:56:02.498642844 +0000 UTC m=+86.402761584 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:55:30 crc kubenswrapper[4754]: I1005 20:55:30.518859 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:30 crc kubenswrapper[4754]: I1005 20:55:30.518925 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:30 crc kubenswrapper[4754]: I1005 20:55:30.518952 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:30 crc kubenswrapper[4754]: I1005 20:55:30.518984 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:30 crc kubenswrapper[4754]: I1005 20:55:30.519010 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:30Z","lastTransitionTime":"2025-10-05T20:55:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:30 crc kubenswrapper[4754]: I1005 20:55:30.600059 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 20:55:30 crc kubenswrapper[4754]: I1005 20:55:30.600132 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 20:55:30 crc kubenswrapper[4754]: I1005 20:55:30.600172 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 20:55:30 crc kubenswrapper[4754]: I1005 20:55:30.600227 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 20:55:30 crc kubenswrapper[4754]: E1005 20:55:30.600352 4754 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 05 20:55:30 crc kubenswrapper[4754]: E1005 20:55:30.600524 4754 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 05 20:55:30 crc kubenswrapper[4754]: E1005 20:55:30.600576 4754 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 05 20:55:30 crc kubenswrapper[4754]: E1005 20:55:30.600583 4754 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 05 20:55:30 crc kubenswrapper[4754]: E1005 20:55:30.600696 4754 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 05 20:55:30 crc kubenswrapper[4754]: E1005 20:55:30.600552 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-05 20:56:02.600476005 +0000 UTC m=+86.504594745 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 05 20:55:30 crc kubenswrapper[4754]: E1005 20:55:30.600529 4754 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 05 20:55:30 crc kubenswrapper[4754]: E1005 20:55:30.600908 4754 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 05 20:55:30 crc kubenswrapper[4754]: E1005 20:55:30.600949 4754 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 05 20:55:30 crc kubenswrapper[4754]: E1005 20:55:30.600822 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-05 20:56:02.600793613 +0000 UTC m=+86.504912363 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 05 20:55:30 crc kubenswrapper[4754]: E1005 20:55:30.601029 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-05 20:56:02.600993858 +0000 UTC m=+86.505112758 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 05 20:55:30 crc kubenswrapper[4754]: E1005 20:55:30.601068 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-05 20:56:02.601048579 +0000 UTC m=+86.505167559 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 05 20:55:30 crc kubenswrapper[4754]: I1005 20:55:30.622419 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:30 crc kubenswrapper[4754]: I1005 20:55:30.622486 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:30 crc kubenswrapper[4754]: I1005 20:55:30.622564 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:30 crc kubenswrapper[4754]: I1005 20:55:30.622596 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:30 crc kubenswrapper[4754]: I1005 20:55:30.622622 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:30Z","lastTransitionTime":"2025-10-05T20:55:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:30 crc kubenswrapper[4754]: I1005 20:55:30.726765 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:30 crc kubenswrapper[4754]: I1005 20:55:30.727287 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:30 crc kubenswrapper[4754]: I1005 20:55:30.727424 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:30 crc kubenswrapper[4754]: I1005 20:55:30.727647 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:30 crc kubenswrapper[4754]: I1005 20:55:30.727807 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:30Z","lastTransitionTime":"2025-10-05T20:55:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:30 crc kubenswrapper[4754]: I1005 20:55:30.832576 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:30 crc kubenswrapper[4754]: I1005 20:55:30.832629 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:30 crc kubenswrapper[4754]: I1005 20:55:30.832648 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:30 crc kubenswrapper[4754]: I1005 20:55:30.832674 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:30 crc kubenswrapper[4754]: I1005 20:55:30.832692 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:30Z","lastTransitionTime":"2025-10-05T20:55:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:30 crc kubenswrapper[4754]: I1005 20:55:30.837285 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 20:55:30 crc kubenswrapper[4754]: I1005 20:55:30.837379 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 20:55:30 crc kubenswrapper[4754]: E1005 20:55:30.837610 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 20:55:30 crc kubenswrapper[4754]: I1005 20:55:30.837833 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 20:55:30 crc kubenswrapper[4754]: E1005 20:55:30.837846 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 20:55:30 crc kubenswrapper[4754]: E1005 20:55:30.838174 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 20:55:30 crc kubenswrapper[4754]: I1005 20:55:30.936513 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:30 crc kubenswrapper[4754]: I1005 20:55:30.936585 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:30 crc kubenswrapper[4754]: I1005 20:55:30.936603 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:30 crc kubenswrapper[4754]: I1005 20:55:30.936635 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:30 crc kubenswrapper[4754]: I1005 20:55:30.936655 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:30Z","lastTransitionTime":"2025-10-05T20:55:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:31 crc kubenswrapper[4754]: I1005 20:55:31.039616 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:31 crc kubenswrapper[4754]: I1005 20:55:31.039691 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:31 crc kubenswrapper[4754]: I1005 20:55:31.039709 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:31 crc kubenswrapper[4754]: I1005 20:55:31.039738 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:31 crc kubenswrapper[4754]: I1005 20:55:31.039756 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:31Z","lastTransitionTime":"2025-10-05T20:55:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:31 crc kubenswrapper[4754]: I1005 20:55:31.143897 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:31 crc kubenswrapper[4754]: I1005 20:55:31.143965 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:31 crc kubenswrapper[4754]: I1005 20:55:31.143986 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:31 crc kubenswrapper[4754]: I1005 20:55:31.144016 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:31 crc kubenswrapper[4754]: I1005 20:55:31.144035 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:31Z","lastTransitionTime":"2025-10-05T20:55:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:31 crc kubenswrapper[4754]: I1005 20:55:31.247780 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:31 crc kubenswrapper[4754]: I1005 20:55:31.247867 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:31 crc kubenswrapper[4754]: I1005 20:55:31.247896 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:31 crc kubenswrapper[4754]: I1005 20:55:31.247932 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:31 crc kubenswrapper[4754]: I1005 20:55:31.247954 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:31Z","lastTransitionTime":"2025-10-05T20:55:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:31 crc kubenswrapper[4754]: I1005 20:55:31.351342 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:31 crc kubenswrapper[4754]: I1005 20:55:31.351416 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:31 crc kubenswrapper[4754]: I1005 20:55:31.351443 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:31 crc kubenswrapper[4754]: I1005 20:55:31.351476 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:31 crc kubenswrapper[4754]: I1005 20:55:31.351541 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:31Z","lastTransitionTime":"2025-10-05T20:55:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:31 crc kubenswrapper[4754]: I1005 20:55:31.454643 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:31 crc kubenswrapper[4754]: I1005 20:55:31.454714 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:31 crc kubenswrapper[4754]: I1005 20:55:31.454735 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:31 crc kubenswrapper[4754]: I1005 20:55:31.454762 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:31 crc kubenswrapper[4754]: I1005 20:55:31.454781 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:31Z","lastTransitionTime":"2025-10-05T20:55:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:31 crc kubenswrapper[4754]: I1005 20:55:31.558383 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:31 crc kubenswrapper[4754]: I1005 20:55:31.558450 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:31 crc kubenswrapper[4754]: I1005 20:55:31.558468 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:31 crc kubenswrapper[4754]: I1005 20:55:31.558527 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:31 crc kubenswrapper[4754]: I1005 20:55:31.558548 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:31Z","lastTransitionTime":"2025-10-05T20:55:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:31 crc kubenswrapper[4754]: I1005 20:55:31.661378 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:31 crc kubenswrapper[4754]: I1005 20:55:31.661452 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:31 crc kubenswrapper[4754]: I1005 20:55:31.661472 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:31 crc kubenswrapper[4754]: I1005 20:55:31.661536 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:31 crc kubenswrapper[4754]: I1005 20:55:31.661559 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:31Z","lastTransitionTime":"2025-10-05T20:55:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:31 crc kubenswrapper[4754]: I1005 20:55:31.764916 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:31 crc kubenswrapper[4754]: I1005 20:55:31.765346 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:31 crc kubenswrapper[4754]: I1005 20:55:31.765535 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:31 crc kubenswrapper[4754]: I1005 20:55:31.765690 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:31 crc kubenswrapper[4754]: I1005 20:55:31.765845 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:31Z","lastTransitionTime":"2025-10-05T20:55:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:31 crc kubenswrapper[4754]: I1005 20:55:31.837197 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nwrnt" Oct 05 20:55:31 crc kubenswrapper[4754]: E1005 20:55:31.837605 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nwrnt" podUID="7dc728ea-2601-44f0-bcab-2913f034007d" Oct 05 20:55:31 crc kubenswrapper[4754]: I1005 20:55:31.870331 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:31 crc kubenswrapper[4754]: I1005 20:55:31.870533 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:31 crc kubenswrapper[4754]: I1005 20:55:31.870559 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:31 crc kubenswrapper[4754]: I1005 20:55:31.870589 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:31 crc kubenswrapper[4754]: I1005 20:55:31.870608 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:31Z","lastTransitionTime":"2025-10-05T20:55:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:31 crc kubenswrapper[4754]: I1005 20:55:31.974634 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:31 crc kubenswrapper[4754]: I1005 20:55:31.974706 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:31 crc kubenswrapper[4754]: I1005 20:55:31.974729 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:31 crc kubenswrapper[4754]: I1005 20:55:31.974758 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:31 crc kubenswrapper[4754]: I1005 20:55:31.974776 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:31Z","lastTransitionTime":"2025-10-05T20:55:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:32 crc kubenswrapper[4754]: I1005 20:55:32.085226 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:32 crc kubenswrapper[4754]: I1005 20:55:32.085311 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:32 crc kubenswrapper[4754]: I1005 20:55:32.085334 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:32 crc kubenswrapper[4754]: I1005 20:55:32.085364 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:32 crc kubenswrapper[4754]: I1005 20:55:32.085384 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:32Z","lastTransitionTime":"2025-10-05T20:55:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:32 crc kubenswrapper[4754]: I1005 20:55:32.189322 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:32 crc kubenswrapper[4754]: I1005 20:55:32.189396 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:32 crc kubenswrapper[4754]: I1005 20:55:32.189415 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:32 crc kubenswrapper[4754]: I1005 20:55:32.189442 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:32 crc kubenswrapper[4754]: I1005 20:55:32.189463 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:32Z","lastTransitionTime":"2025-10-05T20:55:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:32 crc kubenswrapper[4754]: I1005 20:55:32.293148 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:32 crc kubenswrapper[4754]: I1005 20:55:32.293233 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:32 crc kubenswrapper[4754]: I1005 20:55:32.293257 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:32 crc kubenswrapper[4754]: I1005 20:55:32.293291 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:32 crc kubenswrapper[4754]: I1005 20:55:32.293311 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:32Z","lastTransitionTime":"2025-10-05T20:55:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:32 crc kubenswrapper[4754]: I1005 20:55:32.396418 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:32 crc kubenswrapper[4754]: I1005 20:55:32.396486 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:32 crc kubenswrapper[4754]: I1005 20:55:32.396538 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:32 crc kubenswrapper[4754]: I1005 20:55:32.396568 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:32 crc kubenswrapper[4754]: I1005 20:55:32.396589 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:32Z","lastTransitionTime":"2025-10-05T20:55:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:32 crc kubenswrapper[4754]: I1005 20:55:32.500381 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:32 crc kubenswrapper[4754]: I1005 20:55:32.500451 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:32 crc kubenswrapper[4754]: I1005 20:55:32.500471 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:32 crc kubenswrapper[4754]: I1005 20:55:32.500525 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:32 crc kubenswrapper[4754]: I1005 20:55:32.500545 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:32Z","lastTransitionTime":"2025-10-05T20:55:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:32 crc kubenswrapper[4754]: I1005 20:55:32.604374 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:32 crc kubenswrapper[4754]: I1005 20:55:32.604439 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:32 crc kubenswrapper[4754]: I1005 20:55:32.604459 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:32 crc kubenswrapper[4754]: I1005 20:55:32.604523 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:32 crc kubenswrapper[4754]: I1005 20:55:32.604543 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:32Z","lastTransitionTime":"2025-10-05T20:55:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:32 crc kubenswrapper[4754]: I1005 20:55:32.708154 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:32 crc kubenswrapper[4754]: I1005 20:55:32.708220 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:32 crc kubenswrapper[4754]: I1005 20:55:32.708240 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:32 crc kubenswrapper[4754]: I1005 20:55:32.708267 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:32 crc kubenswrapper[4754]: I1005 20:55:32.708287 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:32Z","lastTransitionTime":"2025-10-05T20:55:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:32 crc kubenswrapper[4754]: I1005 20:55:32.812709 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:32 crc kubenswrapper[4754]: I1005 20:55:32.812791 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:32 crc kubenswrapper[4754]: I1005 20:55:32.812816 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:32 crc kubenswrapper[4754]: I1005 20:55:32.812847 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:32 crc kubenswrapper[4754]: I1005 20:55:32.812868 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:32Z","lastTransitionTime":"2025-10-05T20:55:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:32 crc kubenswrapper[4754]: I1005 20:55:32.837090 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 20:55:32 crc kubenswrapper[4754]: I1005 20:55:32.837131 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 20:55:32 crc kubenswrapper[4754]: I1005 20:55:32.837106 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 20:55:32 crc kubenswrapper[4754]: E1005 20:55:32.837345 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 20:55:32 crc kubenswrapper[4754]: E1005 20:55:32.837621 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 20:55:32 crc kubenswrapper[4754]: E1005 20:55:32.837885 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 20:55:32 crc kubenswrapper[4754]: I1005 20:55:32.917962 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:32 crc kubenswrapper[4754]: I1005 20:55:32.918022 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:32 crc kubenswrapper[4754]: I1005 20:55:32.918042 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:32 crc kubenswrapper[4754]: I1005 20:55:32.918067 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:32 crc kubenswrapper[4754]: I1005 20:55:32.918090 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:32Z","lastTransitionTime":"2025-10-05T20:55:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:33 crc kubenswrapper[4754]: I1005 20:55:33.021706 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:33 crc kubenswrapper[4754]: I1005 20:55:33.021765 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:33 crc kubenswrapper[4754]: I1005 20:55:33.021781 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:33 crc kubenswrapper[4754]: I1005 20:55:33.021812 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:33 crc kubenswrapper[4754]: I1005 20:55:33.021831 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:33Z","lastTransitionTime":"2025-10-05T20:55:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:33 crc kubenswrapper[4754]: I1005 20:55:33.125414 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:33 crc kubenswrapper[4754]: I1005 20:55:33.125536 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:33 crc kubenswrapper[4754]: I1005 20:55:33.125558 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:33 crc kubenswrapper[4754]: I1005 20:55:33.125588 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:33 crc kubenswrapper[4754]: I1005 20:55:33.125609 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:33Z","lastTransitionTime":"2025-10-05T20:55:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:33 crc kubenswrapper[4754]: I1005 20:55:33.229423 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:33 crc kubenswrapper[4754]: I1005 20:55:33.229547 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:33 crc kubenswrapper[4754]: I1005 20:55:33.229568 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:33 crc kubenswrapper[4754]: I1005 20:55:33.229598 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:33 crc kubenswrapper[4754]: I1005 20:55:33.229617 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:33Z","lastTransitionTime":"2025-10-05T20:55:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:33 crc kubenswrapper[4754]: I1005 20:55:33.332975 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:33 crc kubenswrapper[4754]: I1005 20:55:33.333082 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:33 crc kubenswrapper[4754]: I1005 20:55:33.333101 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:33 crc kubenswrapper[4754]: I1005 20:55:33.333133 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:33 crc kubenswrapper[4754]: I1005 20:55:33.333157 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:33Z","lastTransitionTime":"2025-10-05T20:55:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:33 crc kubenswrapper[4754]: I1005 20:55:33.437152 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:33 crc kubenswrapper[4754]: I1005 20:55:33.437228 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:33 crc kubenswrapper[4754]: I1005 20:55:33.437248 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:33 crc kubenswrapper[4754]: I1005 20:55:33.437273 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:33 crc kubenswrapper[4754]: I1005 20:55:33.437289 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:33Z","lastTransitionTime":"2025-10-05T20:55:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:33 crc kubenswrapper[4754]: I1005 20:55:33.523010 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 05 20:55:33 crc kubenswrapper[4754]: I1005 20:55:33.542773 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:33 crc kubenswrapper[4754]: I1005 20:55:33.542898 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Oct 05 20:55:33 crc kubenswrapper[4754]: I1005 20:55:33.542948 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:33 crc kubenswrapper[4754]: I1005 20:55:33.542982 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:33 crc kubenswrapper[4754]: I1005 20:55:33.543077 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:33 crc kubenswrapper[4754]: I1005 20:55:33.543165 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:33Z","lastTransitionTime":"2025-10-05T20:55:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:33 crc kubenswrapper[4754]: I1005 20:55:33.548258 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"934bcd66-9c03-48b9-82dd-5093338f9c8f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9042df2f43243d3d79ba9a3940f8dd4f6c7a89ef9e3510e81fe60b1944d7e83b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5366bbf46a91667355f43f5df7335a536b395cda788b90b35ef18e2b53bc4c02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://600bc3bb551b747cf2f18e26745883232d7a37ac5877fad0c67100e52d7a43aa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78f74982180048eb50b9be79fa72c04f62f925b54fe511920f37e6407f864631\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:33Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:33 crc kubenswrapper[4754]: I1005 20:55:33.572635 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:33Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:33 crc kubenswrapper[4754]: I1005 20:55:33.594722 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9ad070bb699b643d29b82bcc1126ce3b585f7492f5ae2817fa491b309b1691d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:33Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:33 crc kubenswrapper[4754]: I1005 20:55:33.614019 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-57sr4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e854c7b9-1229-4da4-80e9-fea6479459b8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceb577e1055d84007936a50d0b3d54b9ae0299a1fc1a6618b37163a6d3fbb9cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vw2ps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:58Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-57sr4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:33Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:33 crc kubenswrapper[4754]: I1005 20:55:33.640258 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cn76s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02d5e3f9-73c2-4496-9aca-0787184aef19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7ba2b076839f4a5f586d2e6c4f4d7188254128b782bfd2180e22bf3ea5ffdd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpx5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cn76s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:33Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:33 crc kubenswrapper[4754]: I1005 20:55:33.646818 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:33 crc kubenswrapper[4754]: I1005 20:55:33.646881 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:33 crc kubenswrapper[4754]: I1005 20:55:33.646899 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:33 crc kubenswrapper[4754]: I1005 20:55:33.646931 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:33 crc kubenswrapper[4754]: I1005 20:55:33.646949 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:33Z","lastTransitionTime":"2025-10-05T20:55:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:33 crc kubenswrapper[4754]: I1005 20:55:33.661725 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ss7lm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f9ef817-0622-4185-a7fb-57221690aa75\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ba613b0d7753a003c572c3733ed394707c45b6f8072cb7aa88f3c7312b8882b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wpwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:55:02Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ss7lm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:33Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:33 crc kubenswrapper[4754]: I1005 20:55:33.682103 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-nwrnt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7dc728ea-2601-44f0-bcab-2913f034007d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-846xq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-846xq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:55:13Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-nwrnt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:33Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:33 crc kubenswrapper[4754]: I1005 20:55:33.708641 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7506d496-65ff-41dc-8230-24c8ee653a2a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33da5a6bb3cd3b042c193bc571e6622f2f87f27e560cdd0bbb9be31e2010d707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df6d700a947d19c84b4f65fd2dd2f8d1f4f1f80fa9b4800c2acfa8dae3912639\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35ca8ee77a0e58cf969bd67b277a3392614c4e73b963e6f6002097709370715d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e32c2154f46ba10c509fe40c357304aa8e82d04e9906bbaf2ccecd47b42279e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8907b9b69b249c34ec6db1e460d641b31711b75d6d31d4caafd867339e6571d0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1005 20:54:52.589115 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 20:54:52.590181 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1304624737/tls.crt::/tmp/serving-cert-1304624737/tls.key\\\\\\\"\\\\nI1005 20:54:58.070544 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1005 20:54:58.090352 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1005 20:54:58.090382 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1005 20:54:58.090412 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1005 20:54:58.090418 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1005 20:54:58.144656 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1005 20:54:58.144688 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 20:54:58.144695 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 20:54:58.144699 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1005 20:54:58.144702 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1005 20:54:58.144705 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1005 20:54:58.144708 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1005 20:54:58.145024 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1005 20:54:58.148112 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4efb34f3493cc8854bc4cff6b29f63973c52b754404d7a21daec4e1f612d327d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54c568b637f3737667433047b956c7eb22fdf0f9301acdfd3b49fb84c39de6ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54c568b637f3737667433047b956c7eb22fdf0f9301acdfd3b49fb84c39de6ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:33Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:33 crc kubenswrapper[4754]: I1005 20:55:33.732197 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c76e6e4740035a6ca54191a770c4516e28101aae69afbb89739174c2f076ef5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9eededc0cb1af94fb7b5daba08785f6b077a85e9cb5ddd8b9f397b658194f61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:33Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:33 crc kubenswrapper[4754]: I1005 20:55:33.750661 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:33 crc kubenswrapper[4754]: I1005 20:55:33.750731 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:33 crc kubenswrapper[4754]: I1005 20:55:33.750748 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:33 crc kubenswrapper[4754]: I1005 20:55:33.750776 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:33 crc kubenswrapper[4754]: I1005 20:55:33.750795 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:33Z","lastTransitionTime":"2025-10-05T20:55:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:33 crc kubenswrapper[4754]: I1005 20:55:33.756833 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tskt7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a68e9c89-465b-4c01-a470-e21a480ee465\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba6edba14cf321b4cd2f7bb44067a3ecf49f1594e76c0197e807ff3daa45e9b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d14ec99610d2167b484bd7712af7bb6d993d26c74b2828da1c89314f16fda22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d14ec99610d2167b484bd7712af7bb6d993d26c74b2828da1c89314f16fda22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f6f08dd853b25881e2f3c66206aa80d0c493518284217780aa40ac4b69f0c0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f6f08dd853b25881e2f3c66206aa80d0c493518284217780aa40ac4b69f0c0b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://71026d00ec1a3d91b019deb3c9f43ee52fb05b8f2d6374e1a5b762c9e5c1b10d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71026d00ec1a3d91b019deb3c9f43ee52fb05b8f2d6374e1a5b762c9e5c1b10d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b9910d2470acb16ebb436d7030b3161061607cb2a5ef50d482d8d5c4f48a8ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b9910d2470acb16ebb436d7030b3161061607cb2a5ef50d482d8d5c4f48a8ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0a93b61eacc55111995e69c15f89a210b8aa15468ac9ac053b7cf6b007d301e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b0a93b61eacc55111995e69c15f89a210b8aa15468ac9ac053b7cf6b007d301e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cafb3208e7e3bcdd01898a68fc5bbf4cb52ca0800b879b77a60bb102794257f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cafb3208e7e3bcdd01898a68fc5bbf4cb52ca0800b879b77a60bb102794257f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tskt7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:33Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:33 crc kubenswrapper[4754]: I1005 20:55:33.777215 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be95a413-6a59-45b4-84b7-b43cae694a26\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10b211cf6e47dbd1ee42c82e49de0a13a45581b92a0c4f6bbba0326c7149e41d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2ggg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4bb8c5903e8b70dd19eebb111097d96f725e17de44e06c3e74c067376f6c7360\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2ggg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-b2h9k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:33Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:33 crc kubenswrapper[4754]: I1005 20:55:33.801225 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rddkl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"15afc2a1-5cb7-41c6-954b-446c36822f50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://866d85ee126f8641d12f1a05fabc3de7b9392c324e9fe06ea92214155c05334c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqv8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a175565f83cfcba7825006eafb0f83efb70ec3d415c2c039c6499d5f6576e89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqv8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:55:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rddkl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:33Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:33 crc kubenswrapper[4754]: I1005 20:55:33.825229 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e380f9b2a0e319dc325c54d1247a6db98346292548fa688cb41a2ca1b3048419\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:33Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:33 crc kubenswrapper[4754]: I1005 20:55:33.837176 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nwrnt" Oct 05 20:55:33 crc kubenswrapper[4754]: E1005 20:55:33.837420 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nwrnt" podUID="7dc728ea-2601-44f0-bcab-2913f034007d" Oct 05 20:55:33 crc kubenswrapper[4754]: I1005 20:55:33.853749 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:33 crc kubenswrapper[4754]: I1005 20:55:33.853832 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:33 crc kubenswrapper[4754]: I1005 20:55:33.853857 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:33 crc kubenswrapper[4754]: I1005 20:55:33.853892 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:33 crc kubenswrapper[4754]: I1005 20:55:33.853914 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:33Z","lastTransitionTime":"2025-10-05T20:55:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:33 crc kubenswrapper[4754]: I1005 20:55:33.881227 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:33Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:33 crc kubenswrapper[4754]: I1005 20:55:33.937139 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9309179-b92a-4fa4-9fba-747b8f911936\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d969d3d50d6ad280fc6654df9441bab238bb538f277ed6d1273779813ffbaaed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://800a731ed8ef57714b2318af14054ef23f0368936630887a093fc77167b33d12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8cd9b08e83dc174413cc37891c14562816787028471a4201680c86dd620b7a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93b729f3f7f993db0d6bc3ba60c04930aada3cef8272ffcc3c6a8a242a9b1475\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dde1519111fc3af5d14463fdb58d4f5c62a7045e7de98c60981c7e4da37b0e7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ea160a5b37aea7f2970ae4dd2614b09d3dcf5dcbd1fc103b9997f8ecc3a804c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ea160a5b37aea7f2970ae4dd2614b09d3dcf5dcbd1fc103b9997f8ecc3a804c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://17ae6aa6c53e7dd061f70b41428ef50bbc7ffcc65352fad1a4a9d3c69dbd2ba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17ae6aa6c53e7dd061f70b41428ef50bbc7ffcc65352fad1a4a9d3c69dbd2ba9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://ad4cd69df7608c7d87d839fe09d4557c674c7efbc4730f4e4d3daaf9a26f5aee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad4cd69df7608c7d87d839fe09d4557c674c7efbc4730f4e4d3daaf9a26f5aee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:36Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:33Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:33 crc kubenswrapper[4754]: I1005 20:55:33.956315 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:33 crc kubenswrapper[4754]: I1005 20:55:33.956366 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:33 crc kubenswrapper[4754]: I1005 20:55:33.956376 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:33 crc kubenswrapper[4754]: I1005 20:55:33.956395 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:33 crc kubenswrapper[4754]: I1005 20:55:33.956408 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:33Z","lastTransitionTime":"2025-10-05T20:55:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:33 crc kubenswrapper[4754]: I1005 20:55:33.959331 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:33Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:33 crc kubenswrapper[4754]: I1005 20:55:33.977192 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b44790fb-fecb-4ec3-9816-8e23dd72ab33\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a7dc53e7f96e48213d724a2cb4859514df925005644108ae3a0c33b14dcf52d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfdd8b1e542ab676576c311e2923fb02861f92de91fe1f4b3e4b968c549c375d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3a63451ffaa14cacc3b78d11340f4ea81445fa3af96178e9b172d063d9697d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3cfd3f2274a8a46fa36660e94ac6c9c298c791283f12d52330c5594ac131998\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f626cb108716da0b595844ea0251b7a04d49b5908a1d6dd0aef5cb2d402c4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://822223f515b5add6d71d193fd5bc77df029b8d37a1cc07be4145b53c7226bf61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af1c313e0dc0fe3b48feac3cce1ecd4452c9efadf3d19f257c07cc7fcdec875f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af1c313e0dc0fe3b48feac3cce1ecd4452c9efadf3d19f257c07cc7fcdec875f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-05T20:55:26Z\\\",\\\"message\\\":\\\"5:26.770640 6339 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/iptables-alerter-4ln5h in node crc\\\\nI1005 20:55:26.770478 6339 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-apiserver-operator/metrics]} name:Service_openshift-apiserver-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.38:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {8b82f026-5975-4a1b-bb18-08d5d51147ec}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1005 20:55:26.770652 6339 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/iptables-alerter-4ln5h after 0 failed attempt(s)\\\\nI1005 20:55:26.770662 6339 default_network_controller.go:776] Recording success event on pod openshift-network-operator/iptables-alerter-4ln5h\\\\nF1005 20:55:26.770677 6339 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controlle\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:25Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-7rhps_openshift-ovn-kubernetes(b44790fb-fecb-4ec3-9816-8e23dd72ab33)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2acf9637fa25e5d3ffeb05224c11218c47defc679568fb928a498869613de4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35ebeb7a87d5d47b6567917d38eca38a3e622c497fba58971296c4609c287423\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35ebeb7a87d5d47b6567917d38eca38a3e622c497fba58971296c4609c287423\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7rhps\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:33Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:34 crc kubenswrapper[4754]: I1005 20:55:34.060126 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:34 crc kubenswrapper[4754]: I1005 20:55:34.060223 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:34 crc kubenswrapper[4754]: I1005 20:55:34.060243 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:34 crc kubenswrapper[4754]: I1005 20:55:34.060270 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:34 crc kubenswrapper[4754]: I1005 20:55:34.060291 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:34Z","lastTransitionTime":"2025-10-05T20:55:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:34 crc kubenswrapper[4754]: I1005 20:55:34.164229 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:34 crc kubenswrapper[4754]: I1005 20:55:34.164298 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:34 crc kubenswrapper[4754]: I1005 20:55:34.164316 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:34 crc kubenswrapper[4754]: I1005 20:55:34.164347 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:34 crc kubenswrapper[4754]: I1005 20:55:34.164369 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:34Z","lastTransitionTime":"2025-10-05T20:55:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:34 crc kubenswrapper[4754]: I1005 20:55:34.268164 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:34 crc kubenswrapper[4754]: I1005 20:55:34.268222 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:34 crc kubenswrapper[4754]: I1005 20:55:34.268239 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:34 crc kubenswrapper[4754]: I1005 20:55:34.268267 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:34 crc kubenswrapper[4754]: I1005 20:55:34.268287 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:34Z","lastTransitionTime":"2025-10-05T20:55:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:34 crc kubenswrapper[4754]: I1005 20:55:34.371341 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:34 crc kubenswrapper[4754]: I1005 20:55:34.371388 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:34 crc kubenswrapper[4754]: I1005 20:55:34.371400 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:34 crc kubenswrapper[4754]: I1005 20:55:34.371419 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:34 crc kubenswrapper[4754]: I1005 20:55:34.371433 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:34Z","lastTransitionTime":"2025-10-05T20:55:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:34 crc kubenswrapper[4754]: I1005 20:55:34.475050 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:34 crc kubenswrapper[4754]: I1005 20:55:34.475118 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:34 crc kubenswrapper[4754]: I1005 20:55:34.475138 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:34 crc kubenswrapper[4754]: I1005 20:55:34.475175 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:34 crc kubenswrapper[4754]: I1005 20:55:34.475198 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:34Z","lastTransitionTime":"2025-10-05T20:55:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:34 crc kubenswrapper[4754]: I1005 20:55:34.578451 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:34 crc kubenswrapper[4754]: I1005 20:55:34.578533 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:34 crc kubenswrapper[4754]: I1005 20:55:34.578552 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:34 crc kubenswrapper[4754]: I1005 20:55:34.578577 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:34 crc kubenswrapper[4754]: I1005 20:55:34.578596 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:34Z","lastTransitionTime":"2025-10-05T20:55:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:34 crc kubenswrapper[4754]: I1005 20:55:34.682863 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:34 crc kubenswrapper[4754]: I1005 20:55:34.683230 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:34 crc kubenswrapper[4754]: I1005 20:55:34.683354 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:34 crc kubenswrapper[4754]: I1005 20:55:34.683527 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:34 crc kubenswrapper[4754]: I1005 20:55:34.683669 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:34Z","lastTransitionTime":"2025-10-05T20:55:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:34 crc kubenswrapper[4754]: I1005 20:55:34.787746 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:34 crc kubenswrapper[4754]: I1005 20:55:34.787824 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:34 crc kubenswrapper[4754]: I1005 20:55:34.787844 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:34 crc kubenswrapper[4754]: I1005 20:55:34.787875 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:34 crc kubenswrapper[4754]: I1005 20:55:34.787898 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:34Z","lastTransitionTime":"2025-10-05T20:55:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:34 crc kubenswrapper[4754]: I1005 20:55:34.837127 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 20:55:34 crc kubenswrapper[4754]: I1005 20:55:34.837254 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 20:55:34 crc kubenswrapper[4754]: I1005 20:55:34.837649 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 20:55:34 crc kubenswrapper[4754]: E1005 20:55:34.837785 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 20:55:34 crc kubenswrapper[4754]: E1005 20:55:34.838112 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 20:55:34 crc kubenswrapper[4754]: E1005 20:55:34.838176 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 20:55:34 crc kubenswrapper[4754]: I1005 20:55:34.891156 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:34 crc kubenswrapper[4754]: I1005 20:55:34.891220 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:34 crc kubenswrapper[4754]: I1005 20:55:34.891239 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:34 crc kubenswrapper[4754]: I1005 20:55:34.891265 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:34 crc kubenswrapper[4754]: I1005 20:55:34.891289 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:34Z","lastTransitionTime":"2025-10-05T20:55:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:34 crc kubenswrapper[4754]: I1005 20:55:34.995288 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:34 crc kubenswrapper[4754]: I1005 20:55:34.995374 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:34 crc kubenswrapper[4754]: I1005 20:55:34.995401 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:34 crc kubenswrapper[4754]: I1005 20:55:34.995438 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:34 crc kubenswrapper[4754]: I1005 20:55:34.995465 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:34Z","lastTransitionTime":"2025-10-05T20:55:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:35 crc kubenswrapper[4754]: I1005 20:55:35.005377 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:35 crc kubenswrapper[4754]: I1005 20:55:35.005461 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:35 crc kubenswrapper[4754]: I1005 20:55:35.005481 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:35 crc kubenswrapper[4754]: I1005 20:55:35.005526 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:35 crc kubenswrapper[4754]: I1005 20:55:35.005544 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:35Z","lastTransitionTime":"2025-10-05T20:55:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:35 crc kubenswrapper[4754]: E1005 20:55:35.027264 4754 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T20:55:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T20:55:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T20:55:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T20:55:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ede2b442-76b0-443a-a450-252ee8711c80\\\",\\\"systemUUID\\\":\\\"72d081dd-fe7e-42a6-9cf0-f0ee7f733ac9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:35Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:35 crc kubenswrapper[4754]: I1005 20:55:35.033114 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:35 crc kubenswrapper[4754]: I1005 20:55:35.033337 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:35 crc kubenswrapper[4754]: I1005 20:55:35.033484 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:35 crc kubenswrapper[4754]: I1005 20:55:35.033663 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:35 crc kubenswrapper[4754]: I1005 20:55:35.033794 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:35Z","lastTransitionTime":"2025-10-05T20:55:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:35 crc kubenswrapper[4754]: E1005 20:55:35.053543 4754 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T20:55:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T20:55:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T20:55:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T20:55:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ede2b442-76b0-443a-a450-252ee8711c80\\\",\\\"systemUUID\\\":\\\"72d081dd-fe7e-42a6-9cf0-f0ee7f733ac9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:35Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:35 crc kubenswrapper[4754]: I1005 20:55:35.060004 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:35 crc kubenswrapper[4754]: I1005 20:55:35.060219 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:35 crc kubenswrapper[4754]: I1005 20:55:35.060405 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:35 crc kubenswrapper[4754]: I1005 20:55:35.061112 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:35 crc kubenswrapper[4754]: I1005 20:55:35.061261 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:35Z","lastTransitionTime":"2025-10-05T20:55:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:35 crc kubenswrapper[4754]: E1005 20:55:35.082821 4754 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T20:55:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T20:55:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T20:55:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T20:55:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ede2b442-76b0-443a-a450-252ee8711c80\\\",\\\"systemUUID\\\":\\\"72d081dd-fe7e-42a6-9cf0-f0ee7f733ac9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:35Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:35 crc kubenswrapper[4754]: I1005 20:55:35.088543 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:35 crc kubenswrapper[4754]: I1005 20:55:35.088734 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:35 crc kubenswrapper[4754]: I1005 20:55:35.088872 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:35 crc kubenswrapper[4754]: I1005 20:55:35.089000 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:35 crc kubenswrapper[4754]: I1005 20:55:35.089129 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:35Z","lastTransitionTime":"2025-10-05T20:55:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:35 crc kubenswrapper[4754]: E1005 20:55:35.110005 4754 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T20:55:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T20:55:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T20:55:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T20:55:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ede2b442-76b0-443a-a450-252ee8711c80\\\",\\\"systemUUID\\\":\\\"72d081dd-fe7e-42a6-9cf0-f0ee7f733ac9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:35Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:35 crc kubenswrapper[4754]: I1005 20:55:35.115308 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:35 crc kubenswrapper[4754]: I1005 20:55:35.115547 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:35 crc kubenswrapper[4754]: I1005 20:55:35.115696 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:35 crc kubenswrapper[4754]: I1005 20:55:35.115831 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:35 crc kubenswrapper[4754]: I1005 20:55:35.115947 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:35Z","lastTransitionTime":"2025-10-05T20:55:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:35 crc kubenswrapper[4754]: E1005 20:55:35.137565 4754 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T20:55:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T20:55:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T20:55:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T20:55:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ede2b442-76b0-443a-a450-252ee8711c80\\\",\\\"systemUUID\\\":\\\"72d081dd-fe7e-42a6-9cf0-f0ee7f733ac9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:35Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:35 crc kubenswrapper[4754]: E1005 20:55:35.138079 4754 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 05 20:55:35 crc kubenswrapper[4754]: I1005 20:55:35.140698 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:35 crc kubenswrapper[4754]: I1005 20:55:35.140909 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:35 crc kubenswrapper[4754]: I1005 20:55:35.141057 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:35 crc kubenswrapper[4754]: I1005 20:55:35.141194 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:35 crc kubenswrapper[4754]: I1005 20:55:35.141331 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:35Z","lastTransitionTime":"2025-10-05T20:55:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:35 crc kubenswrapper[4754]: I1005 20:55:35.246130 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:35 crc kubenswrapper[4754]: I1005 20:55:35.246348 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:35 crc kubenswrapper[4754]: I1005 20:55:35.246559 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:35 crc kubenswrapper[4754]: I1005 20:55:35.246589 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:35 crc kubenswrapper[4754]: I1005 20:55:35.246643 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:35Z","lastTransitionTime":"2025-10-05T20:55:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:35 crc kubenswrapper[4754]: I1005 20:55:35.350230 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:35 crc kubenswrapper[4754]: I1005 20:55:35.350703 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:35 crc kubenswrapper[4754]: I1005 20:55:35.350867 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:35 crc kubenswrapper[4754]: I1005 20:55:35.351043 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:35 crc kubenswrapper[4754]: I1005 20:55:35.351279 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:35Z","lastTransitionTime":"2025-10-05T20:55:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:35 crc kubenswrapper[4754]: I1005 20:55:35.455128 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:35 crc kubenswrapper[4754]: I1005 20:55:35.455554 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:35 crc kubenswrapper[4754]: I1005 20:55:35.455719 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:35 crc kubenswrapper[4754]: I1005 20:55:35.455928 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:35 crc kubenswrapper[4754]: I1005 20:55:35.456168 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:35Z","lastTransitionTime":"2025-10-05T20:55:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:35 crc kubenswrapper[4754]: I1005 20:55:35.560432 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:35 crc kubenswrapper[4754]: I1005 20:55:35.560925 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:35 crc kubenswrapper[4754]: I1005 20:55:35.561058 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:35 crc kubenswrapper[4754]: I1005 20:55:35.561198 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:35 crc kubenswrapper[4754]: I1005 20:55:35.561313 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:35Z","lastTransitionTime":"2025-10-05T20:55:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:35 crc kubenswrapper[4754]: I1005 20:55:35.665219 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:35 crc kubenswrapper[4754]: I1005 20:55:35.665287 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:35 crc kubenswrapper[4754]: I1005 20:55:35.665310 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:35 crc kubenswrapper[4754]: I1005 20:55:35.665341 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:35 crc kubenswrapper[4754]: I1005 20:55:35.665360 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:35Z","lastTransitionTime":"2025-10-05T20:55:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:35 crc kubenswrapper[4754]: I1005 20:55:35.769708 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:35 crc kubenswrapper[4754]: I1005 20:55:35.769809 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:35 crc kubenswrapper[4754]: I1005 20:55:35.769830 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:35 crc kubenswrapper[4754]: I1005 20:55:35.769872 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:35 crc kubenswrapper[4754]: I1005 20:55:35.769895 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:35Z","lastTransitionTime":"2025-10-05T20:55:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:35 crc kubenswrapper[4754]: I1005 20:55:35.836673 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nwrnt" Oct 05 20:55:35 crc kubenswrapper[4754]: E1005 20:55:35.836922 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nwrnt" podUID="7dc728ea-2601-44f0-bcab-2913f034007d" Oct 05 20:55:35 crc kubenswrapper[4754]: I1005 20:55:35.874144 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:35 crc kubenswrapper[4754]: I1005 20:55:35.874203 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:35 crc kubenswrapper[4754]: I1005 20:55:35.874219 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:35 crc kubenswrapper[4754]: I1005 20:55:35.874243 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:35 crc kubenswrapper[4754]: I1005 20:55:35.874260 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:35Z","lastTransitionTime":"2025-10-05T20:55:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:35 crc kubenswrapper[4754]: I1005 20:55:35.978893 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:35 crc kubenswrapper[4754]: I1005 20:55:35.978977 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:35 crc kubenswrapper[4754]: I1005 20:55:35.978991 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:35 crc kubenswrapper[4754]: I1005 20:55:35.979012 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:35 crc kubenswrapper[4754]: I1005 20:55:35.979026 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:35Z","lastTransitionTime":"2025-10-05T20:55:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:36 crc kubenswrapper[4754]: I1005 20:55:36.082472 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:36 crc kubenswrapper[4754]: I1005 20:55:36.082577 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:36 crc kubenswrapper[4754]: I1005 20:55:36.082594 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:36 crc kubenswrapper[4754]: I1005 20:55:36.082625 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:36 crc kubenswrapper[4754]: I1005 20:55:36.082647 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:36Z","lastTransitionTime":"2025-10-05T20:55:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:36 crc kubenswrapper[4754]: I1005 20:55:36.185425 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:36 crc kubenswrapper[4754]: I1005 20:55:36.185529 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:36 crc kubenswrapper[4754]: I1005 20:55:36.185549 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:36 crc kubenswrapper[4754]: I1005 20:55:36.185580 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:36 crc kubenswrapper[4754]: I1005 20:55:36.185605 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:36Z","lastTransitionTime":"2025-10-05T20:55:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:36 crc kubenswrapper[4754]: I1005 20:55:36.288252 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:36 crc kubenswrapper[4754]: I1005 20:55:36.288311 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:36 crc kubenswrapper[4754]: I1005 20:55:36.288329 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:36 crc kubenswrapper[4754]: I1005 20:55:36.288362 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:36 crc kubenswrapper[4754]: I1005 20:55:36.288382 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:36Z","lastTransitionTime":"2025-10-05T20:55:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:36 crc kubenswrapper[4754]: I1005 20:55:36.392223 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:36 crc kubenswrapper[4754]: I1005 20:55:36.392293 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:36 crc kubenswrapper[4754]: I1005 20:55:36.392316 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:36 crc kubenswrapper[4754]: I1005 20:55:36.392344 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:36 crc kubenswrapper[4754]: I1005 20:55:36.392364 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:36Z","lastTransitionTime":"2025-10-05T20:55:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:36 crc kubenswrapper[4754]: I1005 20:55:36.496441 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:36 crc kubenswrapper[4754]: I1005 20:55:36.496550 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:36 crc kubenswrapper[4754]: I1005 20:55:36.496568 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:36 crc kubenswrapper[4754]: I1005 20:55:36.496602 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:36 crc kubenswrapper[4754]: I1005 20:55:36.496620 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:36Z","lastTransitionTime":"2025-10-05T20:55:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:36 crc kubenswrapper[4754]: I1005 20:55:36.600824 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:36 crc kubenswrapper[4754]: I1005 20:55:36.600886 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:36 crc kubenswrapper[4754]: I1005 20:55:36.600903 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:36 crc kubenswrapper[4754]: I1005 20:55:36.600923 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:36 crc kubenswrapper[4754]: I1005 20:55:36.600937 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:36Z","lastTransitionTime":"2025-10-05T20:55:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:36 crc kubenswrapper[4754]: I1005 20:55:36.704569 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:36 crc kubenswrapper[4754]: I1005 20:55:36.704972 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:36 crc kubenswrapper[4754]: I1005 20:55:36.705105 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:36 crc kubenswrapper[4754]: I1005 20:55:36.705228 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:36 crc kubenswrapper[4754]: I1005 20:55:36.705352 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:36Z","lastTransitionTime":"2025-10-05T20:55:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:36 crc kubenswrapper[4754]: I1005 20:55:36.808469 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:36 crc kubenswrapper[4754]: I1005 20:55:36.808537 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:36 crc kubenswrapper[4754]: I1005 20:55:36.808547 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:36 crc kubenswrapper[4754]: I1005 20:55:36.808567 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:36 crc kubenswrapper[4754]: I1005 20:55:36.808579 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:36Z","lastTransitionTime":"2025-10-05T20:55:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:36 crc kubenswrapper[4754]: I1005 20:55:36.836309 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 20:55:36 crc kubenswrapper[4754]: I1005 20:55:36.836370 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 20:55:36 crc kubenswrapper[4754]: I1005 20:55:36.836399 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 20:55:36 crc kubenswrapper[4754]: E1005 20:55:36.836593 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 20:55:36 crc kubenswrapper[4754]: E1005 20:55:36.836691 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 20:55:36 crc kubenswrapper[4754]: E1005 20:55:36.836866 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 20:55:36 crc kubenswrapper[4754]: I1005 20:55:36.862789 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be95a413-6a59-45b4-84b7-b43cae694a26\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10b211cf6e47dbd1ee42c82e49de0a13a45581b92a0c4f6bbba0326c7149e41d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2ggg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4bb8c5903e8b70dd19eebb111097d96f725e17de44e06c3e74c067376f6c7360\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2ggg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-b2h9k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:36Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:36 crc kubenswrapper[4754]: I1005 20:55:36.883357 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rddkl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"15afc2a1-5cb7-41c6-954b-446c36822f50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://866d85ee126f8641d12f1a05fabc3de7b9392c324e9fe06ea92214155c05334c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqv8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a175565f83cfcba7825006eafb0f83efb70ec3d415c2c039c6499d5f6576e89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqv8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:55:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rddkl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:36Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:36 crc kubenswrapper[4754]: I1005 20:55:36.913452 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:36 crc kubenswrapper[4754]: I1005 20:55:36.913856 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:36 crc kubenswrapper[4754]: I1005 20:55:36.914045 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:36 crc kubenswrapper[4754]: I1005 20:55:36.914187 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:36 crc kubenswrapper[4754]: I1005 20:55:36.914309 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:36Z","lastTransitionTime":"2025-10-05T20:55:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:36 crc kubenswrapper[4754]: I1005 20:55:36.914172 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7506d496-65ff-41dc-8230-24c8ee653a2a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33da5a6bb3cd3b042c193bc571e6622f2f87f27e560cdd0bbb9be31e2010d707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df6d700a947d19c84b4f65fd2dd2f8d1f4f1f80fa9b4800c2acfa8dae3912639\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35ca8ee77a0e58cf969bd67b277a3392614c4e73b963e6f6002097709370715d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e32c2154f46ba10c509fe40c357304aa8e82d04e9906bbaf2ccecd47b42279e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8907b9b69b249c34ec6db1e460d641b31711b75d6d31d4caafd867339e6571d0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1005 20:54:52.589115 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 20:54:52.590181 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1304624737/tls.crt::/tmp/serving-cert-1304624737/tls.key\\\\\\\"\\\\nI1005 20:54:58.070544 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1005 20:54:58.090352 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1005 20:54:58.090382 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1005 20:54:58.090412 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1005 20:54:58.090418 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1005 20:54:58.144656 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1005 20:54:58.144688 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 20:54:58.144695 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 20:54:58.144699 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1005 20:54:58.144702 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1005 20:54:58.144705 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1005 20:54:58.144708 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1005 20:54:58.145024 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1005 20:54:58.148112 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4efb34f3493cc8854bc4cff6b29f63973c52b754404d7a21daec4e1f612d327d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54c568b637f3737667433047b956c7eb22fdf0f9301acdfd3b49fb84c39de6ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54c568b637f3737667433047b956c7eb22fdf0f9301acdfd3b49fb84c39de6ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:36Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:36 crc kubenswrapper[4754]: I1005 20:55:36.938195 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c76e6e4740035a6ca54191a770c4516e28101aae69afbb89739174c2f076ef5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9eededc0cb1af94fb7b5daba08785f6b077a85e9cb5ddd8b9f397b658194f61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:36Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:36 crc kubenswrapper[4754]: I1005 20:55:36.962926 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tskt7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a68e9c89-465b-4c01-a470-e21a480ee465\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba6edba14cf321b4cd2f7bb44067a3ecf49f1594e76c0197e807ff3daa45e9b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d14ec99610d2167b484bd7712af7bb6d993d26c74b2828da1c89314f16fda22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d14ec99610d2167b484bd7712af7bb6d993d26c74b2828da1c89314f16fda22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f6f08dd853b25881e2f3c66206aa80d0c493518284217780aa40ac4b69f0c0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f6f08dd853b25881e2f3c66206aa80d0c493518284217780aa40ac4b69f0c0b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://71026d00ec1a3d91b019deb3c9f43ee52fb05b8f2d6374e1a5b762c9e5c1b10d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71026d00ec1a3d91b019deb3c9f43ee52fb05b8f2d6374e1a5b762c9e5c1b10d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b9910d2470acb16ebb436d7030b3161061607cb2a5ef50d482d8d5c4f48a8ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b9910d2470acb16ebb436d7030b3161061607cb2a5ef50d482d8d5c4f48a8ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0a93b61eacc55111995e69c15f89a210b8aa15468ac9ac053b7cf6b007d301e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b0a93b61eacc55111995e69c15f89a210b8aa15468ac9ac053b7cf6b007d301e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cafb3208e7e3bcdd01898a68fc5bbf4cb52ca0800b879b77a60bb102794257f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cafb3208e7e3bcdd01898a68fc5bbf4cb52ca0800b879b77a60bb102794257f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tskt7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:36Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:36 crc kubenswrapper[4754]: I1005 20:55:36.984876 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:36Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:37 crc kubenswrapper[4754]: I1005 20:55:37.006725 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ae81d5f-5e31-4648-a80c-001034f0acf8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02dec08702aeaa03bc26151781909a4c6226fdaf37b921feb2a53ad6fa3c94c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a0c1914423693dd5ca96eddab941162a0f1e9efaf4bd02c47514f092e39e9e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5d1820b7c2ba7c5262f52297111f58e8f05771f3d7e34a59f8f3a05a7fb0bfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c187ac7d3514f4e68984ee98c274c53418df1818159cc9d6c16626d162fa3f97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c187ac7d3514f4e68984ee98c274c53418df1818159cc9d6c16626d162fa3f97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:36Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:37Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:37 crc kubenswrapper[4754]: I1005 20:55:37.018043 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:37 crc kubenswrapper[4754]: I1005 20:55:37.018088 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:37 crc kubenswrapper[4754]: I1005 20:55:37.018103 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:37 crc kubenswrapper[4754]: I1005 20:55:37.018125 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:37 crc kubenswrapper[4754]: I1005 20:55:37.018143 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:37Z","lastTransitionTime":"2025-10-05T20:55:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:37 crc kubenswrapper[4754]: I1005 20:55:37.028416 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e380f9b2a0e319dc325c54d1247a6db98346292548fa688cb41a2ca1b3048419\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:37Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:37 crc kubenswrapper[4754]: I1005 20:55:37.061309 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9309179-b92a-4fa4-9fba-747b8f911936\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d969d3d50d6ad280fc6654df9441bab238bb538f277ed6d1273779813ffbaaed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://800a731ed8ef57714b2318af14054ef23f0368936630887a093fc77167b33d12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8cd9b08e83dc174413cc37891c14562816787028471a4201680c86dd620b7a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93b729f3f7f993db0d6bc3ba60c04930aada3cef8272ffcc3c6a8a242a9b1475\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dde1519111fc3af5d14463fdb58d4f5c62a7045e7de98c60981c7e4da37b0e7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ea160a5b37aea7f2970ae4dd2614b09d3dcf5dcbd1fc103b9997f8ecc3a804c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ea160a5b37aea7f2970ae4dd2614b09d3dcf5dcbd1fc103b9997f8ecc3a804c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://17ae6aa6c53e7dd061f70b41428ef50bbc7ffcc65352fad1a4a9d3c69dbd2ba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17ae6aa6c53e7dd061f70b41428ef50bbc7ffcc65352fad1a4a9d3c69dbd2ba9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://ad4cd69df7608c7d87d839fe09d4557c674c7efbc4730f4e4d3daaf9a26f5aee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad4cd69df7608c7d87d839fe09d4557c674c7efbc4730f4e4d3daaf9a26f5aee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:36Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:37Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:37 crc kubenswrapper[4754]: I1005 20:55:37.077722 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:37Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:37 crc kubenswrapper[4754]: I1005 20:55:37.110184 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b44790fb-fecb-4ec3-9816-8e23dd72ab33\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a7dc53e7f96e48213d724a2cb4859514df925005644108ae3a0c33b14dcf52d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfdd8b1e542ab676576c311e2923fb02861f92de91fe1f4b3e4b968c549c375d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3a63451ffaa14cacc3b78d11340f4ea81445fa3af96178e9b172d063d9697d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3cfd3f2274a8a46fa36660e94ac6c9c298c791283f12d52330c5594ac131998\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f626cb108716da0b595844ea0251b7a04d49b5908a1d6dd0aef5cb2d402c4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://822223f515b5add6d71d193fd5bc77df029b8d37a1cc07be4145b53c7226bf61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af1c313e0dc0fe3b48feac3cce1ecd4452c9efadf3d19f257c07cc7fcdec875f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af1c313e0dc0fe3b48feac3cce1ecd4452c9efadf3d19f257c07cc7fcdec875f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-05T20:55:26Z\\\",\\\"message\\\":\\\"5:26.770640 6339 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/iptables-alerter-4ln5h in node crc\\\\nI1005 20:55:26.770478 6339 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-apiserver-operator/metrics]} name:Service_openshift-apiserver-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.38:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {8b82f026-5975-4a1b-bb18-08d5d51147ec}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1005 20:55:26.770652 6339 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/iptables-alerter-4ln5h after 0 failed attempt(s)\\\\nI1005 20:55:26.770662 6339 default_network_controller.go:776] Recording success event on pod openshift-network-operator/iptables-alerter-4ln5h\\\\nF1005 20:55:26.770677 6339 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controlle\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:25Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-7rhps_openshift-ovn-kubernetes(b44790fb-fecb-4ec3-9816-8e23dd72ab33)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2acf9637fa25e5d3ffeb05224c11218c47defc679568fb928a498869613de4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35ebeb7a87d5d47b6567917d38eca38a3e622c497fba58971296c4609c287423\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35ebeb7a87d5d47b6567917d38eca38a3e622c497fba58971296c4609c287423\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7rhps\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:37Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:37 crc kubenswrapper[4754]: I1005 20:55:37.121367 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:37 crc kubenswrapper[4754]: I1005 20:55:37.121627 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:37 crc kubenswrapper[4754]: I1005 20:55:37.121889 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:37 crc kubenswrapper[4754]: I1005 20:55:37.122132 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:37 crc kubenswrapper[4754]: I1005 20:55:37.122418 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:37Z","lastTransitionTime":"2025-10-05T20:55:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:37 crc kubenswrapper[4754]: I1005 20:55:37.126352 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-57sr4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e854c7b9-1229-4da4-80e9-fea6479459b8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceb577e1055d84007936a50d0b3d54b9ae0299a1fc1a6618b37163a6d3fbb9cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vw2ps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:58Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-57sr4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:37Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:37 crc kubenswrapper[4754]: I1005 20:55:37.144111 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cn76s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02d5e3f9-73c2-4496-9aca-0787184aef19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7ba2b076839f4a5f586d2e6c4f4d7188254128b782bfd2180e22bf3ea5ffdd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpx5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cn76s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:37Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:37 crc kubenswrapper[4754]: I1005 20:55:37.162301 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ss7lm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f9ef817-0622-4185-a7fb-57221690aa75\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ba613b0d7753a003c572c3733ed394707c45b6f8072cb7aa88f3c7312b8882b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wpwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:55:02Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ss7lm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:37Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:37 crc kubenswrapper[4754]: I1005 20:55:37.179903 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-nwrnt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7dc728ea-2601-44f0-bcab-2913f034007d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-846xq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-846xq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:55:13Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-nwrnt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:37Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:37 crc kubenswrapper[4754]: I1005 20:55:37.200387 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"934bcd66-9c03-48b9-82dd-5093338f9c8f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9042df2f43243d3d79ba9a3940f8dd4f6c7a89ef9e3510e81fe60b1944d7e83b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5366bbf46a91667355f43f5df7335a536b395cda788b90b35ef18e2b53bc4c02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://600bc3bb551b747cf2f18e26745883232d7a37ac5877fad0c67100e52d7a43aa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78f74982180048eb50b9be79fa72c04f62f925b54fe511920f37e6407f864631\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:37Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:37 crc kubenswrapper[4754]: I1005 20:55:37.222690 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:37Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:37 crc kubenswrapper[4754]: I1005 20:55:37.229580 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:37 crc kubenswrapper[4754]: I1005 20:55:37.229635 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:37 crc kubenswrapper[4754]: I1005 20:55:37.229648 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:37 crc kubenswrapper[4754]: I1005 20:55:37.229670 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:37 crc kubenswrapper[4754]: I1005 20:55:37.229689 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:37Z","lastTransitionTime":"2025-10-05T20:55:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:37 crc kubenswrapper[4754]: I1005 20:55:37.241638 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9ad070bb699b643d29b82bcc1126ce3b585f7492f5ae2817fa491b309b1691d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:37Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:37 crc kubenswrapper[4754]: I1005 20:55:37.334016 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:37 crc kubenswrapper[4754]: I1005 20:55:37.334114 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:37 crc kubenswrapper[4754]: I1005 20:55:37.334142 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:37 crc kubenswrapper[4754]: I1005 20:55:37.334177 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:37 crc kubenswrapper[4754]: I1005 20:55:37.334253 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:37Z","lastTransitionTime":"2025-10-05T20:55:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:37 crc kubenswrapper[4754]: I1005 20:55:37.438922 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:37 crc kubenswrapper[4754]: I1005 20:55:37.439015 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:37 crc kubenswrapper[4754]: I1005 20:55:37.439042 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:37 crc kubenswrapper[4754]: I1005 20:55:37.439076 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:37 crc kubenswrapper[4754]: I1005 20:55:37.439099 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:37Z","lastTransitionTime":"2025-10-05T20:55:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:37 crc kubenswrapper[4754]: I1005 20:55:37.542043 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:37 crc kubenswrapper[4754]: I1005 20:55:37.542108 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:37 crc kubenswrapper[4754]: I1005 20:55:37.542141 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:37 crc kubenswrapper[4754]: I1005 20:55:37.542168 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:37 crc kubenswrapper[4754]: I1005 20:55:37.542187 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:37Z","lastTransitionTime":"2025-10-05T20:55:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:37 crc kubenswrapper[4754]: I1005 20:55:37.646585 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:37 crc kubenswrapper[4754]: I1005 20:55:37.646657 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:37 crc kubenswrapper[4754]: I1005 20:55:37.646675 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:37 crc kubenswrapper[4754]: I1005 20:55:37.646701 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:37 crc kubenswrapper[4754]: I1005 20:55:37.646720 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:37Z","lastTransitionTime":"2025-10-05T20:55:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:37 crc kubenswrapper[4754]: I1005 20:55:37.751419 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:37 crc kubenswrapper[4754]: I1005 20:55:37.751477 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:37 crc kubenswrapper[4754]: I1005 20:55:37.751529 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:37 crc kubenswrapper[4754]: I1005 20:55:37.751556 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:37 crc kubenswrapper[4754]: I1005 20:55:37.751575 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:37Z","lastTransitionTime":"2025-10-05T20:55:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:37 crc kubenswrapper[4754]: I1005 20:55:37.836284 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nwrnt" Oct 05 20:55:37 crc kubenswrapper[4754]: E1005 20:55:37.836445 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nwrnt" podUID="7dc728ea-2601-44f0-bcab-2913f034007d" Oct 05 20:55:37 crc kubenswrapper[4754]: I1005 20:55:37.854779 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:37 crc kubenswrapper[4754]: I1005 20:55:37.854810 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:37 crc kubenswrapper[4754]: I1005 20:55:37.854818 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:37 crc kubenswrapper[4754]: I1005 20:55:37.854831 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:37 crc kubenswrapper[4754]: I1005 20:55:37.854843 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:37Z","lastTransitionTime":"2025-10-05T20:55:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:37 crc kubenswrapper[4754]: I1005 20:55:37.958269 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:37 crc kubenswrapper[4754]: I1005 20:55:37.958326 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:37 crc kubenswrapper[4754]: I1005 20:55:37.958343 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:37 crc kubenswrapper[4754]: I1005 20:55:37.958366 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:37 crc kubenswrapper[4754]: I1005 20:55:37.958385 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:37Z","lastTransitionTime":"2025-10-05T20:55:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:38 crc kubenswrapper[4754]: I1005 20:55:38.061291 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:38 crc kubenswrapper[4754]: I1005 20:55:38.061349 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:38 crc kubenswrapper[4754]: I1005 20:55:38.061369 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:38 crc kubenswrapper[4754]: I1005 20:55:38.061394 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:38 crc kubenswrapper[4754]: I1005 20:55:38.061413 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:38Z","lastTransitionTime":"2025-10-05T20:55:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:38 crc kubenswrapper[4754]: I1005 20:55:38.164769 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:38 crc kubenswrapper[4754]: I1005 20:55:38.164861 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:38 crc kubenswrapper[4754]: I1005 20:55:38.164880 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:38 crc kubenswrapper[4754]: I1005 20:55:38.164910 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:38 crc kubenswrapper[4754]: I1005 20:55:38.164928 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:38Z","lastTransitionTime":"2025-10-05T20:55:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:38 crc kubenswrapper[4754]: I1005 20:55:38.267988 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:38 crc kubenswrapper[4754]: I1005 20:55:38.268045 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:38 crc kubenswrapper[4754]: I1005 20:55:38.268065 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:38 crc kubenswrapper[4754]: I1005 20:55:38.268091 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:38 crc kubenswrapper[4754]: I1005 20:55:38.268109 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:38Z","lastTransitionTime":"2025-10-05T20:55:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:38 crc kubenswrapper[4754]: I1005 20:55:38.371473 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:38 crc kubenswrapper[4754]: I1005 20:55:38.371576 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:38 crc kubenswrapper[4754]: I1005 20:55:38.371597 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:38 crc kubenswrapper[4754]: I1005 20:55:38.371626 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:38 crc kubenswrapper[4754]: I1005 20:55:38.371646 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:38Z","lastTransitionTime":"2025-10-05T20:55:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:38 crc kubenswrapper[4754]: I1005 20:55:38.476205 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:38 crc kubenswrapper[4754]: I1005 20:55:38.476274 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:38 crc kubenswrapper[4754]: I1005 20:55:38.476301 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:38 crc kubenswrapper[4754]: I1005 20:55:38.476328 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:38 crc kubenswrapper[4754]: I1005 20:55:38.476351 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:38Z","lastTransitionTime":"2025-10-05T20:55:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:38 crc kubenswrapper[4754]: I1005 20:55:38.580750 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:38 crc kubenswrapper[4754]: I1005 20:55:38.580809 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:38 crc kubenswrapper[4754]: I1005 20:55:38.580827 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:38 crc kubenswrapper[4754]: I1005 20:55:38.580852 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:38 crc kubenswrapper[4754]: I1005 20:55:38.580870 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:38Z","lastTransitionTime":"2025-10-05T20:55:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:38 crc kubenswrapper[4754]: I1005 20:55:38.684171 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:38 crc kubenswrapper[4754]: I1005 20:55:38.684241 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:38 crc kubenswrapper[4754]: I1005 20:55:38.684262 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:38 crc kubenswrapper[4754]: I1005 20:55:38.684287 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:38 crc kubenswrapper[4754]: I1005 20:55:38.684309 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:38Z","lastTransitionTime":"2025-10-05T20:55:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:38 crc kubenswrapper[4754]: I1005 20:55:38.787516 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:38 crc kubenswrapper[4754]: I1005 20:55:38.787561 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:38 crc kubenswrapper[4754]: I1005 20:55:38.787574 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:38 crc kubenswrapper[4754]: I1005 20:55:38.787594 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:38 crc kubenswrapper[4754]: I1005 20:55:38.787608 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:38Z","lastTransitionTime":"2025-10-05T20:55:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:38 crc kubenswrapper[4754]: I1005 20:55:38.836676 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 20:55:38 crc kubenswrapper[4754]: I1005 20:55:38.836796 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 20:55:38 crc kubenswrapper[4754]: E1005 20:55:38.836901 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 20:55:38 crc kubenswrapper[4754]: I1005 20:55:38.836907 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 20:55:38 crc kubenswrapper[4754]: E1005 20:55:38.836999 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 20:55:38 crc kubenswrapper[4754]: E1005 20:55:38.837047 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 20:55:38 crc kubenswrapper[4754]: I1005 20:55:38.890301 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:38 crc kubenswrapper[4754]: I1005 20:55:38.890370 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:38 crc kubenswrapper[4754]: I1005 20:55:38.890391 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:38 crc kubenswrapper[4754]: I1005 20:55:38.890419 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:38 crc kubenswrapper[4754]: I1005 20:55:38.890460 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:38Z","lastTransitionTime":"2025-10-05T20:55:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:38 crc kubenswrapper[4754]: I1005 20:55:38.994069 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:38 crc kubenswrapper[4754]: I1005 20:55:38.994147 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:38 crc kubenswrapper[4754]: I1005 20:55:38.994167 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:38 crc kubenswrapper[4754]: I1005 20:55:38.994197 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:38 crc kubenswrapper[4754]: I1005 20:55:38.994218 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:38Z","lastTransitionTime":"2025-10-05T20:55:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:39 crc kubenswrapper[4754]: I1005 20:55:39.097193 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:39 crc kubenswrapper[4754]: I1005 20:55:39.097250 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:39 crc kubenswrapper[4754]: I1005 20:55:39.097269 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:39 crc kubenswrapper[4754]: I1005 20:55:39.097289 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:39 crc kubenswrapper[4754]: I1005 20:55:39.097302 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:39Z","lastTransitionTime":"2025-10-05T20:55:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:39 crc kubenswrapper[4754]: I1005 20:55:39.201230 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:39 crc kubenswrapper[4754]: I1005 20:55:39.201304 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:39 crc kubenswrapper[4754]: I1005 20:55:39.201320 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:39 crc kubenswrapper[4754]: I1005 20:55:39.201349 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:39 crc kubenswrapper[4754]: I1005 20:55:39.201368 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:39Z","lastTransitionTime":"2025-10-05T20:55:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:39 crc kubenswrapper[4754]: I1005 20:55:39.305356 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:39 crc kubenswrapper[4754]: I1005 20:55:39.305436 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:39 crc kubenswrapper[4754]: I1005 20:55:39.305451 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:39 crc kubenswrapper[4754]: I1005 20:55:39.305478 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:39 crc kubenswrapper[4754]: I1005 20:55:39.305525 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:39Z","lastTransitionTime":"2025-10-05T20:55:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:39 crc kubenswrapper[4754]: I1005 20:55:39.408614 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:39 crc kubenswrapper[4754]: I1005 20:55:39.408663 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:39 crc kubenswrapper[4754]: I1005 20:55:39.408673 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:39 crc kubenswrapper[4754]: I1005 20:55:39.408691 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:39 crc kubenswrapper[4754]: I1005 20:55:39.408704 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:39Z","lastTransitionTime":"2025-10-05T20:55:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:39 crc kubenswrapper[4754]: I1005 20:55:39.512164 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:39 crc kubenswrapper[4754]: I1005 20:55:39.512237 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:39 crc kubenswrapper[4754]: I1005 20:55:39.512251 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:39 crc kubenswrapper[4754]: I1005 20:55:39.512273 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:39 crc kubenswrapper[4754]: I1005 20:55:39.512289 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:39Z","lastTransitionTime":"2025-10-05T20:55:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:39 crc kubenswrapper[4754]: I1005 20:55:39.615707 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:39 crc kubenswrapper[4754]: I1005 20:55:39.615763 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:39 crc kubenswrapper[4754]: I1005 20:55:39.615776 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:39 crc kubenswrapper[4754]: I1005 20:55:39.615800 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:39 crc kubenswrapper[4754]: I1005 20:55:39.615815 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:39Z","lastTransitionTime":"2025-10-05T20:55:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:39 crc kubenswrapper[4754]: I1005 20:55:39.720675 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:39 crc kubenswrapper[4754]: I1005 20:55:39.720744 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:39 crc kubenswrapper[4754]: I1005 20:55:39.720762 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:39 crc kubenswrapper[4754]: I1005 20:55:39.720847 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:39 crc kubenswrapper[4754]: I1005 20:55:39.720876 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:39Z","lastTransitionTime":"2025-10-05T20:55:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:39 crc kubenswrapper[4754]: I1005 20:55:39.824897 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:39 crc kubenswrapper[4754]: I1005 20:55:39.824954 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:39 crc kubenswrapper[4754]: I1005 20:55:39.824966 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:39 crc kubenswrapper[4754]: I1005 20:55:39.824990 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:39 crc kubenswrapper[4754]: I1005 20:55:39.825005 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:39Z","lastTransitionTime":"2025-10-05T20:55:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:39 crc kubenswrapper[4754]: I1005 20:55:39.836410 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nwrnt" Oct 05 20:55:39 crc kubenswrapper[4754]: E1005 20:55:39.836654 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nwrnt" podUID="7dc728ea-2601-44f0-bcab-2913f034007d" Oct 05 20:55:39 crc kubenswrapper[4754]: I1005 20:55:39.928327 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:39 crc kubenswrapper[4754]: I1005 20:55:39.928371 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:39 crc kubenswrapper[4754]: I1005 20:55:39.928380 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:39 crc kubenswrapper[4754]: I1005 20:55:39.928398 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:39 crc kubenswrapper[4754]: I1005 20:55:39.928409 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:39Z","lastTransitionTime":"2025-10-05T20:55:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:40 crc kubenswrapper[4754]: I1005 20:55:40.031911 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:40 crc kubenswrapper[4754]: I1005 20:55:40.031970 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:40 crc kubenswrapper[4754]: I1005 20:55:40.031981 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:40 crc kubenswrapper[4754]: I1005 20:55:40.032012 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:40 crc kubenswrapper[4754]: I1005 20:55:40.032024 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:40Z","lastTransitionTime":"2025-10-05T20:55:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:40 crc kubenswrapper[4754]: I1005 20:55:40.136284 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:40 crc kubenswrapper[4754]: I1005 20:55:40.136371 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:40 crc kubenswrapper[4754]: I1005 20:55:40.136387 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:40 crc kubenswrapper[4754]: I1005 20:55:40.136407 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:40 crc kubenswrapper[4754]: I1005 20:55:40.136421 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:40Z","lastTransitionTime":"2025-10-05T20:55:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:40 crc kubenswrapper[4754]: I1005 20:55:40.243222 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:40 crc kubenswrapper[4754]: I1005 20:55:40.243281 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:40 crc kubenswrapper[4754]: I1005 20:55:40.243298 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:40 crc kubenswrapper[4754]: I1005 20:55:40.243327 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:40 crc kubenswrapper[4754]: I1005 20:55:40.243345 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:40Z","lastTransitionTime":"2025-10-05T20:55:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:40 crc kubenswrapper[4754]: I1005 20:55:40.346197 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:40 crc kubenswrapper[4754]: I1005 20:55:40.346262 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:40 crc kubenswrapper[4754]: I1005 20:55:40.346276 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:40 crc kubenswrapper[4754]: I1005 20:55:40.346295 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:40 crc kubenswrapper[4754]: I1005 20:55:40.346308 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:40Z","lastTransitionTime":"2025-10-05T20:55:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:40 crc kubenswrapper[4754]: I1005 20:55:40.450217 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:40 crc kubenswrapper[4754]: I1005 20:55:40.450304 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:40 crc kubenswrapper[4754]: I1005 20:55:40.450326 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:40 crc kubenswrapper[4754]: I1005 20:55:40.450357 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:40 crc kubenswrapper[4754]: I1005 20:55:40.450379 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:40Z","lastTransitionTime":"2025-10-05T20:55:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:40 crc kubenswrapper[4754]: I1005 20:55:40.554322 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:40 crc kubenswrapper[4754]: I1005 20:55:40.554432 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:40 crc kubenswrapper[4754]: I1005 20:55:40.554452 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:40 crc kubenswrapper[4754]: I1005 20:55:40.554486 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:40 crc kubenswrapper[4754]: I1005 20:55:40.554529 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:40Z","lastTransitionTime":"2025-10-05T20:55:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:40 crc kubenswrapper[4754]: I1005 20:55:40.658377 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:40 crc kubenswrapper[4754]: I1005 20:55:40.658450 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:40 crc kubenswrapper[4754]: I1005 20:55:40.658469 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:40 crc kubenswrapper[4754]: I1005 20:55:40.658546 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:40 crc kubenswrapper[4754]: I1005 20:55:40.658567 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:40Z","lastTransitionTime":"2025-10-05T20:55:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:40 crc kubenswrapper[4754]: I1005 20:55:40.762306 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:40 crc kubenswrapper[4754]: I1005 20:55:40.762386 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:40 crc kubenswrapper[4754]: I1005 20:55:40.762405 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:40 crc kubenswrapper[4754]: I1005 20:55:40.762438 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:40 crc kubenswrapper[4754]: I1005 20:55:40.762456 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:40Z","lastTransitionTime":"2025-10-05T20:55:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:40 crc kubenswrapper[4754]: I1005 20:55:40.837226 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 20:55:40 crc kubenswrapper[4754]: I1005 20:55:40.837288 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 20:55:40 crc kubenswrapper[4754]: E1005 20:55:40.837461 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 20:55:40 crc kubenswrapper[4754]: I1005 20:55:40.837481 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 20:55:40 crc kubenswrapper[4754]: E1005 20:55:40.838007 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 20:55:40 crc kubenswrapper[4754]: E1005 20:55:40.838695 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 20:55:40 crc kubenswrapper[4754]: I1005 20:55:40.865900 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:40 crc kubenswrapper[4754]: I1005 20:55:40.865947 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:40 crc kubenswrapper[4754]: I1005 20:55:40.865961 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:40 crc kubenswrapper[4754]: I1005 20:55:40.865984 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:40 crc kubenswrapper[4754]: I1005 20:55:40.865996 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:40Z","lastTransitionTime":"2025-10-05T20:55:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:40 crc kubenswrapper[4754]: I1005 20:55:40.968635 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:40 crc kubenswrapper[4754]: I1005 20:55:40.968685 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:40 crc kubenswrapper[4754]: I1005 20:55:40.968698 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:40 crc kubenswrapper[4754]: I1005 20:55:40.968716 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:40 crc kubenswrapper[4754]: I1005 20:55:40.968728 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:40Z","lastTransitionTime":"2025-10-05T20:55:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:41 crc kubenswrapper[4754]: I1005 20:55:41.071909 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:41 crc kubenswrapper[4754]: I1005 20:55:41.071977 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:41 crc kubenswrapper[4754]: I1005 20:55:41.071990 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:41 crc kubenswrapper[4754]: I1005 20:55:41.072012 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:41 crc kubenswrapper[4754]: I1005 20:55:41.072028 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:41Z","lastTransitionTime":"2025-10-05T20:55:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:41 crc kubenswrapper[4754]: I1005 20:55:41.175823 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:41 crc kubenswrapper[4754]: I1005 20:55:41.175898 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:41 crc kubenswrapper[4754]: I1005 20:55:41.175917 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:41 crc kubenswrapper[4754]: I1005 20:55:41.176397 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:41 crc kubenswrapper[4754]: I1005 20:55:41.176454 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:41Z","lastTransitionTime":"2025-10-05T20:55:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:41 crc kubenswrapper[4754]: I1005 20:55:41.279933 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:41 crc kubenswrapper[4754]: I1005 20:55:41.279966 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:41 crc kubenswrapper[4754]: I1005 20:55:41.279976 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:41 crc kubenswrapper[4754]: I1005 20:55:41.279991 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:41 crc kubenswrapper[4754]: I1005 20:55:41.280001 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:41Z","lastTransitionTime":"2025-10-05T20:55:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:41 crc kubenswrapper[4754]: I1005 20:55:41.382898 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:41 crc kubenswrapper[4754]: I1005 20:55:41.382947 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:41 crc kubenswrapper[4754]: I1005 20:55:41.382963 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:41 crc kubenswrapper[4754]: I1005 20:55:41.382986 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:41 crc kubenswrapper[4754]: I1005 20:55:41.383003 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:41Z","lastTransitionTime":"2025-10-05T20:55:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:41 crc kubenswrapper[4754]: I1005 20:55:41.486823 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:41 crc kubenswrapper[4754]: I1005 20:55:41.486904 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:41 crc kubenswrapper[4754]: I1005 20:55:41.486924 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:41 crc kubenswrapper[4754]: I1005 20:55:41.486949 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:41 crc kubenswrapper[4754]: I1005 20:55:41.486966 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:41Z","lastTransitionTime":"2025-10-05T20:55:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:41 crc kubenswrapper[4754]: I1005 20:55:41.590892 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:41 crc kubenswrapper[4754]: I1005 20:55:41.590963 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:41 crc kubenswrapper[4754]: I1005 20:55:41.590985 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:41 crc kubenswrapper[4754]: I1005 20:55:41.591017 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:41 crc kubenswrapper[4754]: I1005 20:55:41.591036 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:41Z","lastTransitionTime":"2025-10-05T20:55:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:41 crc kubenswrapper[4754]: I1005 20:55:41.693917 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:41 crc kubenswrapper[4754]: I1005 20:55:41.693951 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:41 crc kubenswrapper[4754]: I1005 20:55:41.693962 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:41 crc kubenswrapper[4754]: I1005 20:55:41.693980 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:41 crc kubenswrapper[4754]: I1005 20:55:41.693992 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:41Z","lastTransitionTime":"2025-10-05T20:55:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:41 crc kubenswrapper[4754]: I1005 20:55:41.797202 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:41 crc kubenswrapper[4754]: I1005 20:55:41.797250 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:41 crc kubenswrapper[4754]: I1005 20:55:41.797260 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:41 crc kubenswrapper[4754]: I1005 20:55:41.797277 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:41 crc kubenswrapper[4754]: I1005 20:55:41.797289 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:41Z","lastTransitionTime":"2025-10-05T20:55:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:41 crc kubenswrapper[4754]: I1005 20:55:41.836725 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nwrnt" Oct 05 20:55:41 crc kubenswrapper[4754]: E1005 20:55:41.836917 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nwrnt" podUID="7dc728ea-2601-44f0-bcab-2913f034007d" Oct 05 20:55:41 crc kubenswrapper[4754]: I1005 20:55:41.900443 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:41 crc kubenswrapper[4754]: I1005 20:55:41.900482 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:41 crc kubenswrapper[4754]: I1005 20:55:41.900519 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:41 crc kubenswrapper[4754]: I1005 20:55:41.900538 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:41 crc kubenswrapper[4754]: I1005 20:55:41.900551 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:41Z","lastTransitionTime":"2025-10-05T20:55:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:42 crc kubenswrapper[4754]: I1005 20:55:42.002864 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:42 crc kubenswrapper[4754]: I1005 20:55:42.002904 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:42 crc kubenswrapper[4754]: I1005 20:55:42.002915 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:42 crc kubenswrapper[4754]: I1005 20:55:42.002936 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:42 crc kubenswrapper[4754]: I1005 20:55:42.002947 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:42Z","lastTransitionTime":"2025-10-05T20:55:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:42 crc kubenswrapper[4754]: I1005 20:55:42.105321 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:42 crc kubenswrapper[4754]: I1005 20:55:42.105371 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:42 crc kubenswrapper[4754]: I1005 20:55:42.105384 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:42 crc kubenswrapper[4754]: I1005 20:55:42.105403 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:42 crc kubenswrapper[4754]: I1005 20:55:42.105419 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:42Z","lastTransitionTime":"2025-10-05T20:55:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:42 crc kubenswrapper[4754]: I1005 20:55:42.207771 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:42 crc kubenswrapper[4754]: I1005 20:55:42.207836 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:42 crc kubenswrapper[4754]: I1005 20:55:42.207847 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:42 crc kubenswrapper[4754]: I1005 20:55:42.207866 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:42 crc kubenswrapper[4754]: I1005 20:55:42.207877 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:42Z","lastTransitionTime":"2025-10-05T20:55:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:42 crc kubenswrapper[4754]: I1005 20:55:42.311292 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:42 crc kubenswrapper[4754]: I1005 20:55:42.311337 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:42 crc kubenswrapper[4754]: I1005 20:55:42.311347 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:42 crc kubenswrapper[4754]: I1005 20:55:42.311367 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:42 crc kubenswrapper[4754]: I1005 20:55:42.311379 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:42Z","lastTransitionTime":"2025-10-05T20:55:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:42 crc kubenswrapper[4754]: I1005 20:55:42.413282 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:42 crc kubenswrapper[4754]: I1005 20:55:42.413315 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:42 crc kubenswrapper[4754]: I1005 20:55:42.413324 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:42 crc kubenswrapper[4754]: I1005 20:55:42.413339 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:42 crc kubenswrapper[4754]: I1005 20:55:42.413350 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:42Z","lastTransitionTime":"2025-10-05T20:55:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:42 crc kubenswrapper[4754]: I1005 20:55:42.515927 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:42 crc kubenswrapper[4754]: I1005 20:55:42.515986 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:42 crc kubenswrapper[4754]: I1005 20:55:42.516003 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:42 crc kubenswrapper[4754]: I1005 20:55:42.516049 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:42 crc kubenswrapper[4754]: I1005 20:55:42.516072 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:42Z","lastTransitionTime":"2025-10-05T20:55:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:42 crc kubenswrapper[4754]: I1005 20:55:42.618751 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:42 crc kubenswrapper[4754]: I1005 20:55:42.618796 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:42 crc kubenswrapper[4754]: I1005 20:55:42.618805 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:42 crc kubenswrapper[4754]: I1005 20:55:42.618822 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:42 crc kubenswrapper[4754]: I1005 20:55:42.618834 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:42Z","lastTransitionTime":"2025-10-05T20:55:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:42 crc kubenswrapper[4754]: I1005 20:55:42.721879 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:42 crc kubenswrapper[4754]: I1005 20:55:42.721950 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:42 crc kubenswrapper[4754]: I1005 20:55:42.721969 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:42 crc kubenswrapper[4754]: I1005 20:55:42.721999 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:42 crc kubenswrapper[4754]: I1005 20:55:42.722037 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:42Z","lastTransitionTime":"2025-10-05T20:55:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:42 crc kubenswrapper[4754]: I1005 20:55:42.824489 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:42 crc kubenswrapper[4754]: I1005 20:55:42.824624 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:42 crc kubenswrapper[4754]: I1005 20:55:42.824644 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:42 crc kubenswrapper[4754]: I1005 20:55:42.824674 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:42 crc kubenswrapper[4754]: I1005 20:55:42.824695 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:42Z","lastTransitionTime":"2025-10-05T20:55:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:42 crc kubenswrapper[4754]: I1005 20:55:42.842108 4754 scope.go:117] "RemoveContainer" containerID="af1c313e0dc0fe3b48feac3cce1ecd4452c9efadf3d19f257c07cc7fcdec875f" Oct 05 20:55:42 crc kubenswrapper[4754]: E1005 20:55:42.842391 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-7rhps_openshift-ovn-kubernetes(b44790fb-fecb-4ec3-9816-8e23dd72ab33)\"" pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" podUID="b44790fb-fecb-4ec3-9816-8e23dd72ab33" Oct 05 20:55:42 crc kubenswrapper[4754]: I1005 20:55:42.842687 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 20:55:42 crc kubenswrapper[4754]: E1005 20:55:42.842798 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 20:55:42 crc kubenswrapper[4754]: I1005 20:55:42.843015 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 20:55:42 crc kubenswrapper[4754]: E1005 20:55:42.843109 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 20:55:42 crc kubenswrapper[4754]: I1005 20:55:42.843317 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 20:55:42 crc kubenswrapper[4754]: E1005 20:55:42.843411 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 20:55:42 crc kubenswrapper[4754]: I1005 20:55:42.927632 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:42 crc kubenswrapper[4754]: I1005 20:55:42.927690 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:42 crc kubenswrapper[4754]: I1005 20:55:42.927704 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:42 crc kubenswrapper[4754]: I1005 20:55:42.927729 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:42 crc kubenswrapper[4754]: I1005 20:55:42.927746 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:42Z","lastTransitionTime":"2025-10-05T20:55:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:43 crc kubenswrapper[4754]: I1005 20:55:43.030354 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:43 crc kubenswrapper[4754]: I1005 20:55:43.030418 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:43 crc kubenswrapper[4754]: I1005 20:55:43.030443 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:43 crc kubenswrapper[4754]: I1005 20:55:43.030473 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:43 crc kubenswrapper[4754]: I1005 20:55:43.030520 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:43Z","lastTransitionTime":"2025-10-05T20:55:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:43 crc kubenswrapper[4754]: I1005 20:55:43.133743 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:43 crc kubenswrapper[4754]: I1005 20:55:43.133790 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:43 crc kubenswrapper[4754]: I1005 20:55:43.133802 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:43 crc kubenswrapper[4754]: I1005 20:55:43.133827 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:43 crc kubenswrapper[4754]: I1005 20:55:43.133844 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:43Z","lastTransitionTime":"2025-10-05T20:55:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:43 crc kubenswrapper[4754]: I1005 20:55:43.237626 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:43 crc kubenswrapper[4754]: I1005 20:55:43.237671 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:43 crc kubenswrapper[4754]: I1005 20:55:43.237684 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:43 crc kubenswrapper[4754]: I1005 20:55:43.237705 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:43 crc kubenswrapper[4754]: I1005 20:55:43.237721 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:43Z","lastTransitionTime":"2025-10-05T20:55:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:43 crc kubenswrapper[4754]: I1005 20:55:43.341173 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:43 crc kubenswrapper[4754]: I1005 20:55:43.341222 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:43 crc kubenswrapper[4754]: I1005 20:55:43.341235 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:43 crc kubenswrapper[4754]: I1005 20:55:43.341252 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:43 crc kubenswrapper[4754]: I1005 20:55:43.341264 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:43Z","lastTransitionTime":"2025-10-05T20:55:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:43 crc kubenswrapper[4754]: I1005 20:55:43.444444 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:43 crc kubenswrapper[4754]: I1005 20:55:43.444514 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:43 crc kubenswrapper[4754]: I1005 20:55:43.444530 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:43 crc kubenswrapper[4754]: I1005 20:55:43.444552 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:43 crc kubenswrapper[4754]: I1005 20:55:43.444564 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:43Z","lastTransitionTime":"2025-10-05T20:55:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:43 crc kubenswrapper[4754]: I1005 20:55:43.549359 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:43 crc kubenswrapper[4754]: I1005 20:55:43.549417 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:43 crc kubenswrapper[4754]: I1005 20:55:43.549440 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:43 crc kubenswrapper[4754]: I1005 20:55:43.549468 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:43 crc kubenswrapper[4754]: I1005 20:55:43.549487 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:43Z","lastTransitionTime":"2025-10-05T20:55:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:43 crc kubenswrapper[4754]: I1005 20:55:43.653095 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:43 crc kubenswrapper[4754]: I1005 20:55:43.653156 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:43 crc kubenswrapper[4754]: I1005 20:55:43.653176 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:43 crc kubenswrapper[4754]: I1005 20:55:43.653203 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:43 crc kubenswrapper[4754]: I1005 20:55:43.653226 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:43Z","lastTransitionTime":"2025-10-05T20:55:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:43 crc kubenswrapper[4754]: I1005 20:55:43.756847 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:43 crc kubenswrapper[4754]: I1005 20:55:43.756946 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:43 crc kubenswrapper[4754]: I1005 20:55:43.756976 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:43 crc kubenswrapper[4754]: I1005 20:55:43.757010 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:43 crc kubenswrapper[4754]: I1005 20:55:43.757039 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:43Z","lastTransitionTime":"2025-10-05T20:55:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:43 crc kubenswrapper[4754]: I1005 20:55:43.837266 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nwrnt" Oct 05 20:55:43 crc kubenswrapper[4754]: E1005 20:55:43.837431 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nwrnt" podUID="7dc728ea-2601-44f0-bcab-2913f034007d" Oct 05 20:55:43 crc kubenswrapper[4754]: I1005 20:55:43.859882 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:43 crc kubenswrapper[4754]: I1005 20:55:43.859956 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:43 crc kubenswrapper[4754]: I1005 20:55:43.859977 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:43 crc kubenswrapper[4754]: I1005 20:55:43.860004 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:43 crc kubenswrapper[4754]: I1005 20:55:43.860024 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:43Z","lastTransitionTime":"2025-10-05T20:55:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:43 crc kubenswrapper[4754]: I1005 20:55:43.962947 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:43 crc kubenswrapper[4754]: I1005 20:55:43.963002 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:43 crc kubenswrapper[4754]: I1005 20:55:43.963015 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:43 crc kubenswrapper[4754]: I1005 20:55:43.963036 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:43 crc kubenswrapper[4754]: I1005 20:55:43.963050 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:43Z","lastTransitionTime":"2025-10-05T20:55:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:44 crc kubenswrapper[4754]: I1005 20:55:44.065368 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:44 crc kubenswrapper[4754]: I1005 20:55:44.065420 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:44 crc kubenswrapper[4754]: I1005 20:55:44.065431 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:44 crc kubenswrapper[4754]: I1005 20:55:44.065449 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:44 crc kubenswrapper[4754]: I1005 20:55:44.065460 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:44Z","lastTransitionTime":"2025-10-05T20:55:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:44 crc kubenswrapper[4754]: I1005 20:55:44.167436 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:44 crc kubenswrapper[4754]: I1005 20:55:44.167484 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:44 crc kubenswrapper[4754]: I1005 20:55:44.167517 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:44 crc kubenswrapper[4754]: I1005 20:55:44.167539 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:44 crc kubenswrapper[4754]: I1005 20:55:44.167559 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:44Z","lastTransitionTime":"2025-10-05T20:55:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:44 crc kubenswrapper[4754]: I1005 20:55:44.270310 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:44 crc kubenswrapper[4754]: I1005 20:55:44.270391 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:44 crc kubenswrapper[4754]: I1005 20:55:44.270405 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:44 crc kubenswrapper[4754]: I1005 20:55:44.270431 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:44 crc kubenswrapper[4754]: I1005 20:55:44.270445 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:44Z","lastTransitionTime":"2025-10-05T20:55:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:44 crc kubenswrapper[4754]: I1005 20:55:44.374187 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:44 crc kubenswrapper[4754]: I1005 20:55:44.374248 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:44 crc kubenswrapper[4754]: I1005 20:55:44.374263 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:44 crc kubenswrapper[4754]: I1005 20:55:44.374289 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:44 crc kubenswrapper[4754]: I1005 20:55:44.374306 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:44Z","lastTransitionTime":"2025-10-05T20:55:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:44 crc kubenswrapper[4754]: I1005 20:55:44.478173 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:44 crc kubenswrapper[4754]: I1005 20:55:44.478260 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:44 crc kubenswrapper[4754]: I1005 20:55:44.478280 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:44 crc kubenswrapper[4754]: I1005 20:55:44.478311 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:44 crc kubenswrapper[4754]: I1005 20:55:44.478331 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:44Z","lastTransitionTime":"2025-10-05T20:55:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:44 crc kubenswrapper[4754]: I1005 20:55:44.594579 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:44 crc kubenswrapper[4754]: I1005 20:55:44.594769 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:44 crc kubenswrapper[4754]: I1005 20:55:44.594863 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:44 crc kubenswrapper[4754]: I1005 20:55:44.594948 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:44 crc kubenswrapper[4754]: I1005 20:55:44.595046 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:44Z","lastTransitionTime":"2025-10-05T20:55:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:44 crc kubenswrapper[4754]: I1005 20:55:44.697779 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:44 crc kubenswrapper[4754]: I1005 20:55:44.697835 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:44 crc kubenswrapper[4754]: I1005 20:55:44.697848 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:44 crc kubenswrapper[4754]: I1005 20:55:44.697872 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:44 crc kubenswrapper[4754]: I1005 20:55:44.697885 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:44Z","lastTransitionTime":"2025-10-05T20:55:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:44 crc kubenswrapper[4754]: I1005 20:55:44.801348 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:44 crc kubenswrapper[4754]: I1005 20:55:44.801426 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:44 crc kubenswrapper[4754]: I1005 20:55:44.801448 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:44 crc kubenswrapper[4754]: I1005 20:55:44.801477 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:44 crc kubenswrapper[4754]: I1005 20:55:44.801533 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:44Z","lastTransitionTime":"2025-10-05T20:55:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:44 crc kubenswrapper[4754]: I1005 20:55:44.836812 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 20:55:44 crc kubenswrapper[4754]: I1005 20:55:44.836908 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 20:55:44 crc kubenswrapper[4754]: I1005 20:55:44.836844 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 20:55:44 crc kubenswrapper[4754]: E1005 20:55:44.837043 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 20:55:44 crc kubenswrapper[4754]: E1005 20:55:44.837222 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 20:55:44 crc kubenswrapper[4754]: E1005 20:55:44.837316 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 20:55:44 crc kubenswrapper[4754]: I1005 20:55:44.904574 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:44 crc kubenswrapper[4754]: I1005 20:55:44.904608 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:44 crc kubenswrapper[4754]: I1005 20:55:44.904617 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:44 crc kubenswrapper[4754]: I1005 20:55:44.904632 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:44 crc kubenswrapper[4754]: I1005 20:55:44.904643 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:44Z","lastTransitionTime":"2025-10-05T20:55:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:45 crc kubenswrapper[4754]: I1005 20:55:45.007243 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:45 crc kubenswrapper[4754]: I1005 20:55:45.007305 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:45 crc kubenswrapper[4754]: I1005 20:55:45.007325 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:45 crc kubenswrapper[4754]: I1005 20:55:45.007355 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:45 crc kubenswrapper[4754]: I1005 20:55:45.007373 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:45Z","lastTransitionTime":"2025-10-05T20:55:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:45 crc kubenswrapper[4754]: I1005 20:55:45.110524 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:45 crc kubenswrapper[4754]: I1005 20:55:45.110596 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:45 crc kubenswrapper[4754]: I1005 20:55:45.110614 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:45 crc kubenswrapper[4754]: I1005 20:55:45.110640 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:45 crc kubenswrapper[4754]: I1005 20:55:45.110660 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:45Z","lastTransitionTime":"2025-10-05T20:55:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:45 crc kubenswrapper[4754]: I1005 20:55:45.213767 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:45 crc kubenswrapper[4754]: I1005 20:55:45.214139 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:45 crc kubenswrapper[4754]: I1005 20:55:45.214218 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:45 crc kubenswrapper[4754]: I1005 20:55:45.214692 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:45 crc kubenswrapper[4754]: I1005 20:55:45.214775 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:45Z","lastTransitionTime":"2025-10-05T20:55:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:45 crc kubenswrapper[4754]: I1005 20:55:45.318683 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:45 crc kubenswrapper[4754]: I1005 20:55:45.318742 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:45 crc kubenswrapper[4754]: I1005 20:55:45.318763 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:45 crc kubenswrapper[4754]: I1005 20:55:45.318788 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:45 crc kubenswrapper[4754]: I1005 20:55:45.318806 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:45Z","lastTransitionTime":"2025-10-05T20:55:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:45 crc kubenswrapper[4754]: I1005 20:55:45.421597 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:45 crc kubenswrapper[4754]: I1005 20:55:45.421643 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:45 crc kubenswrapper[4754]: I1005 20:55:45.421654 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:45 crc kubenswrapper[4754]: I1005 20:55:45.421674 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:45 crc kubenswrapper[4754]: I1005 20:55:45.421687 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:45Z","lastTransitionTime":"2025-10-05T20:55:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:45 crc kubenswrapper[4754]: I1005 20:55:45.462468 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:45 crc kubenswrapper[4754]: I1005 20:55:45.462578 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:45 crc kubenswrapper[4754]: I1005 20:55:45.462597 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:45 crc kubenswrapper[4754]: I1005 20:55:45.462623 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:45 crc kubenswrapper[4754]: I1005 20:55:45.462641 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:45Z","lastTransitionTime":"2025-10-05T20:55:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:45 crc kubenswrapper[4754]: E1005 20:55:45.478861 4754 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T20:55:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T20:55:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T20:55:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T20:55:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ede2b442-76b0-443a-a450-252ee8711c80\\\",\\\"systemUUID\\\":\\\"72d081dd-fe7e-42a6-9cf0-f0ee7f733ac9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:45Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:45 crc kubenswrapper[4754]: I1005 20:55:45.483797 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:45 crc kubenswrapper[4754]: I1005 20:55:45.483845 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:45 crc kubenswrapper[4754]: I1005 20:55:45.483859 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:45 crc kubenswrapper[4754]: I1005 20:55:45.483884 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:45 crc kubenswrapper[4754]: I1005 20:55:45.483905 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:45Z","lastTransitionTime":"2025-10-05T20:55:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:45 crc kubenswrapper[4754]: E1005 20:55:45.499185 4754 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T20:55:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T20:55:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T20:55:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T20:55:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ede2b442-76b0-443a-a450-252ee8711c80\\\",\\\"systemUUID\\\":\\\"72d081dd-fe7e-42a6-9cf0-f0ee7f733ac9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:45Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:45 crc kubenswrapper[4754]: I1005 20:55:45.503234 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:45 crc kubenswrapper[4754]: I1005 20:55:45.503265 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:45 crc kubenswrapper[4754]: I1005 20:55:45.503279 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:45 crc kubenswrapper[4754]: I1005 20:55:45.503324 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:45 crc kubenswrapper[4754]: I1005 20:55:45.503336 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:45Z","lastTransitionTime":"2025-10-05T20:55:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:45 crc kubenswrapper[4754]: E1005 20:55:45.518313 4754 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T20:55:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T20:55:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T20:55:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T20:55:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ede2b442-76b0-443a-a450-252ee8711c80\\\",\\\"systemUUID\\\":\\\"72d081dd-fe7e-42a6-9cf0-f0ee7f733ac9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:45Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:45 crc kubenswrapper[4754]: I1005 20:55:45.523192 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:45 crc kubenswrapper[4754]: I1005 20:55:45.523305 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:45 crc kubenswrapper[4754]: I1005 20:55:45.523322 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:45 crc kubenswrapper[4754]: I1005 20:55:45.523346 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:45 crc kubenswrapper[4754]: I1005 20:55:45.523362 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:45Z","lastTransitionTime":"2025-10-05T20:55:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:45 crc kubenswrapper[4754]: E1005 20:55:45.539581 4754 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T20:55:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T20:55:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T20:55:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T20:55:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ede2b442-76b0-443a-a450-252ee8711c80\\\",\\\"systemUUID\\\":\\\"72d081dd-fe7e-42a6-9cf0-f0ee7f733ac9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:45Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:45 crc kubenswrapper[4754]: I1005 20:55:45.546211 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:45 crc kubenswrapper[4754]: I1005 20:55:45.546544 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:45 crc kubenswrapper[4754]: I1005 20:55:45.546750 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:45 crc kubenswrapper[4754]: I1005 20:55:45.546990 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:45 crc kubenswrapper[4754]: I1005 20:55:45.547189 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:45Z","lastTransitionTime":"2025-10-05T20:55:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:45 crc kubenswrapper[4754]: E1005 20:55:45.564824 4754 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T20:55:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T20:55:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T20:55:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T20:55:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ede2b442-76b0-443a-a450-252ee8711c80\\\",\\\"systemUUID\\\":\\\"72d081dd-fe7e-42a6-9cf0-f0ee7f733ac9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:45Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:45 crc kubenswrapper[4754]: E1005 20:55:45.565591 4754 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 05 20:55:45 crc kubenswrapper[4754]: I1005 20:55:45.568108 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:45 crc kubenswrapper[4754]: I1005 20:55:45.568169 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:45 crc kubenswrapper[4754]: I1005 20:55:45.568189 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:45 crc kubenswrapper[4754]: I1005 20:55:45.568216 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:45 crc kubenswrapper[4754]: I1005 20:55:45.568237 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:45Z","lastTransitionTime":"2025-10-05T20:55:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:45 crc kubenswrapper[4754]: I1005 20:55:45.606051 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7dc728ea-2601-44f0-bcab-2913f034007d-metrics-certs\") pod \"network-metrics-daemon-nwrnt\" (UID: \"7dc728ea-2601-44f0-bcab-2913f034007d\") " pod="openshift-multus/network-metrics-daemon-nwrnt" Oct 05 20:55:45 crc kubenswrapper[4754]: E1005 20:55:45.606231 4754 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 05 20:55:45 crc kubenswrapper[4754]: E1005 20:55:45.606305 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7dc728ea-2601-44f0-bcab-2913f034007d-metrics-certs podName:7dc728ea-2601-44f0-bcab-2913f034007d nodeName:}" failed. No retries permitted until 2025-10-05 20:56:17.606281381 +0000 UTC m=+101.510400101 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/7dc728ea-2601-44f0-bcab-2913f034007d-metrics-certs") pod "network-metrics-daemon-nwrnt" (UID: "7dc728ea-2601-44f0-bcab-2913f034007d") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 05 20:55:45 crc kubenswrapper[4754]: I1005 20:55:45.671971 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:45 crc kubenswrapper[4754]: I1005 20:55:45.672013 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:45 crc kubenswrapper[4754]: I1005 20:55:45.672022 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:45 crc kubenswrapper[4754]: I1005 20:55:45.672037 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:45 crc kubenswrapper[4754]: I1005 20:55:45.672051 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:45Z","lastTransitionTime":"2025-10-05T20:55:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:45 crc kubenswrapper[4754]: I1005 20:55:45.774712 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:45 crc kubenswrapper[4754]: I1005 20:55:45.774757 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:45 crc kubenswrapper[4754]: I1005 20:55:45.774770 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:45 crc kubenswrapper[4754]: I1005 20:55:45.774793 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:45 crc kubenswrapper[4754]: I1005 20:55:45.774805 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:45Z","lastTransitionTime":"2025-10-05T20:55:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:45 crc kubenswrapper[4754]: I1005 20:55:45.836540 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nwrnt" Oct 05 20:55:45 crc kubenswrapper[4754]: E1005 20:55:45.836701 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nwrnt" podUID="7dc728ea-2601-44f0-bcab-2913f034007d" Oct 05 20:55:45 crc kubenswrapper[4754]: I1005 20:55:45.877847 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:45 crc kubenswrapper[4754]: I1005 20:55:45.878069 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:45 crc kubenswrapper[4754]: I1005 20:55:45.878145 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:45 crc kubenswrapper[4754]: I1005 20:55:45.878209 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:45 crc kubenswrapper[4754]: I1005 20:55:45.878273 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:45Z","lastTransitionTime":"2025-10-05T20:55:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:45 crc kubenswrapper[4754]: I1005 20:55:45.982106 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:45 crc kubenswrapper[4754]: I1005 20:55:45.982175 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:45 crc kubenswrapper[4754]: I1005 20:55:45.982189 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:45 crc kubenswrapper[4754]: I1005 20:55:45.982218 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:45 crc kubenswrapper[4754]: I1005 20:55:45.982235 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:45Z","lastTransitionTime":"2025-10-05T20:55:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:46 crc kubenswrapper[4754]: I1005 20:55:46.085383 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:46 crc kubenswrapper[4754]: I1005 20:55:46.085423 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:46 crc kubenswrapper[4754]: I1005 20:55:46.085436 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:46 crc kubenswrapper[4754]: I1005 20:55:46.085457 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:46 crc kubenswrapper[4754]: I1005 20:55:46.085469 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:46Z","lastTransitionTime":"2025-10-05T20:55:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:46 crc kubenswrapper[4754]: I1005 20:55:46.189757 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:46 crc kubenswrapper[4754]: I1005 20:55:46.190113 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:46 crc kubenswrapper[4754]: I1005 20:55:46.190222 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:46 crc kubenswrapper[4754]: I1005 20:55:46.190337 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:46 crc kubenswrapper[4754]: I1005 20:55:46.190425 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:46Z","lastTransitionTime":"2025-10-05T20:55:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:46 crc kubenswrapper[4754]: I1005 20:55:46.294526 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:46 crc kubenswrapper[4754]: I1005 20:55:46.294557 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:46 crc kubenswrapper[4754]: I1005 20:55:46.294566 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:46 crc kubenswrapper[4754]: I1005 20:55:46.294584 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:46 crc kubenswrapper[4754]: I1005 20:55:46.294596 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:46Z","lastTransitionTime":"2025-10-05T20:55:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:46 crc kubenswrapper[4754]: I1005 20:55:46.397181 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:46 crc kubenswrapper[4754]: I1005 20:55:46.397556 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:46 crc kubenswrapper[4754]: I1005 20:55:46.397697 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:46 crc kubenswrapper[4754]: I1005 20:55:46.397860 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:46 crc kubenswrapper[4754]: I1005 20:55:46.397992 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:46Z","lastTransitionTime":"2025-10-05T20:55:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:46 crc kubenswrapper[4754]: I1005 20:55:46.501280 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:46 crc kubenswrapper[4754]: I1005 20:55:46.501320 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:46 crc kubenswrapper[4754]: I1005 20:55:46.501339 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:46 crc kubenswrapper[4754]: I1005 20:55:46.501363 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:46 crc kubenswrapper[4754]: I1005 20:55:46.501383 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:46Z","lastTransitionTime":"2025-10-05T20:55:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:46 crc kubenswrapper[4754]: I1005 20:55:46.604482 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:46 crc kubenswrapper[4754]: I1005 20:55:46.604577 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:46 crc kubenswrapper[4754]: I1005 20:55:46.604596 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:46 crc kubenswrapper[4754]: I1005 20:55:46.604627 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:46 crc kubenswrapper[4754]: I1005 20:55:46.604651 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:46Z","lastTransitionTime":"2025-10-05T20:55:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:46 crc kubenswrapper[4754]: I1005 20:55:46.708038 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:46 crc kubenswrapper[4754]: I1005 20:55:46.708094 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:46 crc kubenswrapper[4754]: I1005 20:55:46.708112 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:46 crc kubenswrapper[4754]: I1005 20:55:46.708137 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:46 crc kubenswrapper[4754]: I1005 20:55:46.708156 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:46Z","lastTransitionTime":"2025-10-05T20:55:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:46 crc kubenswrapper[4754]: I1005 20:55:46.811420 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:46 crc kubenswrapper[4754]: I1005 20:55:46.811480 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:46 crc kubenswrapper[4754]: I1005 20:55:46.811526 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:46 crc kubenswrapper[4754]: I1005 20:55:46.811552 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:46 crc kubenswrapper[4754]: I1005 20:55:46.811570 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:46Z","lastTransitionTime":"2025-10-05T20:55:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:46 crc kubenswrapper[4754]: I1005 20:55:46.837315 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 20:55:46 crc kubenswrapper[4754]: I1005 20:55:46.837363 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 20:55:46 crc kubenswrapper[4754]: E1005 20:55:46.837866 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 20:55:46 crc kubenswrapper[4754]: E1005 20:55:46.837922 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 20:55:46 crc kubenswrapper[4754]: I1005 20:55:46.837394 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 20:55:46 crc kubenswrapper[4754]: E1005 20:55:46.838328 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 20:55:46 crc kubenswrapper[4754]: I1005 20:55:46.854454 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-nwrnt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7dc728ea-2601-44f0-bcab-2913f034007d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-846xq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-846xq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:55:13Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-nwrnt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:46Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:46 crc kubenswrapper[4754]: I1005 20:55:46.874344 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"934bcd66-9c03-48b9-82dd-5093338f9c8f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9042df2f43243d3d79ba9a3940f8dd4f6c7a89ef9e3510e81fe60b1944d7e83b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5366bbf46a91667355f43f5df7335a536b395cda788b90b35ef18e2b53bc4c02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://600bc3bb551b747cf2f18e26745883232d7a37ac5877fad0c67100e52d7a43aa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78f74982180048eb50b9be79fa72c04f62f925b54fe511920f37e6407f864631\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:46Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:46 crc kubenswrapper[4754]: I1005 20:55:46.892240 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:46Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:46 crc kubenswrapper[4754]: I1005 20:55:46.905145 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9ad070bb699b643d29b82bcc1126ce3b585f7492f5ae2817fa491b309b1691d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:46Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:46 crc kubenswrapper[4754]: I1005 20:55:46.914677 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:46 crc kubenswrapper[4754]: I1005 20:55:46.914709 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:46 crc kubenswrapper[4754]: I1005 20:55:46.914738 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:46 crc kubenswrapper[4754]: I1005 20:55:46.914756 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:46 crc kubenswrapper[4754]: I1005 20:55:46.914769 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:46Z","lastTransitionTime":"2025-10-05T20:55:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:46 crc kubenswrapper[4754]: I1005 20:55:46.917872 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-57sr4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e854c7b9-1229-4da4-80e9-fea6479459b8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceb577e1055d84007936a50d0b3d54b9ae0299a1fc1a6618b37163a6d3fbb9cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vw2ps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:58Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-57sr4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:46Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:46 crc kubenswrapper[4754]: I1005 20:55:46.935846 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cn76s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02d5e3f9-73c2-4496-9aca-0787184aef19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7ba2b076839f4a5f586d2e6c4f4d7188254128b782bfd2180e22bf3ea5ffdd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpx5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cn76s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:46Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:46 crc kubenswrapper[4754]: I1005 20:55:46.952959 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ss7lm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f9ef817-0622-4185-a7fb-57221690aa75\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ba613b0d7753a003c572c3733ed394707c45b6f8072cb7aa88f3c7312b8882b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wpwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:55:02Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ss7lm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:46Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:46 crc kubenswrapper[4754]: I1005 20:55:46.976582 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7506d496-65ff-41dc-8230-24c8ee653a2a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33da5a6bb3cd3b042c193bc571e6622f2f87f27e560cdd0bbb9be31e2010d707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df6d700a947d19c84b4f65fd2dd2f8d1f4f1f80fa9b4800c2acfa8dae3912639\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35ca8ee77a0e58cf969bd67b277a3392614c4e73b963e6f6002097709370715d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e32c2154f46ba10c509fe40c357304aa8e82d04e9906bbaf2ccecd47b42279e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8907b9b69b249c34ec6db1e460d641b31711b75d6d31d4caafd867339e6571d0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1005 20:54:52.589115 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 20:54:52.590181 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1304624737/tls.crt::/tmp/serving-cert-1304624737/tls.key\\\\\\\"\\\\nI1005 20:54:58.070544 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1005 20:54:58.090352 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1005 20:54:58.090382 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1005 20:54:58.090412 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1005 20:54:58.090418 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1005 20:54:58.144656 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1005 20:54:58.144688 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 20:54:58.144695 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 20:54:58.144699 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1005 20:54:58.144702 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1005 20:54:58.144705 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1005 20:54:58.144708 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1005 20:54:58.145024 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1005 20:54:58.148112 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4efb34f3493cc8854bc4cff6b29f63973c52b754404d7a21daec4e1f612d327d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54c568b637f3737667433047b956c7eb22fdf0f9301acdfd3b49fb84c39de6ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54c568b637f3737667433047b956c7eb22fdf0f9301acdfd3b49fb84c39de6ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:46Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:46 crc kubenswrapper[4754]: I1005 20:55:46.998687 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c76e6e4740035a6ca54191a770c4516e28101aae69afbb89739174c2f076ef5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9eededc0cb1af94fb7b5daba08785f6b077a85e9cb5ddd8b9f397b658194f61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:46Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:47 crc kubenswrapper[4754]: I1005 20:55:47.017892 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:47 crc kubenswrapper[4754]: I1005 20:55:47.017943 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:47 crc kubenswrapper[4754]: I1005 20:55:47.017957 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:47 crc kubenswrapper[4754]: I1005 20:55:47.017977 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:47 crc kubenswrapper[4754]: I1005 20:55:47.017991 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:47Z","lastTransitionTime":"2025-10-05T20:55:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:47 crc kubenswrapper[4754]: I1005 20:55:47.022065 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tskt7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a68e9c89-465b-4c01-a470-e21a480ee465\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba6edba14cf321b4cd2f7bb44067a3ecf49f1594e76c0197e807ff3daa45e9b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d14ec99610d2167b484bd7712af7bb6d993d26c74b2828da1c89314f16fda22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d14ec99610d2167b484bd7712af7bb6d993d26c74b2828da1c89314f16fda22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f6f08dd853b25881e2f3c66206aa80d0c493518284217780aa40ac4b69f0c0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f6f08dd853b25881e2f3c66206aa80d0c493518284217780aa40ac4b69f0c0b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://71026d00ec1a3d91b019deb3c9f43ee52fb05b8f2d6374e1a5b762c9e5c1b10d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71026d00ec1a3d91b019deb3c9f43ee52fb05b8f2d6374e1a5b762c9e5c1b10d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b9910d2470acb16ebb436d7030b3161061607cb2a5ef50d482d8d5c4f48a8ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b9910d2470acb16ebb436d7030b3161061607cb2a5ef50d482d8d5c4f48a8ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0a93b61eacc55111995e69c15f89a210b8aa15468ac9ac053b7cf6b007d301e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b0a93b61eacc55111995e69c15f89a210b8aa15468ac9ac053b7cf6b007d301e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cafb3208e7e3bcdd01898a68fc5bbf4cb52ca0800b879b77a60bb102794257f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cafb3208e7e3bcdd01898a68fc5bbf4cb52ca0800b879b77a60bb102794257f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tskt7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:47Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:47 crc kubenswrapper[4754]: I1005 20:55:47.040986 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be95a413-6a59-45b4-84b7-b43cae694a26\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10b211cf6e47dbd1ee42c82e49de0a13a45581b92a0c4f6bbba0326c7149e41d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2ggg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4bb8c5903e8b70dd19eebb111097d96f725e17de44e06c3e74c067376f6c7360\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2ggg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-b2h9k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:47Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:47 crc kubenswrapper[4754]: I1005 20:55:47.060161 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rddkl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"15afc2a1-5cb7-41c6-954b-446c36822f50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://866d85ee126f8641d12f1a05fabc3de7b9392c324e9fe06ea92214155c05334c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqv8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a175565f83cfcba7825006eafb0f83efb70ec3d415c2c039c6499d5f6576e89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqv8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:55:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rddkl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:47Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:47 crc kubenswrapper[4754]: I1005 20:55:47.076204 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ae81d5f-5e31-4648-a80c-001034f0acf8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02dec08702aeaa03bc26151781909a4c6226fdaf37b921feb2a53ad6fa3c94c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a0c1914423693dd5ca96eddab941162a0f1e9efaf4bd02c47514f092e39e9e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5d1820b7c2ba7c5262f52297111f58e8f05771f3d7e34a59f8f3a05a7fb0bfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c187ac7d3514f4e68984ee98c274c53418df1818159cc9d6c16626d162fa3f97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c187ac7d3514f4e68984ee98c274c53418df1818159cc9d6c16626d162fa3f97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:36Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:47Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:47 crc kubenswrapper[4754]: I1005 20:55:47.095285 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e380f9b2a0e319dc325c54d1247a6db98346292548fa688cb41a2ca1b3048419\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:47Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:47 crc kubenswrapper[4754]: I1005 20:55:47.111843 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:47Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:47 crc kubenswrapper[4754]: I1005 20:55:47.120951 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:47 crc kubenswrapper[4754]: I1005 20:55:47.120992 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:47 crc kubenswrapper[4754]: I1005 20:55:47.121009 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:47 crc kubenswrapper[4754]: I1005 20:55:47.121034 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:47 crc kubenswrapper[4754]: I1005 20:55:47.121053 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:47Z","lastTransitionTime":"2025-10-05T20:55:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:47 crc kubenswrapper[4754]: I1005 20:55:47.141947 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9309179-b92a-4fa4-9fba-747b8f911936\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d969d3d50d6ad280fc6654df9441bab238bb538f277ed6d1273779813ffbaaed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://800a731ed8ef57714b2318af14054ef23f0368936630887a093fc77167b33d12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8cd9b08e83dc174413cc37891c14562816787028471a4201680c86dd620b7a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93b729f3f7f993db0d6bc3ba60c04930aada3cef8272ffcc3c6a8a242a9b1475\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dde1519111fc3af5d14463fdb58d4f5c62a7045e7de98c60981c7e4da37b0e7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ea160a5b37aea7f2970ae4dd2614b09d3dcf5dcbd1fc103b9997f8ecc3a804c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ea160a5b37aea7f2970ae4dd2614b09d3dcf5dcbd1fc103b9997f8ecc3a804c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://17ae6aa6c53e7dd061f70b41428ef50bbc7ffcc65352fad1a4a9d3c69dbd2ba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17ae6aa6c53e7dd061f70b41428ef50bbc7ffcc65352fad1a4a9d3c69dbd2ba9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://ad4cd69df7608c7d87d839fe09d4557c674c7efbc4730f4e4d3daaf9a26f5aee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad4cd69df7608c7d87d839fe09d4557c674c7efbc4730f4e4d3daaf9a26f5aee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:36Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:47Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:47 crc kubenswrapper[4754]: I1005 20:55:47.160655 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:47Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:47 crc kubenswrapper[4754]: I1005 20:55:47.185884 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b44790fb-fecb-4ec3-9816-8e23dd72ab33\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a7dc53e7f96e48213d724a2cb4859514df925005644108ae3a0c33b14dcf52d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfdd8b1e542ab676576c311e2923fb02861f92de91fe1f4b3e4b968c549c375d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3a63451ffaa14cacc3b78d11340f4ea81445fa3af96178e9b172d063d9697d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3cfd3f2274a8a46fa36660e94ac6c9c298c791283f12d52330c5594ac131998\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f626cb108716da0b595844ea0251b7a04d49b5908a1d6dd0aef5cb2d402c4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://822223f515b5add6d71d193fd5bc77df029b8d37a1cc07be4145b53c7226bf61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af1c313e0dc0fe3b48feac3cce1ecd4452c9efadf3d19f257c07cc7fcdec875f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af1c313e0dc0fe3b48feac3cce1ecd4452c9efadf3d19f257c07cc7fcdec875f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-05T20:55:26Z\\\",\\\"message\\\":\\\"5:26.770640 6339 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/iptables-alerter-4ln5h in node crc\\\\nI1005 20:55:26.770478 6339 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-apiserver-operator/metrics]} name:Service_openshift-apiserver-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.38:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {8b82f026-5975-4a1b-bb18-08d5d51147ec}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1005 20:55:26.770652 6339 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/iptables-alerter-4ln5h after 0 failed attempt(s)\\\\nI1005 20:55:26.770662 6339 default_network_controller.go:776] Recording success event on pod openshift-network-operator/iptables-alerter-4ln5h\\\\nF1005 20:55:26.770677 6339 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controlle\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:25Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-7rhps_openshift-ovn-kubernetes(b44790fb-fecb-4ec3-9816-8e23dd72ab33)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2acf9637fa25e5d3ffeb05224c11218c47defc679568fb928a498869613de4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35ebeb7a87d5d47b6567917d38eca38a3e622c497fba58971296c4609c287423\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35ebeb7a87d5d47b6567917d38eca38a3e622c497fba58971296c4609c287423\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7rhps\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:47Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:47 crc kubenswrapper[4754]: I1005 20:55:47.226988 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:47 crc kubenswrapper[4754]: I1005 20:55:47.227042 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:47 crc kubenswrapper[4754]: I1005 20:55:47.227114 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:47 crc kubenswrapper[4754]: I1005 20:55:47.227144 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:47 crc kubenswrapper[4754]: I1005 20:55:47.227155 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:47Z","lastTransitionTime":"2025-10-05T20:55:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:47 crc kubenswrapper[4754]: I1005 20:55:47.332379 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:47 crc kubenswrapper[4754]: I1005 20:55:47.332467 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:47 crc kubenswrapper[4754]: I1005 20:55:47.332490 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:47 crc kubenswrapper[4754]: I1005 20:55:47.332560 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:47 crc kubenswrapper[4754]: I1005 20:55:47.332581 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:47Z","lastTransitionTime":"2025-10-05T20:55:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:47 crc kubenswrapper[4754]: I1005 20:55:47.399827 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-cn76s_02d5e3f9-73c2-4496-9aca-0787184aef19/kube-multus/0.log" Oct 05 20:55:47 crc kubenswrapper[4754]: I1005 20:55:47.399882 4754 generic.go:334] "Generic (PLEG): container finished" podID="02d5e3f9-73c2-4496-9aca-0787184aef19" containerID="f7ba2b076839f4a5f586d2e6c4f4d7188254128b782bfd2180e22bf3ea5ffdd7" exitCode=1 Oct 05 20:55:47 crc kubenswrapper[4754]: I1005 20:55:47.399921 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-cn76s" event={"ID":"02d5e3f9-73c2-4496-9aca-0787184aef19","Type":"ContainerDied","Data":"f7ba2b076839f4a5f586d2e6c4f4d7188254128b782bfd2180e22bf3ea5ffdd7"} Oct 05 20:55:47 crc kubenswrapper[4754]: I1005 20:55:47.400435 4754 scope.go:117] "RemoveContainer" containerID="f7ba2b076839f4a5f586d2e6c4f4d7188254128b782bfd2180e22bf3ea5ffdd7" Oct 05 20:55:47 crc kubenswrapper[4754]: I1005 20:55:47.418435 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7506d496-65ff-41dc-8230-24c8ee653a2a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33da5a6bb3cd3b042c193bc571e6622f2f87f27e560cdd0bbb9be31e2010d707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df6d700a947d19c84b4f65fd2dd2f8d1f4f1f80fa9b4800c2acfa8dae3912639\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35ca8ee77a0e58cf969bd67b277a3392614c4e73b963e6f6002097709370715d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e32c2154f46ba10c509fe40c357304aa8e82d04e9906bbaf2ccecd47b42279e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8907b9b69b249c34ec6db1e460d641b31711b75d6d31d4caafd867339e6571d0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1005 20:54:52.589115 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 20:54:52.590181 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1304624737/tls.crt::/tmp/serving-cert-1304624737/tls.key\\\\\\\"\\\\nI1005 20:54:58.070544 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1005 20:54:58.090352 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1005 20:54:58.090382 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1005 20:54:58.090412 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1005 20:54:58.090418 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1005 20:54:58.144656 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1005 20:54:58.144688 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 20:54:58.144695 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 20:54:58.144699 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1005 20:54:58.144702 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1005 20:54:58.144705 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1005 20:54:58.144708 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1005 20:54:58.145024 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1005 20:54:58.148112 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4efb34f3493cc8854bc4cff6b29f63973c52b754404d7a21daec4e1f612d327d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54c568b637f3737667433047b956c7eb22fdf0f9301acdfd3b49fb84c39de6ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54c568b637f3737667433047b956c7eb22fdf0f9301acdfd3b49fb84c39de6ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:47Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:47 crc kubenswrapper[4754]: I1005 20:55:47.429425 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c76e6e4740035a6ca54191a770c4516e28101aae69afbb89739174c2f076ef5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9eededc0cb1af94fb7b5daba08785f6b077a85e9cb5ddd8b9f397b658194f61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:47Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:47 crc kubenswrapper[4754]: I1005 20:55:47.436290 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:47 crc kubenswrapper[4754]: I1005 20:55:47.436334 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:47 crc kubenswrapper[4754]: I1005 20:55:47.436349 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:47 crc kubenswrapper[4754]: I1005 20:55:47.436373 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:47 crc kubenswrapper[4754]: I1005 20:55:47.436386 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:47Z","lastTransitionTime":"2025-10-05T20:55:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:47 crc kubenswrapper[4754]: I1005 20:55:47.446361 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tskt7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a68e9c89-465b-4c01-a470-e21a480ee465\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba6edba14cf321b4cd2f7bb44067a3ecf49f1594e76c0197e807ff3daa45e9b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d14ec99610d2167b484bd7712af7bb6d993d26c74b2828da1c89314f16fda22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d14ec99610d2167b484bd7712af7bb6d993d26c74b2828da1c89314f16fda22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f6f08dd853b25881e2f3c66206aa80d0c493518284217780aa40ac4b69f0c0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f6f08dd853b25881e2f3c66206aa80d0c493518284217780aa40ac4b69f0c0b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://71026d00ec1a3d91b019deb3c9f43ee52fb05b8f2d6374e1a5b762c9e5c1b10d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71026d00ec1a3d91b019deb3c9f43ee52fb05b8f2d6374e1a5b762c9e5c1b10d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b9910d2470acb16ebb436d7030b3161061607cb2a5ef50d482d8d5c4f48a8ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b9910d2470acb16ebb436d7030b3161061607cb2a5ef50d482d8d5c4f48a8ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0a93b61eacc55111995e69c15f89a210b8aa15468ac9ac053b7cf6b007d301e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b0a93b61eacc55111995e69c15f89a210b8aa15468ac9ac053b7cf6b007d301e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cafb3208e7e3bcdd01898a68fc5bbf4cb52ca0800b879b77a60bb102794257f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cafb3208e7e3bcdd01898a68fc5bbf4cb52ca0800b879b77a60bb102794257f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tskt7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:47Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:47 crc kubenswrapper[4754]: I1005 20:55:47.459310 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be95a413-6a59-45b4-84b7-b43cae694a26\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10b211cf6e47dbd1ee42c82e49de0a13a45581b92a0c4f6bbba0326c7149e41d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2ggg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4bb8c5903e8b70dd19eebb111097d96f725e17de44e06c3e74c067376f6c7360\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2ggg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-b2h9k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:47Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:47 crc kubenswrapper[4754]: I1005 20:55:47.472357 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rddkl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"15afc2a1-5cb7-41c6-954b-446c36822f50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://866d85ee126f8641d12f1a05fabc3de7b9392c324e9fe06ea92214155c05334c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqv8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a175565f83cfcba7825006eafb0f83efb70ec3d415c2c039c6499d5f6576e89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqv8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:55:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rddkl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:47Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:47 crc kubenswrapper[4754]: I1005 20:55:47.484436 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ae81d5f-5e31-4648-a80c-001034f0acf8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02dec08702aeaa03bc26151781909a4c6226fdaf37b921feb2a53ad6fa3c94c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a0c1914423693dd5ca96eddab941162a0f1e9efaf4bd02c47514f092e39e9e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5d1820b7c2ba7c5262f52297111f58e8f05771f3d7e34a59f8f3a05a7fb0bfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c187ac7d3514f4e68984ee98c274c53418df1818159cc9d6c16626d162fa3f97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c187ac7d3514f4e68984ee98c274c53418df1818159cc9d6c16626d162fa3f97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:36Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:47Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:47 crc kubenswrapper[4754]: I1005 20:55:47.501969 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e380f9b2a0e319dc325c54d1247a6db98346292548fa688cb41a2ca1b3048419\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:47Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:47 crc kubenswrapper[4754]: I1005 20:55:47.517294 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:47Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:47 crc kubenswrapper[4754]: I1005 20:55:47.539000 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:47 crc kubenswrapper[4754]: I1005 20:55:47.539030 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:47 crc kubenswrapper[4754]: I1005 20:55:47.539040 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:47 crc kubenswrapper[4754]: I1005 20:55:47.539059 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:47 crc kubenswrapper[4754]: I1005 20:55:47.539073 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:47Z","lastTransitionTime":"2025-10-05T20:55:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:47 crc kubenswrapper[4754]: I1005 20:55:47.540150 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9309179-b92a-4fa4-9fba-747b8f911936\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d969d3d50d6ad280fc6654df9441bab238bb538f277ed6d1273779813ffbaaed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://800a731ed8ef57714b2318af14054ef23f0368936630887a093fc77167b33d12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8cd9b08e83dc174413cc37891c14562816787028471a4201680c86dd620b7a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93b729f3f7f993db0d6bc3ba60c04930aada3cef8272ffcc3c6a8a242a9b1475\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dde1519111fc3af5d14463fdb58d4f5c62a7045e7de98c60981c7e4da37b0e7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ea160a5b37aea7f2970ae4dd2614b09d3dcf5dcbd1fc103b9997f8ecc3a804c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ea160a5b37aea7f2970ae4dd2614b09d3dcf5dcbd1fc103b9997f8ecc3a804c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://17ae6aa6c53e7dd061f70b41428ef50bbc7ffcc65352fad1a4a9d3c69dbd2ba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17ae6aa6c53e7dd061f70b41428ef50bbc7ffcc65352fad1a4a9d3c69dbd2ba9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://ad4cd69df7608c7d87d839fe09d4557c674c7efbc4730f4e4d3daaf9a26f5aee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad4cd69df7608c7d87d839fe09d4557c674c7efbc4730f4e4d3daaf9a26f5aee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:36Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:47Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:47 crc kubenswrapper[4754]: I1005 20:55:47.555390 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:47Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:47 crc kubenswrapper[4754]: I1005 20:55:47.575596 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b44790fb-fecb-4ec3-9816-8e23dd72ab33\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a7dc53e7f96e48213d724a2cb4859514df925005644108ae3a0c33b14dcf52d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfdd8b1e542ab676576c311e2923fb02861f92de91fe1f4b3e4b968c549c375d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3a63451ffaa14cacc3b78d11340f4ea81445fa3af96178e9b172d063d9697d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3cfd3f2274a8a46fa36660e94ac6c9c298c791283f12d52330c5594ac131998\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f626cb108716da0b595844ea0251b7a04d49b5908a1d6dd0aef5cb2d402c4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://822223f515b5add6d71d193fd5bc77df029b8d37a1cc07be4145b53c7226bf61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af1c313e0dc0fe3b48feac3cce1ecd4452c9efadf3d19f257c07cc7fcdec875f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af1c313e0dc0fe3b48feac3cce1ecd4452c9efadf3d19f257c07cc7fcdec875f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-05T20:55:26Z\\\",\\\"message\\\":\\\"5:26.770640 6339 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/iptables-alerter-4ln5h in node crc\\\\nI1005 20:55:26.770478 6339 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-apiserver-operator/metrics]} name:Service_openshift-apiserver-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.38:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {8b82f026-5975-4a1b-bb18-08d5d51147ec}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1005 20:55:26.770652 6339 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/iptables-alerter-4ln5h after 0 failed attempt(s)\\\\nI1005 20:55:26.770662 6339 default_network_controller.go:776] Recording success event on pod openshift-network-operator/iptables-alerter-4ln5h\\\\nF1005 20:55:26.770677 6339 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controlle\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:25Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-7rhps_openshift-ovn-kubernetes(b44790fb-fecb-4ec3-9816-8e23dd72ab33)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2acf9637fa25e5d3ffeb05224c11218c47defc679568fb928a498869613de4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35ebeb7a87d5d47b6567917d38eca38a3e622c497fba58971296c4609c287423\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35ebeb7a87d5d47b6567917d38eca38a3e622c497fba58971296c4609c287423\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7rhps\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:47Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:47 crc kubenswrapper[4754]: I1005 20:55:47.590124 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"934bcd66-9c03-48b9-82dd-5093338f9c8f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9042df2f43243d3d79ba9a3940f8dd4f6c7a89ef9e3510e81fe60b1944d7e83b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5366bbf46a91667355f43f5df7335a536b395cda788b90b35ef18e2b53bc4c02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://600bc3bb551b747cf2f18e26745883232d7a37ac5877fad0c67100e52d7a43aa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78f74982180048eb50b9be79fa72c04f62f925b54fe511920f37e6407f864631\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:47Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:47 crc kubenswrapper[4754]: I1005 20:55:47.603690 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:47Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:47 crc kubenswrapper[4754]: I1005 20:55:47.618405 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9ad070bb699b643d29b82bcc1126ce3b585f7492f5ae2817fa491b309b1691d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:47Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:47 crc kubenswrapper[4754]: I1005 20:55:47.633107 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-57sr4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e854c7b9-1229-4da4-80e9-fea6479459b8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceb577e1055d84007936a50d0b3d54b9ae0299a1fc1a6618b37163a6d3fbb9cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vw2ps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:58Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-57sr4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:47Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:47 crc kubenswrapper[4754]: I1005 20:55:47.643899 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:47 crc kubenswrapper[4754]: I1005 20:55:47.643935 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:47 crc kubenswrapper[4754]: I1005 20:55:47.643944 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:47 crc kubenswrapper[4754]: I1005 20:55:47.643961 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:47 crc kubenswrapper[4754]: I1005 20:55:47.643975 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:47Z","lastTransitionTime":"2025-10-05T20:55:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:47 crc kubenswrapper[4754]: I1005 20:55:47.646921 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cn76s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02d5e3f9-73c2-4496-9aca-0787184aef19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7ba2b076839f4a5f586d2e6c4f4d7188254128b782bfd2180e22bf3ea5ffdd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7ba2b076839f4a5f586d2e6c4f4d7188254128b782bfd2180e22bf3ea5ffdd7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-05T20:55:46Z\\\",\\\"message\\\":\\\"2025-10-05T20:55:01+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_d9d50ba4-5e54-4920-83d9-a3faa08e80be\\\\n2025-10-05T20:55:01+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_d9d50ba4-5e54-4920-83d9-a3faa08e80be to /host/opt/cni/bin/\\\\n2025-10-05T20:55:01Z [verbose] multus-daemon started\\\\n2025-10-05T20:55:01Z [verbose] Readiness Indicator file check\\\\n2025-10-05T20:55:46Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpx5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cn76s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:47Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:47 crc kubenswrapper[4754]: I1005 20:55:47.657741 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ss7lm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f9ef817-0622-4185-a7fb-57221690aa75\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ba613b0d7753a003c572c3733ed394707c45b6f8072cb7aa88f3c7312b8882b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wpwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:55:02Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ss7lm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:47Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:47 crc kubenswrapper[4754]: I1005 20:55:47.668390 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-nwrnt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7dc728ea-2601-44f0-bcab-2913f034007d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-846xq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-846xq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:55:13Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-nwrnt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:47Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:47 crc kubenswrapper[4754]: I1005 20:55:47.746579 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:47 crc kubenswrapper[4754]: I1005 20:55:47.746627 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:47 crc kubenswrapper[4754]: I1005 20:55:47.746639 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:47 crc kubenswrapper[4754]: I1005 20:55:47.746658 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:47 crc kubenswrapper[4754]: I1005 20:55:47.746671 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:47Z","lastTransitionTime":"2025-10-05T20:55:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:47 crc kubenswrapper[4754]: I1005 20:55:47.836603 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nwrnt" Oct 05 20:55:47 crc kubenswrapper[4754]: E1005 20:55:47.836745 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nwrnt" podUID="7dc728ea-2601-44f0-bcab-2913f034007d" Oct 05 20:55:47 crc kubenswrapper[4754]: I1005 20:55:47.849937 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:47 crc kubenswrapper[4754]: I1005 20:55:47.850015 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:47 crc kubenswrapper[4754]: I1005 20:55:47.850039 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:47 crc kubenswrapper[4754]: I1005 20:55:47.850074 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:47 crc kubenswrapper[4754]: I1005 20:55:47.850108 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:47Z","lastTransitionTime":"2025-10-05T20:55:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:47 crc kubenswrapper[4754]: I1005 20:55:47.953635 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:47 crc kubenswrapper[4754]: I1005 20:55:47.953688 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:47 crc kubenswrapper[4754]: I1005 20:55:47.953706 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:47 crc kubenswrapper[4754]: I1005 20:55:47.953729 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:47 crc kubenswrapper[4754]: I1005 20:55:47.953748 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:47Z","lastTransitionTime":"2025-10-05T20:55:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:48 crc kubenswrapper[4754]: I1005 20:55:48.056073 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:48 crc kubenswrapper[4754]: I1005 20:55:48.056111 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:48 crc kubenswrapper[4754]: I1005 20:55:48.056124 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:48 crc kubenswrapper[4754]: I1005 20:55:48.056143 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:48 crc kubenswrapper[4754]: I1005 20:55:48.056156 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:48Z","lastTransitionTime":"2025-10-05T20:55:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:48 crc kubenswrapper[4754]: I1005 20:55:48.159379 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:48 crc kubenswrapper[4754]: I1005 20:55:48.159434 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:48 crc kubenswrapper[4754]: I1005 20:55:48.159464 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:48 crc kubenswrapper[4754]: I1005 20:55:48.159482 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:48 crc kubenswrapper[4754]: I1005 20:55:48.159509 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:48Z","lastTransitionTime":"2025-10-05T20:55:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:48 crc kubenswrapper[4754]: I1005 20:55:48.264170 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:48 crc kubenswrapper[4754]: I1005 20:55:48.264212 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:48 crc kubenswrapper[4754]: I1005 20:55:48.264224 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:48 crc kubenswrapper[4754]: I1005 20:55:48.264240 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:48 crc kubenswrapper[4754]: I1005 20:55:48.264252 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:48Z","lastTransitionTime":"2025-10-05T20:55:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:48 crc kubenswrapper[4754]: I1005 20:55:48.367977 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:48 crc kubenswrapper[4754]: I1005 20:55:48.368045 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:48 crc kubenswrapper[4754]: I1005 20:55:48.368056 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:48 crc kubenswrapper[4754]: I1005 20:55:48.368077 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:48 crc kubenswrapper[4754]: I1005 20:55:48.368092 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:48Z","lastTransitionTime":"2025-10-05T20:55:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:48 crc kubenswrapper[4754]: I1005 20:55:48.412180 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-cn76s_02d5e3f9-73c2-4496-9aca-0787184aef19/kube-multus/0.log" Oct 05 20:55:48 crc kubenswrapper[4754]: I1005 20:55:48.412264 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-cn76s" event={"ID":"02d5e3f9-73c2-4496-9aca-0787184aef19","Type":"ContainerStarted","Data":"5ed23b9cd9b57bb169a8ebbf16b7ff8d196eb14330df0edf2b9fcd2a8c9e2c88"} Oct 05 20:55:48 crc kubenswrapper[4754]: I1005 20:55:48.435643 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e380f9b2a0e319dc325c54d1247a6db98346292548fa688cb41a2ca1b3048419\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:48Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:48 crc kubenswrapper[4754]: I1005 20:55:48.451191 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:48Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:48 crc kubenswrapper[4754]: I1005 20:55:48.468945 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ae81d5f-5e31-4648-a80c-001034f0acf8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02dec08702aeaa03bc26151781909a4c6226fdaf37b921feb2a53ad6fa3c94c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a0c1914423693dd5ca96eddab941162a0f1e9efaf4bd02c47514f092e39e9e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5d1820b7c2ba7c5262f52297111f58e8f05771f3d7e34a59f8f3a05a7fb0bfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c187ac7d3514f4e68984ee98c274c53418df1818159cc9d6c16626d162fa3f97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c187ac7d3514f4e68984ee98c274c53418df1818159cc9d6c16626d162fa3f97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:36Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:48Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:48 crc kubenswrapper[4754]: I1005 20:55:48.472932 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:48 crc kubenswrapper[4754]: I1005 20:55:48.472981 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:48 crc kubenswrapper[4754]: I1005 20:55:48.473003 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:48 crc kubenswrapper[4754]: I1005 20:55:48.473032 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:48 crc kubenswrapper[4754]: I1005 20:55:48.473054 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:48Z","lastTransitionTime":"2025-10-05T20:55:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:48 crc kubenswrapper[4754]: I1005 20:55:48.491675 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b44790fb-fecb-4ec3-9816-8e23dd72ab33\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a7dc53e7f96e48213d724a2cb4859514df925005644108ae3a0c33b14dcf52d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfdd8b1e542ab676576c311e2923fb02861f92de91fe1f4b3e4b968c549c375d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3a63451ffaa14cacc3b78d11340f4ea81445fa3af96178e9b172d063d9697d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3cfd3f2274a8a46fa36660e94ac6c9c298c791283f12d52330c5594ac131998\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f626cb108716da0b595844ea0251b7a04d49b5908a1d6dd0aef5cb2d402c4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://822223f515b5add6d71d193fd5bc77df029b8d37a1cc07be4145b53c7226bf61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af1c313e0dc0fe3b48feac3cce1ecd4452c9efadf3d19f257c07cc7fcdec875f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af1c313e0dc0fe3b48feac3cce1ecd4452c9efadf3d19f257c07cc7fcdec875f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-05T20:55:26Z\\\",\\\"message\\\":\\\"5:26.770640 6339 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/iptables-alerter-4ln5h in node crc\\\\nI1005 20:55:26.770478 6339 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-apiserver-operator/metrics]} name:Service_openshift-apiserver-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.38:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {8b82f026-5975-4a1b-bb18-08d5d51147ec}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1005 20:55:26.770652 6339 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/iptables-alerter-4ln5h after 0 failed attempt(s)\\\\nI1005 20:55:26.770662 6339 default_network_controller.go:776] Recording success event on pod openshift-network-operator/iptables-alerter-4ln5h\\\\nF1005 20:55:26.770677 6339 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controlle\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:25Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-7rhps_openshift-ovn-kubernetes(b44790fb-fecb-4ec3-9816-8e23dd72ab33)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2acf9637fa25e5d3ffeb05224c11218c47defc679568fb928a498869613de4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35ebeb7a87d5d47b6567917d38eca38a3e622c497fba58971296c4609c287423\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35ebeb7a87d5d47b6567917d38eca38a3e622c497fba58971296c4609c287423\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7rhps\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:48Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:48 crc kubenswrapper[4754]: I1005 20:55:48.513540 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9309179-b92a-4fa4-9fba-747b8f911936\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d969d3d50d6ad280fc6654df9441bab238bb538f277ed6d1273779813ffbaaed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://800a731ed8ef57714b2318af14054ef23f0368936630887a093fc77167b33d12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8cd9b08e83dc174413cc37891c14562816787028471a4201680c86dd620b7a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93b729f3f7f993db0d6bc3ba60c04930aada3cef8272ffcc3c6a8a242a9b1475\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dde1519111fc3af5d14463fdb58d4f5c62a7045e7de98c60981c7e4da37b0e7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ea160a5b37aea7f2970ae4dd2614b09d3dcf5dcbd1fc103b9997f8ecc3a804c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ea160a5b37aea7f2970ae4dd2614b09d3dcf5dcbd1fc103b9997f8ecc3a804c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://17ae6aa6c53e7dd061f70b41428ef50bbc7ffcc65352fad1a4a9d3c69dbd2ba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17ae6aa6c53e7dd061f70b41428ef50bbc7ffcc65352fad1a4a9d3c69dbd2ba9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://ad4cd69df7608c7d87d839fe09d4557c674c7efbc4730f4e4d3daaf9a26f5aee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad4cd69df7608c7d87d839fe09d4557c674c7efbc4730f4e4d3daaf9a26f5aee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:36Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:48Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:48 crc kubenswrapper[4754]: I1005 20:55:48.525303 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:48Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:48 crc kubenswrapper[4754]: I1005 20:55:48.543353 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9ad070bb699b643d29b82bcc1126ce3b585f7492f5ae2817fa491b309b1691d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:48Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:48 crc kubenswrapper[4754]: I1005 20:55:48.559881 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-57sr4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e854c7b9-1229-4da4-80e9-fea6479459b8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceb577e1055d84007936a50d0b3d54b9ae0299a1fc1a6618b37163a6d3fbb9cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vw2ps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:58Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-57sr4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:48Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:48 crc kubenswrapper[4754]: I1005 20:55:48.575883 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:48 crc kubenswrapper[4754]: I1005 20:55:48.575909 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:48 crc kubenswrapper[4754]: I1005 20:55:48.575918 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:48 crc kubenswrapper[4754]: I1005 20:55:48.575935 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:48 crc kubenswrapper[4754]: I1005 20:55:48.575946 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:48Z","lastTransitionTime":"2025-10-05T20:55:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:48 crc kubenswrapper[4754]: I1005 20:55:48.579928 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cn76s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02d5e3f9-73c2-4496-9aca-0787184aef19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ed23b9cd9b57bb169a8ebbf16b7ff8d196eb14330df0edf2b9fcd2a8c9e2c88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7ba2b076839f4a5f586d2e6c4f4d7188254128b782bfd2180e22bf3ea5ffdd7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-05T20:55:46Z\\\",\\\"message\\\":\\\"2025-10-05T20:55:01+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_d9d50ba4-5e54-4920-83d9-a3faa08e80be\\\\n2025-10-05T20:55:01+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_d9d50ba4-5e54-4920-83d9-a3faa08e80be to /host/opt/cni/bin/\\\\n2025-10-05T20:55:01Z [verbose] multus-daemon started\\\\n2025-10-05T20:55:01Z [verbose] Readiness Indicator file check\\\\n2025-10-05T20:55:46Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpx5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cn76s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:48Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:48 crc kubenswrapper[4754]: I1005 20:55:48.589920 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ss7lm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f9ef817-0622-4185-a7fb-57221690aa75\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ba613b0d7753a003c572c3733ed394707c45b6f8072cb7aa88f3c7312b8882b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wpwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:55:02Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ss7lm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:48Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:48 crc kubenswrapper[4754]: I1005 20:55:48.604119 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-nwrnt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7dc728ea-2601-44f0-bcab-2913f034007d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-846xq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-846xq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:55:13Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-nwrnt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:48Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:48 crc kubenswrapper[4754]: I1005 20:55:48.620999 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"934bcd66-9c03-48b9-82dd-5093338f9c8f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9042df2f43243d3d79ba9a3940f8dd4f6c7a89ef9e3510e81fe60b1944d7e83b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5366bbf46a91667355f43f5df7335a536b395cda788b90b35ef18e2b53bc4c02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://600bc3bb551b747cf2f18e26745883232d7a37ac5877fad0c67100e52d7a43aa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78f74982180048eb50b9be79fa72c04f62f925b54fe511920f37e6407f864631\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:48Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:48 crc kubenswrapper[4754]: I1005 20:55:48.638342 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:48Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:48 crc kubenswrapper[4754]: I1005 20:55:48.656170 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tskt7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a68e9c89-465b-4c01-a470-e21a480ee465\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba6edba14cf321b4cd2f7bb44067a3ecf49f1594e76c0197e807ff3daa45e9b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d14ec99610d2167b484bd7712af7bb6d993d26c74b2828da1c89314f16fda22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d14ec99610d2167b484bd7712af7bb6d993d26c74b2828da1c89314f16fda22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f6f08dd853b25881e2f3c66206aa80d0c493518284217780aa40ac4b69f0c0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f6f08dd853b25881e2f3c66206aa80d0c493518284217780aa40ac4b69f0c0b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://71026d00ec1a3d91b019deb3c9f43ee52fb05b8f2d6374e1a5b762c9e5c1b10d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71026d00ec1a3d91b019deb3c9f43ee52fb05b8f2d6374e1a5b762c9e5c1b10d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b9910d2470acb16ebb436d7030b3161061607cb2a5ef50d482d8d5c4f48a8ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b9910d2470acb16ebb436d7030b3161061607cb2a5ef50d482d8d5c4f48a8ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0a93b61eacc55111995e69c15f89a210b8aa15468ac9ac053b7cf6b007d301e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b0a93b61eacc55111995e69c15f89a210b8aa15468ac9ac053b7cf6b007d301e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cafb3208e7e3bcdd01898a68fc5bbf4cb52ca0800b879b77a60bb102794257f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cafb3208e7e3bcdd01898a68fc5bbf4cb52ca0800b879b77a60bb102794257f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tskt7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:48Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:48 crc kubenswrapper[4754]: I1005 20:55:48.671653 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be95a413-6a59-45b4-84b7-b43cae694a26\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10b211cf6e47dbd1ee42c82e49de0a13a45581b92a0c4f6bbba0326c7149e41d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2ggg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4bb8c5903e8b70dd19eebb111097d96f725e17de44e06c3e74c067376f6c7360\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2ggg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-b2h9k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:48Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:48 crc kubenswrapper[4754]: I1005 20:55:48.678768 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:48 crc kubenswrapper[4754]: I1005 20:55:48.678795 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:48 crc kubenswrapper[4754]: I1005 20:55:48.678806 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:48 crc kubenswrapper[4754]: I1005 20:55:48.678820 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:48 crc kubenswrapper[4754]: I1005 20:55:48.678831 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:48Z","lastTransitionTime":"2025-10-05T20:55:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:48 crc kubenswrapper[4754]: I1005 20:55:48.684002 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rddkl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"15afc2a1-5cb7-41c6-954b-446c36822f50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://866d85ee126f8641d12f1a05fabc3de7b9392c324e9fe06ea92214155c05334c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqv8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a175565f83cfcba7825006eafb0f83efb70ec3d415c2c039c6499d5f6576e89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqv8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:55:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rddkl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:48Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:48 crc kubenswrapper[4754]: I1005 20:55:48.701728 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7506d496-65ff-41dc-8230-24c8ee653a2a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33da5a6bb3cd3b042c193bc571e6622f2f87f27e560cdd0bbb9be31e2010d707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df6d700a947d19c84b4f65fd2dd2f8d1f4f1f80fa9b4800c2acfa8dae3912639\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35ca8ee77a0e58cf969bd67b277a3392614c4e73b963e6f6002097709370715d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e32c2154f46ba10c509fe40c357304aa8e82d04e9906bbaf2ccecd47b42279e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8907b9b69b249c34ec6db1e460d641b31711b75d6d31d4caafd867339e6571d0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1005 20:54:52.589115 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 20:54:52.590181 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1304624737/tls.crt::/tmp/serving-cert-1304624737/tls.key\\\\\\\"\\\\nI1005 20:54:58.070544 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1005 20:54:58.090352 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1005 20:54:58.090382 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1005 20:54:58.090412 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1005 20:54:58.090418 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1005 20:54:58.144656 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1005 20:54:58.144688 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 20:54:58.144695 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 20:54:58.144699 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1005 20:54:58.144702 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1005 20:54:58.144705 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1005 20:54:58.144708 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1005 20:54:58.145024 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1005 20:54:58.148112 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4efb34f3493cc8854bc4cff6b29f63973c52b754404d7a21daec4e1f612d327d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54c568b637f3737667433047b956c7eb22fdf0f9301acdfd3b49fb84c39de6ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54c568b637f3737667433047b956c7eb22fdf0f9301acdfd3b49fb84c39de6ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:48Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:48 crc kubenswrapper[4754]: I1005 20:55:48.716371 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c76e6e4740035a6ca54191a770c4516e28101aae69afbb89739174c2f076ef5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9eededc0cb1af94fb7b5daba08785f6b077a85e9cb5ddd8b9f397b658194f61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:48Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:48 crc kubenswrapper[4754]: I1005 20:55:48.782017 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:48 crc kubenswrapper[4754]: I1005 20:55:48.782071 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:48 crc kubenswrapper[4754]: I1005 20:55:48.782085 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:48 crc kubenswrapper[4754]: I1005 20:55:48.782103 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:48 crc kubenswrapper[4754]: I1005 20:55:48.782117 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:48Z","lastTransitionTime":"2025-10-05T20:55:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:48 crc kubenswrapper[4754]: I1005 20:55:48.837020 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 20:55:48 crc kubenswrapper[4754]: I1005 20:55:48.837111 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 20:55:48 crc kubenswrapper[4754]: E1005 20:55:48.837161 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 20:55:48 crc kubenswrapper[4754]: I1005 20:55:48.837276 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 20:55:48 crc kubenswrapper[4754]: E1005 20:55:48.837339 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 20:55:48 crc kubenswrapper[4754]: E1005 20:55:48.837467 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 20:55:48 crc kubenswrapper[4754]: I1005 20:55:48.887283 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:48 crc kubenswrapper[4754]: I1005 20:55:48.887317 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:48 crc kubenswrapper[4754]: I1005 20:55:48.887328 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:48 crc kubenswrapper[4754]: I1005 20:55:48.887340 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:48 crc kubenswrapper[4754]: I1005 20:55:48.887350 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:48Z","lastTransitionTime":"2025-10-05T20:55:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:48 crc kubenswrapper[4754]: I1005 20:55:48.990002 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:48 crc kubenswrapper[4754]: I1005 20:55:48.990037 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:48 crc kubenswrapper[4754]: I1005 20:55:48.990047 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:48 crc kubenswrapper[4754]: I1005 20:55:48.990063 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:48 crc kubenswrapper[4754]: I1005 20:55:48.990073 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:48Z","lastTransitionTime":"2025-10-05T20:55:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:49 crc kubenswrapper[4754]: I1005 20:55:49.092809 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:49 crc kubenswrapper[4754]: I1005 20:55:49.092855 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:49 crc kubenswrapper[4754]: I1005 20:55:49.092868 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:49 crc kubenswrapper[4754]: I1005 20:55:49.092883 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:49 crc kubenswrapper[4754]: I1005 20:55:49.092895 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:49Z","lastTransitionTime":"2025-10-05T20:55:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:49 crc kubenswrapper[4754]: I1005 20:55:49.197074 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:49 crc kubenswrapper[4754]: I1005 20:55:49.197163 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:49 crc kubenswrapper[4754]: I1005 20:55:49.197182 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:49 crc kubenswrapper[4754]: I1005 20:55:49.197211 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:49 crc kubenswrapper[4754]: I1005 20:55:49.197248 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:49Z","lastTransitionTime":"2025-10-05T20:55:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:49 crc kubenswrapper[4754]: I1005 20:55:49.300907 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:49 crc kubenswrapper[4754]: I1005 20:55:49.301305 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:49 crc kubenswrapper[4754]: I1005 20:55:49.301386 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:49 crc kubenswrapper[4754]: I1005 20:55:49.301481 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:49 crc kubenswrapper[4754]: I1005 20:55:49.301626 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:49Z","lastTransitionTime":"2025-10-05T20:55:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:49 crc kubenswrapper[4754]: I1005 20:55:49.410214 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:49 crc kubenswrapper[4754]: I1005 20:55:49.410267 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:49 crc kubenswrapper[4754]: I1005 20:55:49.410278 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:49 crc kubenswrapper[4754]: I1005 20:55:49.410299 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:49 crc kubenswrapper[4754]: I1005 20:55:49.410310 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:49Z","lastTransitionTime":"2025-10-05T20:55:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:49 crc kubenswrapper[4754]: I1005 20:55:49.513451 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:49 crc kubenswrapper[4754]: I1005 20:55:49.513506 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:49 crc kubenswrapper[4754]: I1005 20:55:49.513521 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:49 crc kubenswrapper[4754]: I1005 20:55:49.513538 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:49 crc kubenswrapper[4754]: I1005 20:55:49.513549 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:49Z","lastTransitionTime":"2025-10-05T20:55:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:49 crc kubenswrapper[4754]: I1005 20:55:49.618114 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:49 crc kubenswrapper[4754]: I1005 20:55:49.618195 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:49 crc kubenswrapper[4754]: I1005 20:55:49.618212 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:49 crc kubenswrapper[4754]: I1005 20:55:49.618244 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:49 crc kubenswrapper[4754]: I1005 20:55:49.618260 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:49Z","lastTransitionTime":"2025-10-05T20:55:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:49 crc kubenswrapper[4754]: I1005 20:55:49.722159 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:49 crc kubenswrapper[4754]: I1005 20:55:49.722227 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:49 crc kubenswrapper[4754]: I1005 20:55:49.722244 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:49 crc kubenswrapper[4754]: I1005 20:55:49.722270 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:49 crc kubenswrapper[4754]: I1005 20:55:49.722290 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:49Z","lastTransitionTime":"2025-10-05T20:55:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:49 crc kubenswrapper[4754]: I1005 20:55:49.825185 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:49 crc kubenswrapper[4754]: I1005 20:55:49.825283 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:49 crc kubenswrapper[4754]: I1005 20:55:49.825301 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:49 crc kubenswrapper[4754]: I1005 20:55:49.825326 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:49 crc kubenswrapper[4754]: I1005 20:55:49.825345 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:49Z","lastTransitionTime":"2025-10-05T20:55:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:49 crc kubenswrapper[4754]: I1005 20:55:49.836780 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nwrnt" Oct 05 20:55:49 crc kubenswrapper[4754]: E1005 20:55:49.837000 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nwrnt" podUID="7dc728ea-2601-44f0-bcab-2913f034007d" Oct 05 20:55:49 crc kubenswrapper[4754]: I1005 20:55:49.928663 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:49 crc kubenswrapper[4754]: I1005 20:55:49.928755 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:49 crc kubenswrapper[4754]: I1005 20:55:49.928773 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:49 crc kubenswrapper[4754]: I1005 20:55:49.928799 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:49 crc kubenswrapper[4754]: I1005 20:55:49.928815 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:49Z","lastTransitionTime":"2025-10-05T20:55:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:50 crc kubenswrapper[4754]: I1005 20:55:50.032229 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:50 crc kubenswrapper[4754]: I1005 20:55:50.032275 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:50 crc kubenswrapper[4754]: I1005 20:55:50.032287 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:50 crc kubenswrapper[4754]: I1005 20:55:50.032303 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:50 crc kubenswrapper[4754]: I1005 20:55:50.032314 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:50Z","lastTransitionTime":"2025-10-05T20:55:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:50 crc kubenswrapper[4754]: I1005 20:55:50.136052 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:50 crc kubenswrapper[4754]: I1005 20:55:50.136100 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:50 crc kubenswrapper[4754]: I1005 20:55:50.136112 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:50 crc kubenswrapper[4754]: I1005 20:55:50.136133 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:50 crc kubenswrapper[4754]: I1005 20:55:50.136149 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:50Z","lastTransitionTime":"2025-10-05T20:55:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:50 crc kubenswrapper[4754]: I1005 20:55:50.239444 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:50 crc kubenswrapper[4754]: I1005 20:55:50.239621 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:50 crc kubenswrapper[4754]: I1005 20:55:50.239649 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:50 crc kubenswrapper[4754]: I1005 20:55:50.239679 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:50 crc kubenswrapper[4754]: I1005 20:55:50.239700 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:50Z","lastTransitionTime":"2025-10-05T20:55:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:50 crc kubenswrapper[4754]: I1005 20:55:50.343282 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:50 crc kubenswrapper[4754]: I1005 20:55:50.343318 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:50 crc kubenswrapper[4754]: I1005 20:55:50.343327 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:50 crc kubenswrapper[4754]: I1005 20:55:50.343342 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:50 crc kubenswrapper[4754]: I1005 20:55:50.343352 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:50Z","lastTransitionTime":"2025-10-05T20:55:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:50 crc kubenswrapper[4754]: I1005 20:55:50.446863 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:50 crc kubenswrapper[4754]: I1005 20:55:50.446929 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:50 crc kubenswrapper[4754]: I1005 20:55:50.446948 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:50 crc kubenswrapper[4754]: I1005 20:55:50.446975 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:50 crc kubenswrapper[4754]: I1005 20:55:50.446996 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:50Z","lastTransitionTime":"2025-10-05T20:55:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:50 crc kubenswrapper[4754]: I1005 20:55:50.550489 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:50 crc kubenswrapper[4754]: I1005 20:55:50.551709 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:50 crc kubenswrapper[4754]: I1005 20:55:50.551872 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:50 crc kubenswrapper[4754]: I1005 20:55:50.552015 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:50 crc kubenswrapper[4754]: I1005 20:55:50.552154 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:50Z","lastTransitionTime":"2025-10-05T20:55:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:50 crc kubenswrapper[4754]: I1005 20:55:50.655946 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:50 crc kubenswrapper[4754]: I1005 20:55:50.656012 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:50 crc kubenswrapper[4754]: I1005 20:55:50.656025 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:50 crc kubenswrapper[4754]: I1005 20:55:50.656048 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:50 crc kubenswrapper[4754]: I1005 20:55:50.656063 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:50Z","lastTransitionTime":"2025-10-05T20:55:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:50 crc kubenswrapper[4754]: I1005 20:55:50.760172 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:50 crc kubenswrapper[4754]: I1005 20:55:50.760242 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:50 crc kubenswrapper[4754]: I1005 20:55:50.760260 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:50 crc kubenswrapper[4754]: I1005 20:55:50.760307 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:50 crc kubenswrapper[4754]: I1005 20:55:50.760330 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:50Z","lastTransitionTime":"2025-10-05T20:55:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:50 crc kubenswrapper[4754]: I1005 20:55:50.836978 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 20:55:50 crc kubenswrapper[4754]: I1005 20:55:50.836978 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 20:55:50 crc kubenswrapper[4754]: I1005 20:55:50.836977 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 20:55:50 crc kubenswrapper[4754]: E1005 20:55:50.837318 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 20:55:50 crc kubenswrapper[4754]: E1005 20:55:50.837603 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 20:55:50 crc kubenswrapper[4754]: E1005 20:55:50.837756 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 20:55:50 crc kubenswrapper[4754]: I1005 20:55:50.863889 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:50 crc kubenswrapper[4754]: I1005 20:55:50.863962 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:50 crc kubenswrapper[4754]: I1005 20:55:50.863982 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:50 crc kubenswrapper[4754]: I1005 20:55:50.864012 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:50 crc kubenswrapper[4754]: I1005 20:55:50.864035 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:50Z","lastTransitionTime":"2025-10-05T20:55:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:50 crc kubenswrapper[4754]: I1005 20:55:50.968350 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:50 crc kubenswrapper[4754]: I1005 20:55:50.968420 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:50 crc kubenswrapper[4754]: I1005 20:55:50.968438 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:50 crc kubenswrapper[4754]: I1005 20:55:50.968468 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:50 crc kubenswrapper[4754]: I1005 20:55:50.968486 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:50Z","lastTransitionTime":"2025-10-05T20:55:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:51 crc kubenswrapper[4754]: I1005 20:55:51.072234 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:51 crc kubenswrapper[4754]: I1005 20:55:51.072974 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:51 crc kubenswrapper[4754]: I1005 20:55:51.073221 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:51 crc kubenswrapper[4754]: I1005 20:55:51.073366 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:51 crc kubenswrapper[4754]: I1005 20:55:51.073479 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:51Z","lastTransitionTime":"2025-10-05T20:55:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:51 crc kubenswrapper[4754]: I1005 20:55:51.177035 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:51 crc kubenswrapper[4754]: I1005 20:55:51.177087 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:51 crc kubenswrapper[4754]: I1005 20:55:51.177108 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:51 crc kubenswrapper[4754]: I1005 20:55:51.177138 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:51 crc kubenswrapper[4754]: I1005 20:55:51.177157 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:51Z","lastTransitionTime":"2025-10-05T20:55:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:51 crc kubenswrapper[4754]: I1005 20:55:51.280691 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:51 crc kubenswrapper[4754]: I1005 20:55:51.280810 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:51 crc kubenswrapper[4754]: I1005 20:55:51.280846 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:51 crc kubenswrapper[4754]: I1005 20:55:51.280884 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:51 crc kubenswrapper[4754]: I1005 20:55:51.280909 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:51Z","lastTransitionTime":"2025-10-05T20:55:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:51 crc kubenswrapper[4754]: I1005 20:55:51.385676 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:51 crc kubenswrapper[4754]: I1005 20:55:51.386014 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:51 crc kubenswrapper[4754]: I1005 20:55:51.386152 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:51 crc kubenswrapper[4754]: I1005 20:55:51.386249 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:51 crc kubenswrapper[4754]: I1005 20:55:51.386339 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:51Z","lastTransitionTime":"2025-10-05T20:55:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:51 crc kubenswrapper[4754]: I1005 20:55:51.489730 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:51 crc kubenswrapper[4754]: I1005 20:55:51.489793 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:51 crc kubenswrapper[4754]: I1005 20:55:51.489812 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:51 crc kubenswrapper[4754]: I1005 20:55:51.489832 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:51 crc kubenswrapper[4754]: I1005 20:55:51.489844 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:51Z","lastTransitionTime":"2025-10-05T20:55:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:51 crc kubenswrapper[4754]: I1005 20:55:51.593255 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:51 crc kubenswrapper[4754]: I1005 20:55:51.593364 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:51 crc kubenswrapper[4754]: I1005 20:55:51.593384 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:51 crc kubenswrapper[4754]: I1005 20:55:51.593405 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:51 crc kubenswrapper[4754]: I1005 20:55:51.593419 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:51Z","lastTransitionTime":"2025-10-05T20:55:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:51 crc kubenswrapper[4754]: I1005 20:55:51.695955 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:51 crc kubenswrapper[4754]: I1005 20:55:51.696012 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:51 crc kubenswrapper[4754]: I1005 20:55:51.696026 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:51 crc kubenswrapper[4754]: I1005 20:55:51.696045 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:51 crc kubenswrapper[4754]: I1005 20:55:51.696064 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:51Z","lastTransitionTime":"2025-10-05T20:55:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:51 crc kubenswrapper[4754]: I1005 20:55:51.799557 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:51 crc kubenswrapper[4754]: I1005 20:55:51.799918 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:51 crc kubenswrapper[4754]: I1005 20:55:51.799952 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:51 crc kubenswrapper[4754]: I1005 20:55:51.800008 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:51 crc kubenswrapper[4754]: I1005 20:55:51.800188 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:51Z","lastTransitionTime":"2025-10-05T20:55:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:51 crc kubenswrapper[4754]: I1005 20:55:51.836520 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nwrnt" Oct 05 20:55:51 crc kubenswrapper[4754]: E1005 20:55:51.836756 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nwrnt" podUID="7dc728ea-2601-44f0-bcab-2913f034007d" Oct 05 20:55:51 crc kubenswrapper[4754]: I1005 20:55:51.904202 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:51 crc kubenswrapper[4754]: I1005 20:55:51.904301 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:51 crc kubenswrapper[4754]: I1005 20:55:51.904322 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:51 crc kubenswrapper[4754]: I1005 20:55:51.904352 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:51 crc kubenswrapper[4754]: I1005 20:55:51.904372 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:51Z","lastTransitionTime":"2025-10-05T20:55:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:52 crc kubenswrapper[4754]: I1005 20:55:52.007951 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:52 crc kubenswrapper[4754]: I1005 20:55:52.008015 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:52 crc kubenswrapper[4754]: I1005 20:55:52.008033 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:52 crc kubenswrapper[4754]: I1005 20:55:52.008059 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:52 crc kubenswrapper[4754]: I1005 20:55:52.008079 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:52Z","lastTransitionTime":"2025-10-05T20:55:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:52 crc kubenswrapper[4754]: I1005 20:55:52.155903 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:52 crc kubenswrapper[4754]: I1005 20:55:52.155962 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:52 crc kubenswrapper[4754]: I1005 20:55:52.155976 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:52 crc kubenswrapper[4754]: I1005 20:55:52.156001 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:52 crc kubenswrapper[4754]: I1005 20:55:52.156015 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:52Z","lastTransitionTime":"2025-10-05T20:55:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:52 crc kubenswrapper[4754]: I1005 20:55:52.259755 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:52 crc kubenswrapper[4754]: I1005 20:55:52.259827 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:52 crc kubenswrapper[4754]: I1005 20:55:52.259845 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:52 crc kubenswrapper[4754]: I1005 20:55:52.259871 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:52 crc kubenswrapper[4754]: I1005 20:55:52.259891 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:52Z","lastTransitionTime":"2025-10-05T20:55:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:52 crc kubenswrapper[4754]: I1005 20:55:52.363825 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:52 crc kubenswrapper[4754]: I1005 20:55:52.363871 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:52 crc kubenswrapper[4754]: I1005 20:55:52.363881 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:52 crc kubenswrapper[4754]: I1005 20:55:52.363897 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:52 crc kubenswrapper[4754]: I1005 20:55:52.363910 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:52Z","lastTransitionTime":"2025-10-05T20:55:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:52 crc kubenswrapper[4754]: I1005 20:55:52.467025 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:52 crc kubenswrapper[4754]: I1005 20:55:52.467096 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:52 crc kubenswrapper[4754]: I1005 20:55:52.467115 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:52 crc kubenswrapper[4754]: I1005 20:55:52.467142 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:52 crc kubenswrapper[4754]: I1005 20:55:52.467162 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:52Z","lastTransitionTime":"2025-10-05T20:55:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:52 crc kubenswrapper[4754]: I1005 20:55:52.570626 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:52 crc kubenswrapper[4754]: I1005 20:55:52.570679 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:52 crc kubenswrapper[4754]: I1005 20:55:52.570693 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:52 crc kubenswrapper[4754]: I1005 20:55:52.570715 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:52 crc kubenswrapper[4754]: I1005 20:55:52.570728 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:52Z","lastTransitionTime":"2025-10-05T20:55:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:52 crc kubenswrapper[4754]: I1005 20:55:52.674814 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:52 crc kubenswrapper[4754]: I1005 20:55:52.674897 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:52 crc kubenswrapper[4754]: I1005 20:55:52.674928 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:52 crc kubenswrapper[4754]: I1005 20:55:52.674965 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:52 crc kubenswrapper[4754]: I1005 20:55:52.675291 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:52Z","lastTransitionTime":"2025-10-05T20:55:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:52 crc kubenswrapper[4754]: I1005 20:55:52.779286 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:52 crc kubenswrapper[4754]: I1005 20:55:52.779353 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:52 crc kubenswrapper[4754]: I1005 20:55:52.779374 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:52 crc kubenswrapper[4754]: I1005 20:55:52.779406 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:52 crc kubenswrapper[4754]: I1005 20:55:52.779425 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:52Z","lastTransitionTime":"2025-10-05T20:55:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:52 crc kubenswrapper[4754]: I1005 20:55:52.836786 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 20:55:52 crc kubenswrapper[4754]: I1005 20:55:52.836889 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 20:55:52 crc kubenswrapper[4754]: I1005 20:55:52.836927 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 20:55:52 crc kubenswrapper[4754]: E1005 20:55:52.837107 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 20:55:52 crc kubenswrapper[4754]: E1005 20:55:52.837366 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 20:55:52 crc kubenswrapper[4754]: E1005 20:55:52.837624 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 20:55:52 crc kubenswrapper[4754]: I1005 20:55:52.884145 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:52 crc kubenswrapper[4754]: I1005 20:55:52.884217 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:52 crc kubenswrapper[4754]: I1005 20:55:52.884245 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:52 crc kubenswrapper[4754]: I1005 20:55:52.884280 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:52 crc kubenswrapper[4754]: I1005 20:55:52.884304 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:52Z","lastTransitionTime":"2025-10-05T20:55:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:52 crc kubenswrapper[4754]: I1005 20:55:52.989071 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:52 crc kubenswrapper[4754]: I1005 20:55:52.989143 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:52 crc kubenswrapper[4754]: I1005 20:55:52.989167 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:52 crc kubenswrapper[4754]: I1005 20:55:52.989206 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:52 crc kubenswrapper[4754]: I1005 20:55:52.989232 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:52Z","lastTransitionTime":"2025-10-05T20:55:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:53 crc kubenswrapper[4754]: I1005 20:55:53.092589 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:53 crc kubenswrapper[4754]: I1005 20:55:53.092645 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:53 crc kubenswrapper[4754]: I1005 20:55:53.092664 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:53 crc kubenswrapper[4754]: I1005 20:55:53.092690 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:53 crc kubenswrapper[4754]: I1005 20:55:53.092708 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:53Z","lastTransitionTime":"2025-10-05T20:55:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:53 crc kubenswrapper[4754]: I1005 20:55:53.196388 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:53 crc kubenswrapper[4754]: I1005 20:55:53.196464 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:53 crc kubenswrapper[4754]: I1005 20:55:53.196541 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:53 crc kubenswrapper[4754]: I1005 20:55:53.196580 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:53 crc kubenswrapper[4754]: I1005 20:55:53.196602 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:53Z","lastTransitionTime":"2025-10-05T20:55:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:53 crc kubenswrapper[4754]: I1005 20:55:53.299430 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:53 crc kubenswrapper[4754]: I1005 20:55:53.299561 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:53 crc kubenswrapper[4754]: I1005 20:55:53.299577 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:53 crc kubenswrapper[4754]: I1005 20:55:53.299595 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:53 crc kubenswrapper[4754]: I1005 20:55:53.299608 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:53Z","lastTransitionTime":"2025-10-05T20:55:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:53 crc kubenswrapper[4754]: I1005 20:55:53.403191 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:53 crc kubenswrapper[4754]: I1005 20:55:53.403254 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:53 crc kubenswrapper[4754]: I1005 20:55:53.403275 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:53 crc kubenswrapper[4754]: I1005 20:55:53.403305 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:53 crc kubenswrapper[4754]: I1005 20:55:53.403324 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:53Z","lastTransitionTime":"2025-10-05T20:55:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:53 crc kubenswrapper[4754]: I1005 20:55:53.506685 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:53 crc kubenswrapper[4754]: I1005 20:55:53.506764 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:53 crc kubenswrapper[4754]: I1005 20:55:53.506786 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:53 crc kubenswrapper[4754]: I1005 20:55:53.506814 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:53 crc kubenswrapper[4754]: I1005 20:55:53.506837 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:53Z","lastTransitionTime":"2025-10-05T20:55:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:53 crc kubenswrapper[4754]: I1005 20:55:53.609379 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:53 crc kubenswrapper[4754]: I1005 20:55:53.609438 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:53 crc kubenswrapper[4754]: I1005 20:55:53.609452 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:53 crc kubenswrapper[4754]: I1005 20:55:53.609476 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:53 crc kubenswrapper[4754]: I1005 20:55:53.609495 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:53Z","lastTransitionTime":"2025-10-05T20:55:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:53 crc kubenswrapper[4754]: I1005 20:55:53.712878 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:53 crc kubenswrapper[4754]: I1005 20:55:53.712976 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:53 crc kubenswrapper[4754]: I1005 20:55:53.713002 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:53 crc kubenswrapper[4754]: I1005 20:55:53.713036 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:53 crc kubenswrapper[4754]: I1005 20:55:53.713056 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:53Z","lastTransitionTime":"2025-10-05T20:55:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:53 crc kubenswrapper[4754]: I1005 20:55:53.815596 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:53 crc kubenswrapper[4754]: I1005 20:55:53.815654 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:53 crc kubenswrapper[4754]: I1005 20:55:53.815665 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:53 crc kubenswrapper[4754]: I1005 20:55:53.815686 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:53 crc kubenswrapper[4754]: I1005 20:55:53.815699 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:53Z","lastTransitionTime":"2025-10-05T20:55:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:53 crc kubenswrapper[4754]: I1005 20:55:53.836978 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nwrnt" Oct 05 20:55:53 crc kubenswrapper[4754]: E1005 20:55:53.837301 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nwrnt" podUID="7dc728ea-2601-44f0-bcab-2913f034007d" Oct 05 20:55:53 crc kubenswrapper[4754]: I1005 20:55:53.839625 4754 scope.go:117] "RemoveContainer" containerID="af1c313e0dc0fe3b48feac3cce1ecd4452c9efadf3d19f257c07cc7fcdec875f" Oct 05 20:55:53 crc kubenswrapper[4754]: I1005 20:55:53.861919 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Oct 05 20:55:53 crc kubenswrapper[4754]: I1005 20:55:53.920133 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:53 crc kubenswrapper[4754]: I1005 20:55:53.920190 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:53 crc kubenswrapper[4754]: I1005 20:55:53.920205 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:53 crc kubenswrapper[4754]: I1005 20:55:53.920225 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:53 crc kubenswrapper[4754]: I1005 20:55:53.920562 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:53Z","lastTransitionTime":"2025-10-05T20:55:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:54 crc kubenswrapper[4754]: I1005 20:55:54.026034 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:54 crc kubenswrapper[4754]: I1005 20:55:54.026225 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:54 crc kubenswrapper[4754]: I1005 20:55:54.026432 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:54 crc kubenswrapper[4754]: I1005 20:55:54.026598 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:54 crc kubenswrapper[4754]: I1005 20:55:54.026781 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:54Z","lastTransitionTime":"2025-10-05T20:55:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:54 crc kubenswrapper[4754]: I1005 20:55:54.129954 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:54 crc kubenswrapper[4754]: I1005 20:55:54.129993 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:54 crc kubenswrapper[4754]: I1005 20:55:54.130025 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:54 crc kubenswrapper[4754]: I1005 20:55:54.130076 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:54 crc kubenswrapper[4754]: I1005 20:55:54.130113 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:54Z","lastTransitionTime":"2025-10-05T20:55:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:54 crc kubenswrapper[4754]: I1005 20:55:54.233554 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:54 crc kubenswrapper[4754]: I1005 20:55:54.233612 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:54 crc kubenswrapper[4754]: I1005 20:55:54.233623 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:54 crc kubenswrapper[4754]: I1005 20:55:54.233642 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:54 crc kubenswrapper[4754]: I1005 20:55:54.233653 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:54Z","lastTransitionTime":"2025-10-05T20:55:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:54 crc kubenswrapper[4754]: I1005 20:55:54.336237 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:54 crc kubenswrapper[4754]: I1005 20:55:54.336294 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:54 crc kubenswrapper[4754]: I1005 20:55:54.336309 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:54 crc kubenswrapper[4754]: I1005 20:55:54.336328 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:54 crc kubenswrapper[4754]: I1005 20:55:54.336338 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:54Z","lastTransitionTime":"2025-10-05T20:55:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:54 crc kubenswrapper[4754]: I1005 20:55:54.440014 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:54 crc kubenswrapper[4754]: I1005 20:55:54.440066 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:54 crc kubenswrapper[4754]: I1005 20:55:54.440081 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:54 crc kubenswrapper[4754]: I1005 20:55:54.440297 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:54 crc kubenswrapper[4754]: I1005 20:55:54.440312 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:54Z","lastTransitionTime":"2025-10-05T20:55:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:54 crc kubenswrapper[4754]: I1005 20:55:54.444421 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7rhps_b44790fb-fecb-4ec3-9816-8e23dd72ab33/ovnkube-controller/2.log" Oct 05 20:55:54 crc kubenswrapper[4754]: I1005 20:55:54.447961 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" event={"ID":"b44790fb-fecb-4ec3-9816-8e23dd72ab33","Type":"ContainerStarted","Data":"7c990164cb7213726a9086a630586e38da00d62fdb57aed8418bbec265a54705"} Oct 05 20:55:54 crc kubenswrapper[4754]: I1005 20:55:54.448896 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" Oct 05 20:55:54 crc kubenswrapper[4754]: I1005 20:55:54.469188 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e380f9b2a0e319dc325c54d1247a6db98346292548fa688cb41a2ca1b3048419\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:54Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:54 crc kubenswrapper[4754]: I1005 20:55:54.482690 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:54Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:54 crc kubenswrapper[4754]: I1005 20:55:54.495971 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"860bd54f-d440-4642-bbe1-740e500c7551\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be822fa37e7f0aabf2715000b5a12edf3fe5c3df2999739551d2a89242c18129\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6482afbcf8935223f52760fa7bf9ab57e151e8fe9999f0ca43fc6db9f9f7b6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6482afbcf8935223f52760fa7bf9ab57e151e8fe9999f0ca43fc6db9f9f7b6f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:36Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:54Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:54 crc kubenswrapper[4754]: I1005 20:55:54.511624 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ae81d5f-5e31-4648-a80c-001034f0acf8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02dec08702aeaa03bc26151781909a4c6226fdaf37b921feb2a53ad6fa3c94c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a0c1914423693dd5ca96eddab941162a0f1e9efaf4bd02c47514f092e39e9e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5d1820b7c2ba7c5262f52297111f58e8f05771f3d7e34a59f8f3a05a7fb0bfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c187ac7d3514f4e68984ee98c274c53418df1818159cc9d6c16626d162fa3f97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c187ac7d3514f4e68984ee98c274c53418df1818159cc9d6c16626d162fa3f97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:36Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:54Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:54 crc kubenswrapper[4754]: I1005 20:55:54.534978 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b44790fb-fecb-4ec3-9816-8e23dd72ab33\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a7dc53e7f96e48213d724a2cb4859514df925005644108ae3a0c33b14dcf52d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfdd8b1e542ab676576c311e2923fb02861f92de91fe1f4b3e4b968c549c375d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3a63451ffaa14cacc3b78d11340f4ea81445fa3af96178e9b172d063d9697d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3cfd3f2274a8a46fa36660e94ac6c9c298c791283f12d52330c5594ac131998\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f626cb108716da0b595844ea0251b7a04d49b5908a1d6dd0aef5cb2d402c4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://822223f515b5add6d71d193fd5bc77df029b8d37a1cc07be4145b53c7226bf61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c990164cb7213726a9086a630586e38da00d62fdb57aed8418bbec265a54705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af1c313e0dc0fe3b48feac3cce1ecd4452c9efadf3d19f257c07cc7fcdec875f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-05T20:55:26Z\\\",\\\"message\\\":\\\"5:26.770640 6339 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/iptables-alerter-4ln5h in node crc\\\\nI1005 20:55:26.770478 6339 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-apiserver-operator/metrics]} name:Service_openshift-apiserver-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.38:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {8b82f026-5975-4a1b-bb18-08d5d51147ec}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1005 20:55:26.770652 6339 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/iptables-alerter-4ln5h after 0 failed attempt(s)\\\\nI1005 20:55:26.770662 6339 default_network_controller.go:776] Recording success event on pod openshift-network-operator/iptables-alerter-4ln5h\\\\nF1005 20:55:26.770677 6339 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controlle\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:25Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2acf9637fa25e5d3ffeb05224c11218c47defc679568fb928a498869613de4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35ebeb7a87d5d47b6567917d38eca38a3e622c497fba58971296c4609c287423\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35ebeb7a87d5d47b6567917d38eca38a3e622c497fba58971296c4609c287423\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7rhps\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:54Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:54 crc kubenswrapper[4754]: I1005 20:55:54.543495 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:54 crc kubenswrapper[4754]: I1005 20:55:54.543561 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:54 crc kubenswrapper[4754]: I1005 20:55:54.543573 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:54 crc kubenswrapper[4754]: I1005 20:55:54.543593 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:54 crc kubenswrapper[4754]: I1005 20:55:54.543605 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:54Z","lastTransitionTime":"2025-10-05T20:55:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:54 crc kubenswrapper[4754]: I1005 20:55:54.557176 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9309179-b92a-4fa4-9fba-747b8f911936\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d969d3d50d6ad280fc6654df9441bab238bb538f277ed6d1273779813ffbaaed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://800a731ed8ef57714b2318af14054ef23f0368936630887a093fc77167b33d12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8cd9b08e83dc174413cc37891c14562816787028471a4201680c86dd620b7a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93b729f3f7f993db0d6bc3ba60c04930aada3cef8272ffcc3c6a8a242a9b1475\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dde1519111fc3af5d14463fdb58d4f5c62a7045e7de98c60981c7e4da37b0e7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ea160a5b37aea7f2970ae4dd2614b09d3dcf5dcbd1fc103b9997f8ecc3a804c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ea160a5b37aea7f2970ae4dd2614b09d3dcf5dcbd1fc103b9997f8ecc3a804c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://17ae6aa6c53e7dd061f70b41428ef50bbc7ffcc65352fad1a4a9d3c69dbd2ba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17ae6aa6c53e7dd061f70b41428ef50bbc7ffcc65352fad1a4a9d3c69dbd2ba9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://ad4cd69df7608c7d87d839fe09d4557c674c7efbc4730f4e4d3daaf9a26f5aee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad4cd69df7608c7d87d839fe09d4557c674c7efbc4730f4e4d3daaf9a26f5aee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:36Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:54Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:54 crc kubenswrapper[4754]: I1005 20:55:54.570046 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:54Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:54 crc kubenswrapper[4754]: I1005 20:55:54.580833 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9ad070bb699b643d29b82bcc1126ce3b585f7492f5ae2817fa491b309b1691d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:54Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:54 crc kubenswrapper[4754]: I1005 20:55:54.590236 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-57sr4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e854c7b9-1229-4da4-80e9-fea6479459b8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceb577e1055d84007936a50d0b3d54b9ae0299a1fc1a6618b37163a6d3fbb9cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vw2ps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:58Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-57sr4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:54Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:54 crc kubenswrapper[4754]: I1005 20:55:54.603267 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cn76s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02d5e3f9-73c2-4496-9aca-0787184aef19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ed23b9cd9b57bb169a8ebbf16b7ff8d196eb14330df0edf2b9fcd2a8c9e2c88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7ba2b076839f4a5f586d2e6c4f4d7188254128b782bfd2180e22bf3ea5ffdd7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-05T20:55:46Z\\\",\\\"message\\\":\\\"2025-10-05T20:55:01+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_d9d50ba4-5e54-4920-83d9-a3faa08e80be\\\\n2025-10-05T20:55:01+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_d9d50ba4-5e54-4920-83d9-a3faa08e80be to /host/opt/cni/bin/\\\\n2025-10-05T20:55:01Z [verbose] multus-daemon started\\\\n2025-10-05T20:55:01Z [verbose] Readiness Indicator file check\\\\n2025-10-05T20:55:46Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpx5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cn76s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:54Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:54 crc kubenswrapper[4754]: I1005 20:55:54.616970 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ss7lm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f9ef817-0622-4185-a7fb-57221690aa75\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ba613b0d7753a003c572c3733ed394707c45b6f8072cb7aa88f3c7312b8882b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wpwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:55:02Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ss7lm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:54Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:54 crc kubenswrapper[4754]: I1005 20:55:54.631832 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-nwrnt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7dc728ea-2601-44f0-bcab-2913f034007d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-846xq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-846xq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:55:13Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-nwrnt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:54Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:54 crc kubenswrapper[4754]: I1005 20:55:54.646926 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:54 crc kubenswrapper[4754]: I1005 20:55:54.646966 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:54 crc kubenswrapper[4754]: I1005 20:55:54.646976 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:54 crc kubenswrapper[4754]: I1005 20:55:54.646993 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:54 crc kubenswrapper[4754]: I1005 20:55:54.647003 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:54Z","lastTransitionTime":"2025-10-05T20:55:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:54 crc kubenswrapper[4754]: I1005 20:55:54.648706 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"934bcd66-9c03-48b9-82dd-5093338f9c8f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9042df2f43243d3d79ba9a3940f8dd4f6c7a89ef9e3510e81fe60b1944d7e83b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5366bbf46a91667355f43f5df7335a536b395cda788b90b35ef18e2b53bc4c02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://600bc3bb551b747cf2f18e26745883232d7a37ac5877fad0c67100e52d7a43aa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78f74982180048eb50b9be79fa72c04f62f925b54fe511920f37e6407f864631\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:54Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:54 crc kubenswrapper[4754]: I1005 20:55:54.662920 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:54Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:54 crc kubenswrapper[4754]: I1005 20:55:54.684712 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tskt7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a68e9c89-465b-4c01-a470-e21a480ee465\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba6edba14cf321b4cd2f7bb44067a3ecf49f1594e76c0197e807ff3daa45e9b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d14ec99610d2167b484bd7712af7bb6d993d26c74b2828da1c89314f16fda22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d14ec99610d2167b484bd7712af7bb6d993d26c74b2828da1c89314f16fda22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f6f08dd853b25881e2f3c66206aa80d0c493518284217780aa40ac4b69f0c0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f6f08dd853b25881e2f3c66206aa80d0c493518284217780aa40ac4b69f0c0b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://71026d00ec1a3d91b019deb3c9f43ee52fb05b8f2d6374e1a5b762c9e5c1b10d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71026d00ec1a3d91b019deb3c9f43ee52fb05b8f2d6374e1a5b762c9e5c1b10d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b9910d2470acb16ebb436d7030b3161061607cb2a5ef50d482d8d5c4f48a8ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b9910d2470acb16ebb436d7030b3161061607cb2a5ef50d482d8d5c4f48a8ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0a93b61eacc55111995e69c15f89a210b8aa15468ac9ac053b7cf6b007d301e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b0a93b61eacc55111995e69c15f89a210b8aa15468ac9ac053b7cf6b007d301e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cafb3208e7e3bcdd01898a68fc5bbf4cb52ca0800b879b77a60bb102794257f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cafb3208e7e3bcdd01898a68fc5bbf4cb52ca0800b879b77a60bb102794257f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tskt7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:54Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:54 crc kubenswrapper[4754]: I1005 20:55:54.703276 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be95a413-6a59-45b4-84b7-b43cae694a26\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10b211cf6e47dbd1ee42c82e49de0a13a45581b92a0c4f6bbba0326c7149e41d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2ggg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4bb8c5903e8b70dd19eebb111097d96f725e17de44e06c3e74c067376f6c7360\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2ggg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-b2h9k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:54Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:54 crc kubenswrapper[4754]: I1005 20:55:54.723388 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rddkl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"15afc2a1-5cb7-41c6-954b-446c36822f50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://866d85ee126f8641d12f1a05fabc3de7b9392c324e9fe06ea92214155c05334c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqv8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a175565f83cfcba7825006eafb0f83efb70ec3d415c2c039c6499d5f6576e89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqv8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:55:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rddkl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:54Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:54 crc kubenswrapper[4754]: I1005 20:55:54.748397 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7506d496-65ff-41dc-8230-24c8ee653a2a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33da5a6bb3cd3b042c193bc571e6622f2f87f27e560cdd0bbb9be31e2010d707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df6d700a947d19c84b4f65fd2dd2f8d1f4f1f80fa9b4800c2acfa8dae3912639\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35ca8ee77a0e58cf969bd67b277a3392614c4e73b963e6f6002097709370715d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e32c2154f46ba10c509fe40c357304aa8e82d04e9906bbaf2ccecd47b42279e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8907b9b69b249c34ec6db1e460d641b31711b75d6d31d4caafd867339e6571d0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1005 20:54:52.589115 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 20:54:52.590181 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1304624737/tls.crt::/tmp/serving-cert-1304624737/tls.key\\\\\\\"\\\\nI1005 20:54:58.070544 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1005 20:54:58.090352 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1005 20:54:58.090382 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1005 20:54:58.090412 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1005 20:54:58.090418 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1005 20:54:58.144656 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1005 20:54:58.144688 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 20:54:58.144695 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 20:54:58.144699 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1005 20:54:58.144702 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1005 20:54:58.144705 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1005 20:54:58.144708 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1005 20:54:58.145024 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1005 20:54:58.148112 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4efb34f3493cc8854bc4cff6b29f63973c52b754404d7a21daec4e1f612d327d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54c568b637f3737667433047b956c7eb22fdf0f9301acdfd3b49fb84c39de6ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54c568b637f3737667433047b956c7eb22fdf0f9301acdfd3b49fb84c39de6ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:54Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:54 crc kubenswrapper[4754]: I1005 20:55:54.762837 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:54 crc kubenswrapper[4754]: I1005 20:55:54.762899 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:54 crc kubenswrapper[4754]: I1005 20:55:54.762919 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:54 crc kubenswrapper[4754]: I1005 20:55:54.762946 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:54 crc kubenswrapper[4754]: I1005 20:55:54.762967 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:54Z","lastTransitionTime":"2025-10-05T20:55:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:54 crc kubenswrapper[4754]: I1005 20:55:54.788456 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c76e6e4740035a6ca54191a770c4516e28101aae69afbb89739174c2f076ef5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9eededc0cb1af94fb7b5daba08785f6b077a85e9cb5ddd8b9f397b658194f61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:54Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:54 crc kubenswrapper[4754]: I1005 20:55:54.837460 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 20:55:54 crc kubenswrapper[4754]: I1005 20:55:54.837498 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 20:55:54 crc kubenswrapper[4754]: E1005 20:55:54.837844 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 20:55:54 crc kubenswrapper[4754]: I1005 20:55:54.837898 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 20:55:54 crc kubenswrapper[4754]: E1005 20:55:54.838131 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 20:55:54 crc kubenswrapper[4754]: E1005 20:55:54.839701 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 20:55:54 crc kubenswrapper[4754]: I1005 20:55:54.867726 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:54 crc kubenswrapper[4754]: I1005 20:55:54.867768 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:54 crc kubenswrapper[4754]: I1005 20:55:54.867781 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:54 crc kubenswrapper[4754]: I1005 20:55:54.867801 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:54 crc kubenswrapper[4754]: I1005 20:55:54.867821 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:54Z","lastTransitionTime":"2025-10-05T20:55:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:54 crc kubenswrapper[4754]: I1005 20:55:54.972230 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:54 crc kubenswrapper[4754]: I1005 20:55:54.972306 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:54 crc kubenswrapper[4754]: I1005 20:55:54.972328 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:54 crc kubenswrapper[4754]: I1005 20:55:54.972354 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:54 crc kubenswrapper[4754]: I1005 20:55:54.972378 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:54Z","lastTransitionTime":"2025-10-05T20:55:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:55 crc kubenswrapper[4754]: I1005 20:55:55.075724 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:55 crc kubenswrapper[4754]: I1005 20:55:55.075804 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:55 crc kubenswrapper[4754]: I1005 20:55:55.075823 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:55 crc kubenswrapper[4754]: I1005 20:55:55.075852 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:55 crc kubenswrapper[4754]: I1005 20:55:55.075870 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:55Z","lastTransitionTime":"2025-10-05T20:55:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:55 crc kubenswrapper[4754]: I1005 20:55:55.179625 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:55 crc kubenswrapper[4754]: I1005 20:55:55.179734 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:55 crc kubenswrapper[4754]: I1005 20:55:55.179760 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:55 crc kubenswrapper[4754]: I1005 20:55:55.179795 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:55 crc kubenswrapper[4754]: I1005 20:55:55.179817 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:55Z","lastTransitionTime":"2025-10-05T20:55:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:55 crc kubenswrapper[4754]: I1005 20:55:55.283307 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:55 crc kubenswrapper[4754]: I1005 20:55:55.283379 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:55 crc kubenswrapper[4754]: I1005 20:55:55.283400 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:55 crc kubenswrapper[4754]: I1005 20:55:55.283433 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:55 crc kubenswrapper[4754]: I1005 20:55:55.283461 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:55Z","lastTransitionTime":"2025-10-05T20:55:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:55 crc kubenswrapper[4754]: I1005 20:55:55.387027 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:55 crc kubenswrapper[4754]: I1005 20:55:55.387079 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:55 crc kubenswrapper[4754]: I1005 20:55:55.387095 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:55 crc kubenswrapper[4754]: I1005 20:55:55.387119 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:55 crc kubenswrapper[4754]: I1005 20:55:55.387137 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:55Z","lastTransitionTime":"2025-10-05T20:55:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:55 crc kubenswrapper[4754]: I1005 20:55:55.455715 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7rhps_b44790fb-fecb-4ec3-9816-8e23dd72ab33/ovnkube-controller/3.log" Oct 05 20:55:55 crc kubenswrapper[4754]: I1005 20:55:55.456644 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7rhps_b44790fb-fecb-4ec3-9816-8e23dd72ab33/ovnkube-controller/2.log" Oct 05 20:55:55 crc kubenswrapper[4754]: I1005 20:55:55.462106 4754 generic.go:334] "Generic (PLEG): container finished" podID="b44790fb-fecb-4ec3-9816-8e23dd72ab33" containerID="7c990164cb7213726a9086a630586e38da00d62fdb57aed8418bbec265a54705" exitCode=1 Oct 05 20:55:55 crc kubenswrapper[4754]: I1005 20:55:55.462186 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" event={"ID":"b44790fb-fecb-4ec3-9816-8e23dd72ab33","Type":"ContainerDied","Data":"7c990164cb7213726a9086a630586e38da00d62fdb57aed8418bbec265a54705"} Oct 05 20:55:55 crc kubenswrapper[4754]: I1005 20:55:55.462253 4754 scope.go:117] "RemoveContainer" containerID="af1c313e0dc0fe3b48feac3cce1ecd4452c9efadf3d19f257c07cc7fcdec875f" Oct 05 20:55:55 crc kubenswrapper[4754]: I1005 20:55:55.467098 4754 scope.go:117] "RemoveContainer" containerID="7c990164cb7213726a9086a630586e38da00d62fdb57aed8418bbec265a54705" Oct 05 20:55:55 crc kubenswrapper[4754]: E1005 20:55:55.468183 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-7rhps_openshift-ovn-kubernetes(b44790fb-fecb-4ec3-9816-8e23dd72ab33)\"" pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" podUID="b44790fb-fecb-4ec3-9816-8e23dd72ab33" Oct 05 20:55:55 crc kubenswrapper[4754]: I1005 20:55:55.490857 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:55 crc kubenswrapper[4754]: I1005 20:55:55.490924 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:55 crc kubenswrapper[4754]: I1005 20:55:55.490969 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:55 crc kubenswrapper[4754]: I1005 20:55:55.490999 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:55 crc kubenswrapper[4754]: I1005 20:55:55.491020 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:55Z","lastTransitionTime":"2025-10-05T20:55:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:55 crc kubenswrapper[4754]: I1005 20:55:55.491654 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7506d496-65ff-41dc-8230-24c8ee653a2a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33da5a6bb3cd3b042c193bc571e6622f2f87f27e560cdd0bbb9be31e2010d707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df6d700a947d19c84b4f65fd2dd2f8d1f4f1f80fa9b4800c2acfa8dae3912639\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35ca8ee77a0e58cf969bd67b277a3392614c4e73b963e6f6002097709370715d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e32c2154f46ba10c509fe40c357304aa8e82d04e9906bbaf2ccecd47b42279e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8907b9b69b249c34ec6db1e460d641b31711b75d6d31d4caafd867339e6571d0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1005 20:54:52.589115 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 20:54:52.590181 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1304624737/tls.crt::/tmp/serving-cert-1304624737/tls.key\\\\\\\"\\\\nI1005 20:54:58.070544 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1005 20:54:58.090352 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1005 20:54:58.090382 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1005 20:54:58.090412 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1005 20:54:58.090418 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1005 20:54:58.144656 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1005 20:54:58.144688 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 20:54:58.144695 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 20:54:58.144699 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1005 20:54:58.144702 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1005 20:54:58.144705 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1005 20:54:58.144708 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1005 20:54:58.145024 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1005 20:54:58.148112 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4efb34f3493cc8854bc4cff6b29f63973c52b754404d7a21daec4e1f612d327d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54c568b637f3737667433047b956c7eb22fdf0f9301acdfd3b49fb84c39de6ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54c568b637f3737667433047b956c7eb22fdf0f9301acdfd3b49fb84c39de6ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:55Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:55 crc kubenswrapper[4754]: I1005 20:55:55.511335 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c76e6e4740035a6ca54191a770c4516e28101aae69afbb89739174c2f076ef5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9eededc0cb1af94fb7b5daba08785f6b077a85e9cb5ddd8b9f397b658194f61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:55Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:55 crc kubenswrapper[4754]: I1005 20:55:55.541094 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tskt7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a68e9c89-465b-4c01-a470-e21a480ee465\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba6edba14cf321b4cd2f7bb44067a3ecf49f1594e76c0197e807ff3daa45e9b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d14ec99610d2167b484bd7712af7bb6d993d26c74b2828da1c89314f16fda22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d14ec99610d2167b484bd7712af7bb6d993d26c74b2828da1c89314f16fda22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f6f08dd853b25881e2f3c66206aa80d0c493518284217780aa40ac4b69f0c0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f6f08dd853b25881e2f3c66206aa80d0c493518284217780aa40ac4b69f0c0b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://71026d00ec1a3d91b019deb3c9f43ee52fb05b8f2d6374e1a5b762c9e5c1b10d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71026d00ec1a3d91b019deb3c9f43ee52fb05b8f2d6374e1a5b762c9e5c1b10d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b9910d2470acb16ebb436d7030b3161061607cb2a5ef50d482d8d5c4f48a8ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b9910d2470acb16ebb436d7030b3161061607cb2a5ef50d482d8d5c4f48a8ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0a93b61eacc55111995e69c15f89a210b8aa15468ac9ac053b7cf6b007d301e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b0a93b61eacc55111995e69c15f89a210b8aa15468ac9ac053b7cf6b007d301e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cafb3208e7e3bcdd01898a68fc5bbf4cb52ca0800b879b77a60bb102794257f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cafb3208e7e3bcdd01898a68fc5bbf4cb52ca0800b879b77a60bb102794257f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tskt7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:55Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:55 crc kubenswrapper[4754]: I1005 20:55:55.563417 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be95a413-6a59-45b4-84b7-b43cae694a26\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10b211cf6e47dbd1ee42c82e49de0a13a45581b92a0c4f6bbba0326c7149e41d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2ggg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4bb8c5903e8b70dd19eebb111097d96f725e17de44e06c3e74c067376f6c7360\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2ggg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-b2h9k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:55Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:55 crc kubenswrapper[4754]: I1005 20:55:55.584278 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rddkl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"15afc2a1-5cb7-41c6-954b-446c36822f50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://866d85ee126f8641d12f1a05fabc3de7b9392c324e9fe06ea92214155c05334c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqv8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a175565f83cfcba7825006eafb0f83efb70ec3d415c2c039c6499d5f6576e89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqv8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:55:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rddkl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:55Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:55 crc kubenswrapper[4754]: I1005 20:55:55.594648 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:55 crc kubenswrapper[4754]: I1005 20:55:55.594857 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:55 crc kubenswrapper[4754]: I1005 20:55:55.595000 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:55 crc kubenswrapper[4754]: I1005 20:55:55.595185 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:55 crc kubenswrapper[4754]: I1005 20:55:55.595387 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:55Z","lastTransitionTime":"2025-10-05T20:55:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:55 crc kubenswrapper[4754]: I1005 20:55:55.602181 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"860bd54f-d440-4642-bbe1-740e500c7551\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be822fa37e7f0aabf2715000b5a12edf3fe5c3df2999739551d2a89242c18129\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6482afbcf8935223f52760fa7bf9ab57e151e8fe9999f0ca43fc6db9f9f7b6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6482afbcf8935223f52760fa7bf9ab57e151e8fe9999f0ca43fc6db9f9f7b6f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:36Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:55Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:55 crc kubenswrapper[4754]: I1005 20:55:55.622277 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ae81d5f-5e31-4648-a80c-001034f0acf8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02dec08702aeaa03bc26151781909a4c6226fdaf37b921feb2a53ad6fa3c94c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a0c1914423693dd5ca96eddab941162a0f1e9efaf4bd02c47514f092e39e9e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5d1820b7c2ba7c5262f52297111f58e8f05771f3d7e34a59f8f3a05a7fb0bfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c187ac7d3514f4e68984ee98c274c53418df1818159cc9d6c16626d162fa3f97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c187ac7d3514f4e68984ee98c274c53418df1818159cc9d6c16626d162fa3f97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:36Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:55Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:55 crc kubenswrapper[4754]: I1005 20:55:55.649976 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e380f9b2a0e319dc325c54d1247a6db98346292548fa688cb41a2ca1b3048419\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:55Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:55 crc kubenswrapper[4754]: I1005 20:55:55.670826 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:55Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:55 crc kubenswrapper[4754]: I1005 20:55:55.699306 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:55 crc kubenswrapper[4754]: I1005 20:55:55.699396 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:55 crc kubenswrapper[4754]: I1005 20:55:55.699418 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:55 crc kubenswrapper[4754]: I1005 20:55:55.699448 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:55 crc kubenswrapper[4754]: I1005 20:55:55.699470 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:55Z","lastTransitionTime":"2025-10-05T20:55:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:55 crc kubenswrapper[4754]: I1005 20:55:55.720705 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9309179-b92a-4fa4-9fba-747b8f911936\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d969d3d50d6ad280fc6654df9441bab238bb538f277ed6d1273779813ffbaaed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://800a731ed8ef57714b2318af14054ef23f0368936630887a093fc77167b33d12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8cd9b08e83dc174413cc37891c14562816787028471a4201680c86dd620b7a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93b729f3f7f993db0d6bc3ba60c04930aada3cef8272ffcc3c6a8a242a9b1475\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dde1519111fc3af5d14463fdb58d4f5c62a7045e7de98c60981c7e4da37b0e7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ea160a5b37aea7f2970ae4dd2614b09d3dcf5dcbd1fc103b9997f8ecc3a804c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ea160a5b37aea7f2970ae4dd2614b09d3dcf5dcbd1fc103b9997f8ecc3a804c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://17ae6aa6c53e7dd061f70b41428ef50bbc7ffcc65352fad1a4a9d3c69dbd2ba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17ae6aa6c53e7dd061f70b41428ef50bbc7ffcc65352fad1a4a9d3c69dbd2ba9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://ad4cd69df7608c7d87d839fe09d4557c674c7efbc4730f4e4d3daaf9a26f5aee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad4cd69df7608c7d87d839fe09d4557c674c7efbc4730f4e4d3daaf9a26f5aee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:36Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:55Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:55 crc kubenswrapper[4754]: I1005 20:55:55.743683 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:55Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:55 crc kubenswrapper[4754]: I1005 20:55:55.777034 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:55 crc kubenswrapper[4754]: I1005 20:55:55.777104 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:55 crc kubenswrapper[4754]: I1005 20:55:55.777133 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:55 crc kubenswrapper[4754]: I1005 20:55:55.777163 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:55 crc kubenswrapper[4754]: I1005 20:55:55.777183 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:55Z","lastTransitionTime":"2025-10-05T20:55:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:55 crc kubenswrapper[4754]: I1005 20:55:55.783571 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b44790fb-fecb-4ec3-9816-8e23dd72ab33\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a7dc53e7f96e48213d724a2cb4859514df925005644108ae3a0c33b14dcf52d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfdd8b1e542ab676576c311e2923fb02861f92de91fe1f4b3e4b968c549c375d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3a63451ffaa14cacc3b78d11340f4ea81445fa3af96178e9b172d063d9697d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3cfd3f2274a8a46fa36660e94ac6c9c298c791283f12d52330c5594ac131998\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f626cb108716da0b595844ea0251b7a04d49b5908a1d6dd0aef5cb2d402c4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://822223f515b5add6d71d193fd5bc77df029b8d37a1cc07be4145b53c7226bf61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c990164cb7213726a9086a630586e38da00d62fdb57aed8418bbec265a54705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af1c313e0dc0fe3b48feac3cce1ecd4452c9efadf3d19f257c07cc7fcdec875f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-05T20:55:26Z\\\",\\\"message\\\":\\\"5:26.770640 6339 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/iptables-alerter-4ln5h in node crc\\\\nI1005 20:55:26.770478 6339 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-apiserver-operator/metrics]} name:Service_openshift-apiserver-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.38:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {8b82f026-5975-4a1b-bb18-08d5d51147ec}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1005 20:55:26.770652 6339 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/iptables-alerter-4ln5h after 0 failed attempt(s)\\\\nI1005 20:55:26.770662 6339 default_network_controller.go:776] Recording success event on pod openshift-network-operator/iptables-alerter-4ln5h\\\\nF1005 20:55:26.770677 6339 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controlle\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:25Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c990164cb7213726a9086a630586e38da00d62fdb57aed8418bbec265a54705\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-05T20:55:54Z\\\",\\\"message\\\":\\\"ient.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-api/machine-api-operator]} name:Service_openshift-machine-api/machine-api-operator_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.21:8443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {58a148b3-0a7b-4412-b447-f87788c4883f}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1005 20:55:54.809155 6675 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: failed to add event handler: handler {0x1e60340 0x1e60020 0x1e5ffc0} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:54Z is after 2025-08-24\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2acf9637fa25e5d3ffeb05224c11218c47defc679568fb928a498869613de4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35ebeb7a87d5d47b6567917d38eca38a3e622c497fba58971296c4609c287423\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35ebeb7a87d5d47b6567917d38eca38a3e622c497fba58971296c4609c287423\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7rhps\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:55Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:55 crc kubenswrapper[4754]: E1005 20:55:55.802864 4754 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T20:55:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T20:55:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T20:55:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T20:55:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ede2b442-76b0-443a-a450-252ee8711c80\\\",\\\"systemUUID\\\":\\\"72d081dd-fe7e-42a6-9cf0-f0ee7f733ac9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:55Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:55 crc kubenswrapper[4754]: I1005 20:55:55.804119 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-nwrnt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7dc728ea-2601-44f0-bcab-2913f034007d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-846xq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-846xq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:55:13Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-nwrnt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:55Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:55 crc kubenswrapper[4754]: I1005 20:55:55.808590 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:55 crc kubenswrapper[4754]: I1005 20:55:55.808814 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:55 crc kubenswrapper[4754]: I1005 20:55:55.808959 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:55 crc kubenswrapper[4754]: I1005 20:55:55.809113 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:55 crc kubenswrapper[4754]: I1005 20:55:55.809296 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:55Z","lastTransitionTime":"2025-10-05T20:55:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:55 crc kubenswrapper[4754]: I1005 20:55:55.823017 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"934bcd66-9c03-48b9-82dd-5093338f9c8f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9042df2f43243d3d79ba9a3940f8dd4f6c7a89ef9e3510e81fe60b1944d7e83b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5366bbf46a91667355f43f5df7335a536b395cda788b90b35ef18e2b53bc4c02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://600bc3bb551b747cf2f18e26745883232d7a37ac5877fad0c67100e52d7a43aa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78f74982180048eb50b9be79fa72c04f62f925b54fe511920f37e6407f864631\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:55Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:55 crc kubenswrapper[4754]: E1005 20:55:55.834721 4754 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T20:55:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T20:55:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T20:55:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T20:55:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ede2b442-76b0-443a-a450-252ee8711c80\\\",\\\"systemUUID\\\":\\\"72d081dd-fe7e-42a6-9cf0-f0ee7f733ac9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:55Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:55 crc kubenswrapper[4754]: I1005 20:55:55.836855 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nwrnt" Oct 05 20:55:55 crc kubenswrapper[4754]: E1005 20:55:55.837273 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nwrnt" podUID="7dc728ea-2601-44f0-bcab-2913f034007d" Oct 05 20:55:55 crc kubenswrapper[4754]: I1005 20:55:55.840773 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:55 crc kubenswrapper[4754]: I1005 20:55:55.840831 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:55 crc kubenswrapper[4754]: I1005 20:55:55.840851 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:55 crc kubenswrapper[4754]: I1005 20:55:55.840879 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:55 crc kubenswrapper[4754]: I1005 20:55:55.840899 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:55Z","lastTransitionTime":"2025-10-05T20:55:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:55 crc kubenswrapper[4754]: I1005 20:55:55.841320 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:55Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:55 crc kubenswrapper[4754]: E1005 20:55:55.856843 4754 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T20:55:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T20:55:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T20:55:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T20:55:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ede2b442-76b0-443a-a450-252ee8711c80\\\",\\\"systemUUID\\\":\\\"72d081dd-fe7e-42a6-9cf0-f0ee7f733ac9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:55Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:55 crc kubenswrapper[4754]: I1005 20:55:55.867385 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:55 crc kubenswrapper[4754]: I1005 20:55:55.867555 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:55 crc kubenswrapper[4754]: I1005 20:55:55.867583 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:55 crc kubenswrapper[4754]: I1005 20:55:55.867704 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:55 crc kubenswrapper[4754]: I1005 20:55:55.867727 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:55Z","lastTransitionTime":"2025-10-05T20:55:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:55 crc kubenswrapper[4754]: I1005 20:55:55.869833 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9ad070bb699b643d29b82bcc1126ce3b585f7492f5ae2817fa491b309b1691d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:55Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:55 crc kubenswrapper[4754]: I1005 20:55:55.884581 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-57sr4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e854c7b9-1229-4da4-80e9-fea6479459b8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceb577e1055d84007936a50d0b3d54b9ae0299a1fc1a6618b37163a6d3fbb9cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vw2ps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:58Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-57sr4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:55Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:55 crc kubenswrapper[4754]: E1005 20:55:55.895480 4754 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T20:55:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T20:55:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T20:55:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T20:55:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ede2b442-76b0-443a-a450-252ee8711c80\\\",\\\"systemUUID\\\":\\\"72d081dd-fe7e-42a6-9cf0-f0ee7f733ac9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:55Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:55 crc kubenswrapper[4754]: I1005 20:55:55.902483 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:55 crc kubenswrapper[4754]: I1005 20:55:55.902610 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:55 crc kubenswrapper[4754]: I1005 20:55:55.902636 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:55 crc kubenswrapper[4754]: I1005 20:55:55.902664 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:55 crc kubenswrapper[4754]: I1005 20:55:55.902692 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:55Z","lastTransitionTime":"2025-10-05T20:55:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:55 crc kubenswrapper[4754]: I1005 20:55:55.907983 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cn76s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02d5e3f9-73c2-4496-9aca-0787184aef19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ed23b9cd9b57bb169a8ebbf16b7ff8d196eb14330df0edf2b9fcd2a8c9e2c88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7ba2b076839f4a5f586d2e6c4f4d7188254128b782bfd2180e22bf3ea5ffdd7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-05T20:55:46Z\\\",\\\"message\\\":\\\"2025-10-05T20:55:01+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_d9d50ba4-5e54-4920-83d9-a3faa08e80be\\\\n2025-10-05T20:55:01+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_d9d50ba4-5e54-4920-83d9-a3faa08e80be to /host/opt/cni/bin/\\\\n2025-10-05T20:55:01Z [verbose] multus-daemon started\\\\n2025-10-05T20:55:01Z [verbose] Readiness Indicator file check\\\\n2025-10-05T20:55:46Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpx5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cn76s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:55Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:55 crc kubenswrapper[4754]: I1005 20:55:55.921604 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ss7lm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f9ef817-0622-4185-a7fb-57221690aa75\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ba613b0d7753a003c572c3733ed394707c45b6f8072cb7aa88f3c7312b8882b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wpwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:55:02Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ss7lm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:55Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:55 crc kubenswrapper[4754]: E1005 20:55:55.927315 4754 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T20:55:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T20:55:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T20:55:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T20:55:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ede2b442-76b0-443a-a450-252ee8711c80\\\",\\\"systemUUID\\\":\\\"72d081dd-fe7e-42a6-9cf0-f0ee7f733ac9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:55Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:55 crc kubenswrapper[4754]: E1005 20:55:55.927460 4754 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 05 20:55:55 crc kubenswrapper[4754]: I1005 20:55:55.929602 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:55 crc kubenswrapper[4754]: I1005 20:55:55.929649 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:55 crc kubenswrapper[4754]: I1005 20:55:55.929662 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:55 crc kubenswrapper[4754]: I1005 20:55:55.929685 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:55 crc kubenswrapper[4754]: I1005 20:55:55.929700 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:55Z","lastTransitionTime":"2025-10-05T20:55:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:56 crc kubenswrapper[4754]: I1005 20:55:56.033482 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:56 crc kubenswrapper[4754]: I1005 20:55:56.033633 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:56 crc kubenswrapper[4754]: I1005 20:55:56.033654 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:56 crc kubenswrapper[4754]: I1005 20:55:56.033688 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:56 crc kubenswrapper[4754]: I1005 20:55:56.033709 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:56Z","lastTransitionTime":"2025-10-05T20:55:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:56 crc kubenswrapper[4754]: I1005 20:55:56.137462 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:56 crc kubenswrapper[4754]: I1005 20:55:56.137566 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:56 crc kubenswrapper[4754]: I1005 20:55:56.137589 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:56 crc kubenswrapper[4754]: I1005 20:55:56.137619 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:56 crc kubenswrapper[4754]: I1005 20:55:56.137638 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:56Z","lastTransitionTime":"2025-10-05T20:55:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:56 crc kubenswrapper[4754]: I1005 20:55:56.241613 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:56 crc kubenswrapper[4754]: I1005 20:55:56.241714 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:56 crc kubenswrapper[4754]: I1005 20:55:56.241735 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:56 crc kubenswrapper[4754]: I1005 20:55:56.241765 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:56 crc kubenswrapper[4754]: I1005 20:55:56.241784 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:56Z","lastTransitionTime":"2025-10-05T20:55:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:56 crc kubenswrapper[4754]: I1005 20:55:56.344680 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:56 crc kubenswrapper[4754]: I1005 20:55:56.344761 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:56 crc kubenswrapper[4754]: I1005 20:55:56.344782 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:56 crc kubenswrapper[4754]: I1005 20:55:56.344816 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:56 crc kubenswrapper[4754]: I1005 20:55:56.344837 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:56Z","lastTransitionTime":"2025-10-05T20:55:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:56 crc kubenswrapper[4754]: I1005 20:55:56.449296 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:56 crc kubenswrapper[4754]: I1005 20:55:56.449361 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:56 crc kubenswrapper[4754]: I1005 20:55:56.449380 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:56 crc kubenswrapper[4754]: I1005 20:55:56.449408 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:56 crc kubenswrapper[4754]: I1005 20:55:56.449427 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:56Z","lastTransitionTime":"2025-10-05T20:55:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:56 crc kubenswrapper[4754]: I1005 20:55:56.469831 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7rhps_b44790fb-fecb-4ec3-9816-8e23dd72ab33/ovnkube-controller/3.log" Oct 05 20:55:56 crc kubenswrapper[4754]: I1005 20:55:56.476158 4754 scope.go:117] "RemoveContainer" containerID="7c990164cb7213726a9086a630586e38da00d62fdb57aed8418bbec265a54705" Oct 05 20:55:56 crc kubenswrapper[4754]: E1005 20:55:56.476422 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-7rhps_openshift-ovn-kubernetes(b44790fb-fecb-4ec3-9816-8e23dd72ab33)\"" pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" podUID="b44790fb-fecb-4ec3-9816-8e23dd72ab33" Oct 05 20:55:56 crc kubenswrapper[4754]: I1005 20:55:56.495661 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ss7lm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f9ef817-0622-4185-a7fb-57221690aa75\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ba613b0d7753a003c572c3733ed394707c45b6f8072cb7aa88f3c7312b8882b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wpwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:55:02Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ss7lm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:56Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:56 crc kubenswrapper[4754]: I1005 20:55:56.514366 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-nwrnt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7dc728ea-2601-44f0-bcab-2913f034007d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-846xq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-846xq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:55:13Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-nwrnt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:56Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:56 crc kubenswrapper[4754]: I1005 20:55:56.535619 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"934bcd66-9c03-48b9-82dd-5093338f9c8f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9042df2f43243d3d79ba9a3940f8dd4f6c7a89ef9e3510e81fe60b1944d7e83b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5366bbf46a91667355f43f5df7335a536b395cda788b90b35ef18e2b53bc4c02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://600bc3bb551b747cf2f18e26745883232d7a37ac5877fad0c67100e52d7a43aa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78f74982180048eb50b9be79fa72c04f62f925b54fe511920f37e6407f864631\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:56Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:56 crc kubenswrapper[4754]: I1005 20:55:56.552837 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:56 crc kubenswrapper[4754]: I1005 20:55:56.552912 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:56 crc kubenswrapper[4754]: I1005 20:55:56.552935 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:56 crc kubenswrapper[4754]: I1005 20:55:56.552967 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:56 crc kubenswrapper[4754]: I1005 20:55:56.552988 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:56Z","lastTransitionTime":"2025-10-05T20:55:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:56 crc kubenswrapper[4754]: I1005 20:55:56.562160 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:56Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:56 crc kubenswrapper[4754]: I1005 20:55:56.583008 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9ad070bb699b643d29b82bcc1126ce3b585f7492f5ae2817fa491b309b1691d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:56Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:56 crc kubenswrapper[4754]: I1005 20:55:56.610880 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-57sr4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e854c7b9-1229-4da4-80e9-fea6479459b8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceb577e1055d84007936a50d0b3d54b9ae0299a1fc1a6618b37163a6d3fbb9cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vw2ps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:58Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-57sr4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:56Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:56 crc kubenswrapper[4754]: I1005 20:55:56.630339 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cn76s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02d5e3f9-73c2-4496-9aca-0787184aef19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ed23b9cd9b57bb169a8ebbf16b7ff8d196eb14330df0edf2b9fcd2a8c9e2c88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7ba2b076839f4a5f586d2e6c4f4d7188254128b782bfd2180e22bf3ea5ffdd7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-05T20:55:46Z\\\",\\\"message\\\":\\\"2025-10-05T20:55:01+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_d9d50ba4-5e54-4920-83d9-a3faa08e80be\\\\n2025-10-05T20:55:01+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_d9d50ba4-5e54-4920-83d9-a3faa08e80be to /host/opt/cni/bin/\\\\n2025-10-05T20:55:01Z [verbose] multus-daemon started\\\\n2025-10-05T20:55:01Z [verbose] Readiness Indicator file check\\\\n2025-10-05T20:55:46Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpx5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cn76s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:56Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:56 crc kubenswrapper[4754]: I1005 20:55:56.646479 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7506d496-65ff-41dc-8230-24c8ee653a2a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33da5a6bb3cd3b042c193bc571e6622f2f87f27e560cdd0bbb9be31e2010d707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df6d700a947d19c84b4f65fd2dd2f8d1f4f1f80fa9b4800c2acfa8dae3912639\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35ca8ee77a0e58cf969bd67b277a3392614c4e73b963e6f6002097709370715d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e32c2154f46ba10c509fe40c357304aa8e82d04e9906bbaf2ccecd47b42279e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8907b9b69b249c34ec6db1e460d641b31711b75d6d31d4caafd867339e6571d0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1005 20:54:52.589115 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 20:54:52.590181 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1304624737/tls.crt::/tmp/serving-cert-1304624737/tls.key\\\\\\\"\\\\nI1005 20:54:58.070544 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1005 20:54:58.090352 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1005 20:54:58.090382 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1005 20:54:58.090412 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1005 20:54:58.090418 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1005 20:54:58.144656 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1005 20:54:58.144688 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 20:54:58.144695 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 20:54:58.144699 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1005 20:54:58.144702 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1005 20:54:58.144705 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1005 20:54:58.144708 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1005 20:54:58.145024 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1005 20:54:58.148112 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4efb34f3493cc8854bc4cff6b29f63973c52b754404d7a21daec4e1f612d327d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54c568b637f3737667433047b956c7eb22fdf0f9301acdfd3b49fb84c39de6ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54c568b637f3737667433047b956c7eb22fdf0f9301acdfd3b49fb84c39de6ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:56Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:56 crc kubenswrapper[4754]: I1005 20:55:56.656600 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:56 crc kubenswrapper[4754]: I1005 20:55:56.656662 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:56 crc kubenswrapper[4754]: I1005 20:55:56.656683 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:56 crc kubenswrapper[4754]: I1005 20:55:56.656711 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:56 crc kubenswrapper[4754]: I1005 20:55:56.656734 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:56Z","lastTransitionTime":"2025-10-05T20:55:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:56 crc kubenswrapper[4754]: I1005 20:55:56.667616 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c76e6e4740035a6ca54191a770c4516e28101aae69afbb89739174c2f076ef5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9eededc0cb1af94fb7b5daba08785f6b077a85e9cb5ddd8b9f397b658194f61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:56Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:56 crc kubenswrapper[4754]: I1005 20:55:56.693579 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tskt7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a68e9c89-465b-4c01-a470-e21a480ee465\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba6edba14cf321b4cd2f7bb44067a3ecf49f1594e76c0197e807ff3daa45e9b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d14ec99610d2167b484bd7712af7bb6d993d26c74b2828da1c89314f16fda22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d14ec99610d2167b484bd7712af7bb6d993d26c74b2828da1c89314f16fda22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f6f08dd853b25881e2f3c66206aa80d0c493518284217780aa40ac4b69f0c0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f6f08dd853b25881e2f3c66206aa80d0c493518284217780aa40ac4b69f0c0b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://71026d00ec1a3d91b019deb3c9f43ee52fb05b8f2d6374e1a5b762c9e5c1b10d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71026d00ec1a3d91b019deb3c9f43ee52fb05b8f2d6374e1a5b762c9e5c1b10d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b9910d2470acb16ebb436d7030b3161061607cb2a5ef50d482d8d5c4f48a8ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b9910d2470acb16ebb436d7030b3161061607cb2a5ef50d482d8d5c4f48a8ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0a93b61eacc55111995e69c15f89a210b8aa15468ac9ac053b7cf6b007d301e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b0a93b61eacc55111995e69c15f89a210b8aa15468ac9ac053b7cf6b007d301e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cafb3208e7e3bcdd01898a68fc5bbf4cb52ca0800b879b77a60bb102794257f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cafb3208e7e3bcdd01898a68fc5bbf4cb52ca0800b879b77a60bb102794257f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tskt7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:56Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:56 crc kubenswrapper[4754]: I1005 20:55:56.712572 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be95a413-6a59-45b4-84b7-b43cae694a26\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10b211cf6e47dbd1ee42c82e49de0a13a45581b92a0c4f6bbba0326c7149e41d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2ggg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4bb8c5903e8b70dd19eebb111097d96f725e17de44e06c3e74c067376f6c7360\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2ggg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-b2h9k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:56Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:56 crc kubenswrapper[4754]: I1005 20:55:56.735071 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rddkl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"15afc2a1-5cb7-41c6-954b-446c36822f50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://866d85ee126f8641d12f1a05fabc3de7b9392c324e9fe06ea92214155c05334c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqv8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a175565f83cfcba7825006eafb0f83efb70ec3d415c2c039c6499d5f6576e89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqv8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:55:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rddkl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:56Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:56 crc kubenswrapper[4754]: I1005 20:55:56.752137 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"860bd54f-d440-4642-bbe1-740e500c7551\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be822fa37e7f0aabf2715000b5a12edf3fe5c3df2999739551d2a89242c18129\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6482afbcf8935223f52760fa7bf9ab57e151e8fe9999f0ca43fc6db9f9f7b6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6482afbcf8935223f52760fa7bf9ab57e151e8fe9999f0ca43fc6db9f9f7b6f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:36Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:56Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:56 crc kubenswrapper[4754]: I1005 20:55:56.770944 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:56 crc kubenswrapper[4754]: I1005 20:55:56.771035 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:56 crc kubenswrapper[4754]: I1005 20:55:56.771063 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:56 crc kubenswrapper[4754]: I1005 20:55:56.771105 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:56 crc kubenswrapper[4754]: I1005 20:55:56.771135 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:56Z","lastTransitionTime":"2025-10-05T20:55:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:56 crc kubenswrapper[4754]: I1005 20:55:56.778479 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ae81d5f-5e31-4648-a80c-001034f0acf8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02dec08702aeaa03bc26151781909a4c6226fdaf37b921feb2a53ad6fa3c94c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a0c1914423693dd5ca96eddab941162a0f1e9efaf4bd02c47514f092e39e9e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5d1820b7c2ba7c5262f52297111f58e8f05771f3d7e34a59f8f3a05a7fb0bfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c187ac7d3514f4e68984ee98c274c53418df1818159cc9d6c16626d162fa3f97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c187ac7d3514f4e68984ee98c274c53418df1818159cc9d6c16626d162fa3f97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:36Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:56Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:56 crc kubenswrapper[4754]: I1005 20:55:56.798376 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e380f9b2a0e319dc325c54d1247a6db98346292548fa688cb41a2ca1b3048419\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:56Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:56 crc kubenswrapper[4754]: I1005 20:55:56.822816 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:56Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:56 crc kubenswrapper[4754]: I1005 20:55:56.837456 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 20:55:56 crc kubenswrapper[4754]: I1005 20:55:56.837508 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 20:55:56 crc kubenswrapper[4754]: E1005 20:55:56.837688 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 20:55:56 crc kubenswrapper[4754]: I1005 20:55:56.838050 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 20:55:56 crc kubenswrapper[4754]: E1005 20:55:56.838185 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 20:55:56 crc kubenswrapper[4754]: E1005 20:55:56.838432 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 20:55:56 crc kubenswrapper[4754]: I1005 20:55:56.857955 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9309179-b92a-4fa4-9fba-747b8f911936\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d969d3d50d6ad280fc6654df9441bab238bb538f277ed6d1273779813ffbaaed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://800a731ed8ef57714b2318af14054ef23f0368936630887a093fc77167b33d12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8cd9b08e83dc174413cc37891c14562816787028471a4201680c86dd620b7a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93b729f3f7f993db0d6bc3ba60c04930aada3cef8272ffcc3c6a8a242a9b1475\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dde1519111fc3af5d14463fdb58d4f5c62a7045e7de98c60981c7e4da37b0e7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ea160a5b37aea7f2970ae4dd2614b09d3dcf5dcbd1fc103b9997f8ecc3a804c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ea160a5b37aea7f2970ae4dd2614b09d3dcf5dcbd1fc103b9997f8ecc3a804c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://17ae6aa6c53e7dd061f70b41428ef50bbc7ffcc65352fad1a4a9d3c69dbd2ba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17ae6aa6c53e7dd061f70b41428ef50bbc7ffcc65352fad1a4a9d3c69dbd2ba9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://ad4cd69df7608c7d87d839fe09d4557c674c7efbc4730f4e4d3daaf9a26f5aee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad4cd69df7608c7d87d839fe09d4557c674c7efbc4730f4e4d3daaf9a26f5aee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:36Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:56Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:56 crc kubenswrapper[4754]: I1005 20:55:56.874879 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:56 crc kubenswrapper[4754]: I1005 20:55:56.874928 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:56 crc kubenswrapper[4754]: I1005 20:55:56.874966 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:56 crc kubenswrapper[4754]: I1005 20:55:56.874995 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:56 crc kubenswrapper[4754]: I1005 20:55:56.875018 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:56Z","lastTransitionTime":"2025-10-05T20:55:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:56 crc kubenswrapper[4754]: I1005 20:55:56.880058 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:56Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:56 crc kubenswrapper[4754]: I1005 20:55:56.907124 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b44790fb-fecb-4ec3-9816-8e23dd72ab33\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a7dc53e7f96e48213d724a2cb4859514df925005644108ae3a0c33b14dcf52d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfdd8b1e542ab676576c311e2923fb02861f92de91fe1f4b3e4b968c549c375d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3a63451ffaa14cacc3b78d11340f4ea81445fa3af96178e9b172d063d9697d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3cfd3f2274a8a46fa36660e94ac6c9c298c791283f12d52330c5594ac131998\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f626cb108716da0b595844ea0251b7a04d49b5908a1d6dd0aef5cb2d402c4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://822223f515b5add6d71d193fd5bc77df029b8d37a1cc07be4145b53c7226bf61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c990164cb7213726a9086a630586e38da00d62fdb57aed8418bbec265a54705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c990164cb7213726a9086a630586e38da00d62fdb57aed8418bbec265a54705\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-05T20:55:54Z\\\",\\\"message\\\":\\\"ient.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-api/machine-api-operator]} name:Service_openshift-machine-api/machine-api-operator_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.21:8443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {58a148b3-0a7b-4412-b447-f87788c4883f}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1005 20:55:54.809155 6675 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: failed to add event handler: handler {0x1e60340 0x1e60020 0x1e5ffc0} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:54Z is after 2025-08-24\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:53Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-7rhps_openshift-ovn-kubernetes(b44790fb-fecb-4ec3-9816-8e23dd72ab33)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2acf9637fa25e5d3ffeb05224c11218c47defc679568fb928a498869613de4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35ebeb7a87d5d47b6567917d38eca38a3e622c497fba58971296c4609c287423\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35ebeb7a87d5d47b6567917d38eca38a3e622c497fba58971296c4609c287423\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7rhps\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:56Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:56 crc kubenswrapper[4754]: I1005 20:55:56.945289 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9309179-b92a-4fa4-9fba-747b8f911936\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d969d3d50d6ad280fc6654df9441bab238bb538f277ed6d1273779813ffbaaed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://800a731ed8ef57714b2318af14054ef23f0368936630887a093fc77167b33d12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8cd9b08e83dc174413cc37891c14562816787028471a4201680c86dd620b7a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93b729f3f7f993db0d6bc3ba60c04930aada3cef8272ffcc3c6a8a242a9b1475\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dde1519111fc3af5d14463fdb58d4f5c62a7045e7de98c60981c7e4da37b0e7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ea160a5b37aea7f2970ae4dd2614b09d3dcf5dcbd1fc103b9997f8ecc3a804c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ea160a5b37aea7f2970ae4dd2614b09d3dcf5dcbd1fc103b9997f8ecc3a804c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://17ae6aa6c53e7dd061f70b41428ef50bbc7ffcc65352fad1a4a9d3c69dbd2ba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17ae6aa6c53e7dd061f70b41428ef50bbc7ffcc65352fad1a4a9d3c69dbd2ba9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://ad4cd69df7608c7d87d839fe09d4557c674c7efbc4730f4e4d3daaf9a26f5aee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad4cd69df7608c7d87d839fe09d4557c674c7efbc4730f4e4d3daaf9a26f5aee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:36Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:56Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:56 crc kubenswrapper[4754]: I1005 20:55:56.967937 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:56Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:56 crc kubenswrapper[4754]: I1005 20:55:56.978051 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:56 crc kubenswrapper[4754]: I1005 20:55:56.978113 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:56 crc kubenswrapper[4754]: I1005 20:55:56.978126 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:56 crc kubenswrapper[4754]: I1005 20:55:56.978145 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:56 crc kubenswrapper[4754]: I1005 20:55:56.978157 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:56Z","lastTransitionTime":"2025-10-05T20:55:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:56 crc kubenswrapper[4754]: I1005 20:55:56.995398 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b44790fb-fecb-4ec3-9816-8e23dd72ab33\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5a7dc53e7f96e48213d724a2cb4859514df925005644108ae3a0c33b14dcf52d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfdd8b1e542ab676576c311e2923fb02861f92de91fe1f4b3e4b968c549c375d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3a63451ffaa14cacc3b78d11340f4ea81445fa3af96178e9b172d063d9697d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3cfd3f2274a8a46fa36660e94ac6c9c298c791283f12d52330c5594ac131998\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f626cb108716da0b595844ea0251b7a04d49b5908a1d6dd0aef5cb2d402c4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://822223f515b5add6d71d193fd5bc77df029b8d37a1cc07be4145b53c7226bf61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c990164cb7213726a9086a630586e38da00d62fdb57aed8418bbec265a54705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c990164cb7213726a9086a630586e38da00d62fdb57aed8418bbec265a54705\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-05T20:55:54Z\\\",\\\"message\\\":\\\"ient.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-api/machine-api-operator]} name:Service_openshift-machine-api/machine-api-operator_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.21:8443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {58a148b3-0a7b-4412-b447-f87788c4883f}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1005 20:55:54.809155 6675 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: failed to add event handler: handler {0x1e60340 0x1e60020 0x1e5ffc0} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:54Z is after 2025-08-24\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:53Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-7rhps_openshift-ovn-kubernetes(b44790fb-fecb-4ec3-9816-8e23dd72ab33)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2acf9637fa25e5d3ffeb05224c11218c47defc679568fb928a498869613de4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35ebeb7a87d5d47b6567917d38eca38a3e622c497fba58971296c4609c287423\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35ebeb7a87d5d47b6567917d38eca38a3e622c497fba58971296c4609c287423\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gnzsc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7rhps\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:56Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:57 crc kubenswrapper[4754]: I1005 20:55:57.018120 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"934bcd66-9c03-48b9-82dd-5093338f9c8f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9042df2f43243d3d79ba9a3940f8dd4f6c7a89ef9e3510e81fe60b1944d7e83b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5366bbf46a91667355f43f5df7335a536b395cda788b90b35ef18e2b53bc4c02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://600bc3bb551b747cf2f18e26745883232d7a37ac5877fad0c67100e52d7a43aa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78f74982180048eb50b9be79fa72c04f62f925b54fe511920f37e6407f864631\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:57Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:57 crc kubenswrapper[4754]: I1005 20:55:57.036134 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:57Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:57 crc kubenswrapper[4754]: I1005 20:55:57.054349 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9ad070bb699b643d29b82bcc1126ce3b585f7492f5ae2817fa491b309b1691d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:57Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:57 crc kubenswrapper[4754]: I1005 20:55:57.068711 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-57sr4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e854c7b9-1229-4da4-80e9-fea6479459b8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceb577e1055d84007936a50d0b3d54b9ae0299a1fc1a6618b37163a6d3fbb9cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vw2ps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:58Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-57sr4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:57Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:57 crc kubenswrapper[4754]: I1005 20:55:57.083504 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:57 crc kubenswrapper[4754]: I1005 20:55:57.083581 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:57 crc kubenswrapper[4754]: I1005 20:55:57.083604 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:57 crc kubenswrapper[4754]: I1005 20:55:57.083801 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:57 crc kubenswrapper[4754]: I1005 20:55:57.083838 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:57Z","lastTransitionTime":"2025-10-05T20:55:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:57 crc kubenswrapper[4754]: I1005 20:55:57.093452 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cn76s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02d5e3f9-73c2-4496-9aca-0787184aef19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ed23b9cd9b57bb169a8ebbf16b7ff8d196eb14330df0edf2b9fcd2a8c9e2c88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7ba2b076839f4a5f586d2e6c4f4d7188254128b782bfd2180e22bf3ea5ffdd7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-05T20:55:46Z\\\",\\\"message\\\":\\\"2025-10-05T20:55:01+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_d9d50ba4-5e54-4920-83d9-a3faa08e80be\\\\n2025-10-05T20:55:01+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_d9d50ba4-5e54-4920-83d9-a3faa08e80be to /host/opt/cni/bin/\\\\n2025-10-05T20:55:01Z [verbose] multus-daemon started\\\\n2025-10-05T20:55:01Z [verbose] Readiness Indicator file check\\\\n2025-10-05T20:55:46Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpx5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cn76s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:57Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:57 crc kubenswrapper[4754]: I1005 20:55:57.109693 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ss7lm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f9ef817-0622-4185-a7fb-57221690aa75\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ba613b0d7753a003c572c3733ed394707c45b6f8072cb7aa88f3c7312b8882b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wpwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:55:02Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ss7lm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:57Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:57 crc kubenswrapper[4754]: I1005 20:55:57.126852 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-nwrnt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7dc728ea-2601-44f0-bcab-2913f034007d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:13Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-846xq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-846xq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:55:13Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-nwrnt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:57Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:57 crc kubenswrapper[4754]: I1005 20:55:57.150092 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7506d496-65ff-41dc-8230-24c8ee653a2a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33da5a6bb3cd3b042c193bc571e6622f2f87f27e560cdd0bbb9be31e2010d707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df6d700a947d19c84b4f65fd2dd2f8d1f4f1f80fa9b4800c2acfa8dae3912639\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35ca8ee77a0e58cf969bd67b277a3392614c4e73b963e6f6002097709370715d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e32c2154f46ba10c509fe40c357304aa8e82d04e9906bbaf2ccecd47b42279e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8907b9b69b249c34ec6db1e460d641b31711b75d6d31d4caafd867339e6571d0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1005 20:54:52.589115 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 20:54:52.590181 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1304624737/tls.crt::/tmp/serving-cert-1304624737/tls.key\\\\\\\"\\\\nI1005 20:54:58.070544 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1005 20:54:58.090352 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1005 20:54:58.090382 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1005 20:54:58.090412 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1005 20:54:58.090418 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1005 20:54:58.144656 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1005 20:54:58.144688 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 20:54:58.144695 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 20:54:58.144699 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1005 20:54:58.144702 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1005 20:54:58.144705 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1005 20:54:58.144708 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1005 20:54:58.145024 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1005 20:54:58.148112 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4efb34f3493cc8854bc4cff6b29f63973c52b754404d7a21daec4e1f612d327d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54c568b637f3737667433047b956c7eb22fdf0f9301acdfd3b49fb84c39de6ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54c568b637f3737667433047b956c7eb22fdf0f9301acdfd3b49fb84c39de6ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:57Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:57 crc kubenswrapper[4754]: I1005 20:55:57.187598 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:57 crc kubenswrapper[4754]: I1005 20:55:57.187662 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:57 crc kubenswrapper[4754]: I1005 20:55:57.187717 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:57 crc kubenswrapper[4754]: I1005 20:55:57.187749 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:57 crc kubenswrapper[4754]: I1005 20:55:57.187772 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:57Z","lastTransitionTime":"2025-10-05T20:55:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:57 crc kubenswrapper[4754]: I1005 20:55:57.200382 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c76e6e4740035a6ca54191a770c4516e28101aae69afbb89739174c2f076ef5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9eededc0cb1af94fb7b5daba08785f6b077a85e9cb5ddd8b9f397b658194f61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:57Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:57 crc kubenswrapper[4754]: I1005 20:55:57.220965 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tskt7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a68e9c89-465b-4c01-a470-e21a480ee465\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba6edba14cf321b4cd2f7bb44067a3ecf49f1594e76c0197e807ff3daa45e9b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d14ec99610d2167b484bd7712af7bb6d993d26c74b2828da1c89314f16fda22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d14ec99610d2167b484bd7712af7bb6d993d26c74b2828da1c89314f16fda22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f6f08dd853b25881e2f3c66206aa80d0c493518284217780aa40ac4b69f0c0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f6f08dd853b25881e2f3c66206aa80d0c493518284217780aa40ac4b69f0c0b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://71026d00ec1a3d91b019deb3c9f43ee52fb05b8f2d6374e1a5b762c9e5c1b10d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71026d00ec1a3d91b019deb3c9f43ee52fb05b8f2d6374e1a5b762c9e5c1b10d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b9910d2470acb16ebb436d7030b3161061607cb2a5ef50d482d8d5c4f48a8ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b9910d2470acb16ebb436d7030b3161061607cb2a5ef50d482d8d5c4f48a8ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0a93b61eacc55111995e69c15f89a210b8aa15468ac9ac053b7cf6b007d301e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b0a93b61eacc55111995e69c15f89a210b8aa15468ac9ac053b7cf6b007d301e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cafb3208e7e3bcdd01898a68fc5bbf4cb52ca0800b879b77a60bb102794257f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cafb3208e7e3bcdd01898a68fc5bbf4cb52ca0800b879b77a60bb102794257f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:55:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24dst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tskt7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:57Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:57 crc kubenswrapper[4754]: I1005 20:55:57.245462 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be95a413-6a59-45b4-84b7-b43cae694a26\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10b211cf6e47dbd1ee42c82e49de0a13a45581b92a0c4f6bbba0326c7149e41d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2ggg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4bb8c5903e8b70dd19eebb111097d96f725e17de44e06c3e74c067376f6c7360\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2ggg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-b2h9k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:57Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:57 crc kubenswrapper[4754]: I1005 20:55:57.266174 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rddkl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"15afc2a1-5cb7-41c6-954b-446c36822f50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://866d85ee126f8641d12f1a05fabc3de7b9392c324e9fe06ea92214155c05334c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqv8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a175565f83cfcba7825006eafb0f83efb70ec3d415c2c039c6499d5f6576e89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:55:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqv8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:55:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rddkl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:57Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:57 crc kubenswrapper[4754]: I1005 20:55:57.280383 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"860bd54f-d440-4642-bbe1-740e500c7551\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be822fa37e7f0aabf2715000b5a12edf3fe5c3df2999739551d2a89242c18129\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6482afbcf8935223f52760fa7bf9ab57e151e8fe9999f0ca43fc6db9f9f7b6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6482afbcf8935223f52760fa7bf9ab57e151e8fe9999f0ca43fc6db9f9f7b6f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:36Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:57Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:57 crc kubenswrapper[4754]: I1005 20:55:57.290167 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:57 crc kubenswrapper[4754]: I1005 20:55:57.290196 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:57 crc kubenswrapper[4754]: I1005 20:55:57.290207 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:57 crc kubenswrapper[4754]: I1005 20:55:57.290226 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:57 crc kubenswrapper[4754]: I1005 20:55:57.290239 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:57Z","lastTransitionTime":"2025-10-05T20:55:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:57 crc kubenswrapper[4754]: I1005 20:55:57.295429 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ae81d5f-5e31-4648-a80c-001034f0acf8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02dec08702aeaa03bc26151781909a4c6226fdaf37b921feb2a53ad6fa3c94c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a0c1914423693dd5ca96eddab941162a0f1e9efaf4bd02c47514f092e39e9e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5d1820b7c2ba7c5262f52297111f58e8f05771f3d7e34a59f8f3a05a7fb0bfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c187ac7d3514f4e68984ee98c274c53418df1818159cc9d6c16626d162fa3f97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c187ac7d3514f4e68984ee98c274c53418df1818159cc9d6c16626d162fa3f97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T20:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T20:54:38Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T20:54:36Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:57Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:57 crc kubenswrapper[4754]: I1005 20:55:57.316915 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:55:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e380f9b2a0e319dc325c54d1247a6db98346292548fa688cb41a2ca1b3048419\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T20:54:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:57Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:57 crc kubenswrapper[4754]: I1005 20:55:57.336875 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T20:54:58Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T20:55:57Z is after 2025-08-24T17:21:41Z" Oct 05 20:55:57 crc kubenswrapper[4754]: I1005 20:55:57.393962 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:57 crc kubenswrapper[4754]: I1005 20:55:57.394046 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:57 crc kubenswrapper[4754]: I1005 20:55:57.394065 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:57 crc kubenswrapper[4754]: I1005 20:55:57.394105 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:57 crc kubenswrapper[4754]: I1005 20:55:57.394130 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:57Z","lastTransitionTime":"2025-10-05T20:55:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:57 crc kubenswrapper[4754]: I1005 20:55:57.497848 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:57 crc kubenswrapper[4754]: I1005 20:55:57.497940 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:57 crc kubenswrapper[4754]: I1005 20:55:57.497966 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:57 crc kubenswrapper[4754]: I1005 20:55:57.497996 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:57 crc kubenswrapper[4754]: I1005 20:55:57.498023 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:57Z","lastTransitionTime":"2025-10-05T20:55:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:57 crc kubenswrapper[4754]: I1005 20:55:57.602150 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:57 crc kubenswrapper[4754]: I1005 20:55:57.602226 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:57 crc kubenswrapper[4754]: I1005 20:55:57.602251 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:57 crc kubenswrapper[4754]: I1005 20:55:57.602282 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:57 crc kubenswrapper[4754]: I1005 20:55:57.602303 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:57Z","lastTransitionTime":"2025-10-05T20:55:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:57 crc kubenswrapper[4754]: I1005 20:55:57.706538 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:57 crc kubenswrapper[4754]: I1005 20:55:57.706613 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:57 crc kubenswrapper[4754]: I1005 20:55:57.706631 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:57 crc kubenswrapper[4754]: I1005 20:55:57.706658 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:57 crc kubenswrapper[4754]: I1005 20:55:57.706681 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:57Z","lastTransitionTime":"2025-10-05T20:55:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:57 crc kubenswrapper[4754]: I1005 20:55:57.809867 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:57 crc kubenswrapper[4754]: I1005 20:55:57.809955 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:57 crc kubenswrapper[4754]: I1005 20:55:57.809975 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:57 crc kubenswrapper[4754]: I1005 20:55:57.810006 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:57 crc kubenswrapper[4754]: I1005 20:55:57.810025 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:57Z","lastTransitionTime":"2025-10-05T20:55:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:57 crc kubenswrapper[4754]: I1005 20:55:57.837067 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nwrnt" Oct 05 20:55:57 crc kubenswrapper[4754]: E1005 20:55:57.837281 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nwrnt" podUID="7dc728ea-2601-44f0-bcab-2913f034007d" Oct 05 20:55:57 crc kubenswrapper[4754]: I1005 20:55:57.912882 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:57 crc kubenswrapper[4754]: I1005 20:55:57.913003 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:57 crc kubenswrapper[4754]: I1005 20:55:57.913030 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:57 crc kubenswrapper[4754]: I1005 20:55:57.913062 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:57 crc kubenswrapper[4754]: I1005 20:55:57.913086 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:57Z","lastTransitionTime":"2025-10-05T20:55:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:58 crc kubenswrapper[4754]: I1005 20:55:58.015872 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:58 crc kubenswrapper[4754]: I1005 20:55:58.015945 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:58 crc kubenswrapper[4754]: I1005 20:55:58.015971 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:58 crc kubenswrapper[4754]: I1005 20:55:58.016001 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:58 crc kubenswrapper[4754]: I1005 20:55:58.016026 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:58Z","lastTransitionTime":"2025-10-05T20:55:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:58 crc kubenswrapper[4754]: I1005 20:55:58.120412 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:58 crc kubenswrapper[4754]: I1005 20:55:58.120485 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:58 crc kubenswrapper[4754]: I1005 20:55:58.120545 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:58 crc kubenswrapper[4754]: I1005 20:55:58.120583 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:58 crc kubenswrapper[4754]: I1005 20:55:58.120603 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:58Z","lastTransitionTime":"2025-10-05T20:55:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:58 crc kubenswrapper[4754]: I1005 20:55:58.224625 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:58 crc kubenswrapper[4754]: I1005 20:55:58.224703 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:58 crc kubenswrapper[4754]: I1005 20:55:58.224729 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:58 crc kubenswrapper[4754]: I1005 20:55:58.224756 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:58 crc kubenswrapper[4754]: I1005 20:55:58.224811 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:58Z","lastTransitionTime":"2025-10-05T20:55:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:58 crc kubenswrapper[4754]: I1005 20:55:58.333544 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:58 crc kubenswrapper[4754]: I1005 20:55:58.333655 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:58 crc kubenswrapper[4754]: I1005 20:55:58.333681 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:58 crc kubenswrapper[4754]: I1005 20:55:58.333715 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:58 crc kubenswrapper[4754]: I1005 20:55:58.333746 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:58Z","lastTransitionTime":"2025-10-05T20:55:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:58 crc kubenswrapper[4754]: I1005 20:55:58.437786 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:58 crc kubenswrapper[4754]: I1005 20:55:58.437855 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:58 crc kubenswrapper[4754]: I1005 20:55:58.437879 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:58 crc kubenswrapper[4754]: I1005 20:55:58.437912 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:58 crc kubenswrapper[4754]: I1005 20:55:58.437957 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:58Z","lastTransitionTime":"2025-10-05T20:55:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:58 crc kubenswrapper[4754]: I1005 20:55:58.542443 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:58 crc kubenswrapper[4754]: I1005 20:55:58.542563 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:58 crc kubenswrapper[4754]: I1005 20:55:58.542585 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:58 crc kubenswrapper[4754]: I1005 20:55:58.542616 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:58 crc kubenswrapper[4754]: I1005 20:55:58.542640 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:58Z","lastTransitionTime":"2025-10-05T20:55:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:58 crc kubenswrapper[4754]: I1005 20:55:58.647780 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:58 crc kubenswrapper[4754]: I1005 20:55:58.647868 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:58 crc kubenswrapper[4754]: I1005 20:55:58.647889 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:58 crc kubenswrapper[4754]: I1005 20:55:58.647922 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:58 crc kubenswrapper[4754]: I1005 20:55:58.647942 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:58Z","lastTransitionTime":"2025-10-05T20:55:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:58 crc kubenswrapper[4754]: I1005 20:55:58.751918 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:58 crc kubenswrapper[4754]: I1005 20:55:58.752017 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:58 crc kubenswrapper[4754]: I1005 20:55:58.752041 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:58 crc kubenswrapper[4754]: I1005 20:55:58.752077 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:58 crc kubenswrapper[4754]: I1005 20:55:58.752098 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:58Z","lastTransitionTime":"2025-10-05T20:55:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:58 crc kubenswrapper[4754]: I1005 20:55:58.837116 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 20:55:58 crc kubenswrapper[4754]: I1005 20:55:58.837270 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 20:55:58 crc kubenswrapper[4754]: I1005 20:55:58.837158 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 20:55:58 crc kubenswrapper[4754]: E1005 20:55:58.837440 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 20:55:58 crc kubenswrapper[4754]: E1005 20:55:58.837669 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 20:55:58 crc kubenswrapper[4754]: E1005 20:55:58.837785 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 20:55:58 crc kubenswrapper[4754]: I1005 20:55:58.855891 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:58 crc kubenswrapper[4754]: I1005 20:55:58.855958 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:58 crc kubenswrapper[4754]: I1005 20:55:58.855978 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:58 crc kubenswrapper[4754]: I1005 20:55:58.856005 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:58 crc kubenswrapper[4754]: I1005 20:55:58.856024 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:58Z","lastTransitionTime":"2025-10-05T20:55:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:58 crc kubenswrapper[4754]: I1005 20:55:58.959909 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:58 crc kubenswrapper[4754]: I1005 20:55:58.959974 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:58 crc kubenswrapper[4754]: I1005 20:55:58.959987 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:58 crc kubenswrapper[4754]: I1005 20:55:58.960013 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:58 crc kubenswrapper[4754]: I1005 20:55:58.960027 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:58Z","lastTransitionTime":"2025-10-05T20:55:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:59 crc kubenswrapper[4754]: I1005 20:55:59.063754 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:59 crc kubenswrapper[4754]: I1005 20:55:59.063856 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:59 crc kubenswrapper[4754]: I1005 20:55:59.063883 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:59 crc kubenswrapper[4754]: I1005 20:55:59.063915 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:59 crc kubenswrapper[4754]: I1005 20:55:59.063935 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:59Z","lastTransitionTime":"2025-10-05T20:55:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:59 crc kubenswrapper[4754]: I1005 20:55:59.167602 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:59 crc kubenswrapper[4754]: I1005 20:55:59.167682 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:59 crc kubenswrapper[4754]: I1005 20:55:59.167747 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:59 crc kubenswrapper[4754]: I1005 20:55:59.167829 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:59 crc kubenswrapper[4754]: I1005 20:55:59.167857 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:59Z","lastTransitionTime":"2025-10-05T20:55:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:59 crc kubenswrapper[4754]: I1005 20:55:59.272744 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:59 crc kubenswrapper[4754]: I1005 20:55:59.273608 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:59 crc kubenswrapper[4754]: I1005 20:55:59.273633 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:59 crc kubenswrapper[4754]: I1005 20:55:59.273668 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:59 crc kubenswrapper[4754]: I1005 20:55:59.273691 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:59Z","lastTransitionTime":"2025-10-05T20:55:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:59 crc kubenswrapper[4754]: I1005 20:55:59.376268 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:59 crc kubenswrapper[4754]: I1005 20:55:59.376812 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:59 crc kubenswrapper[4754]: I1005 20:55:59.377032 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:59 crc kubenswrapper[4754]: I1005 20:55:59.377179 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:59 crc kubenswrapper[4754]: I1005 20:55:59.377325 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:59Z","lastTransitionTime":"2025-10-05T20:55:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:59 crc kubenswrapper[4754]: I1005 20:55:59.480863 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:59 crc kubenswrapper[4754]: I1005 20:55:59.480913 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:59 crc kubenswrapper[4754]: I1005 20:55:59.480923 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:59 crc kubenswrapper[4754]: I1005 20:55:59.480945 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:59 crc kubenswrapper[4754]: I1005 20:55:59.480956 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:59Z","lastTransitionTime":"2025-10-05T20:55:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:59 crc kubenswrapper[4754]: I1005 20:55:59.584627 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:59 crc kubenswrapper[4754]: I1005 20:55:59.584695 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:59 crc kubenswrapper[4754]: I1005 20:55:59.584715 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:59 crc kubenswrapper[4754]: I1005 20:55:59.584741 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:59 crc kubenswrapper[4754]: I1005 20:55:59.584761 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:59Z","lastTransitionTime":"2025-10-05T20:55:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:59 crc kubenswrapper[4754]: I1005 20:55:59.688867 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:59 crc kubenswrapper[4754]: I1005 20:55:59.688920 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:59 crc kubenswrapper[4754]: I1005 20:55:59.688932 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:59 crc kubenswrapper[4754]: I1005 20:55:59.689007 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:59 crc kubenswrapper[4754]: I1005 20:55:59.689026 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:59Z","lastTransitionTime":"2025-10-05T20:55:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:59 crc kubenswrapper[4754]: I1005 20:55:59.792231 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:59 crc kubenswrapper[4754]: I1005 20:55:59.792276 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:59 crc kubenswrapper[4754]: I1005 20:55:59.792289 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:59 crc kubenswrapper[4754]: I1005 20:55:59.792309 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:59 crc kubenswrapper[4754]: I1005 20:55:59.792323 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:59Z","lastTransitionTime":"2025-10-05T20:55:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:59 crc kubenswrapper[4754]: I1005 20:55:59.836258 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nwrnt" Oct 05 20:55:59 crc kubenswrapper[4754]: E1005 20:55:59.836471 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nwrnt" podUID="7dc728ea-2601-44f0-bcab-2913f034007d" Oct 05 20:55:59 crc kubenswrapper[4754]: I1005 20:55:59.895875 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:59 crc kubenswrapper[4754]: I1005 20:55:59.895966 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:59 crc kubenswrapper[4754]: I1005 20:55:59.895987 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:59 crc kubenswrapper[4754]: I1005 20:55:59.896016 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:59 crc kubenswrapper[4754]: I1005 20:55:59.896043 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:59Z","lastTransitionTime":"2025-10-05T20:55:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:55:59 crc kubenswrapper[4754]: I1005 20:55:59.999124 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:55:59 crc kubenswrapper[4754]: I1005 20:55:59.999187 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:55:59 crc kubenswrapper[4754]: I1005 20:55:59.999204 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:55:59 crc kubenswrapper[4754]: I1005 20:55:59.999229 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:55:59 crc kubenswrapper[4754]: I1005 20:55:59.999249 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:55:59Z","lastTransitionTime":"2025-10-05T20:55:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:56:00 crc kubenswrapper[4754]: I1005 20:56:00.104880 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:56:00 crc kubenswrapper[4754]: I1005 20:56:00.104955 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:56:00 crc kubenswrapper[4754]: I1005 20:56:00.104973 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:56:00 crc kubenswrapper[4754]: I1005 20:56:00.105006 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:56:00 crc kubenswrapper[4754]: I1005 20:56:00.105040 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:56:00Z","lastTransitionTime":"2025-10-05T20:56:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:56:00 crc kubenswrapper[4754]: I1005 20:56:00.208575 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:56:00 crc kubenswrapper[4754]: I1005 20:56:00.208645 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:56:00 crc kubenswrapper[4754]: I1005 20:56:00.208665 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:56:00 crc kubenswrapper[4754]: I1005 20:56:00.208698 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:56:00 crc kubenswrapper[4754]: I1005 20:56:00.208718 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:56:00Z","lastTransitionTime":"2025-10-05T20:56:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:56:00 crc kubenswrapper[4754]: I1005 20:56:00.311478 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:56:00 crc kubenswrapper[4754]: I1005 20:56:00.311592 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:56:00 crc kubenswrapper[4754]: I1005 20:56:00.311624 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:56:00 crc kubenswrapper[4754]: I1005 20:56:00.311664 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:56:00 crc kubenswrapper[4754]: I1005 20:56:00.311689 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:56:00Z","lastTransitionTime":"2025-10-05T20:56:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:56:00 crc kubenswrapper[4754]: I1005 20:56:00.415001 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:56:00 crc kubenswrapper[4754]: I1005 20:56:00.415078 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:56:00 crc kubenswrapper[4754]: I1005 20:56:00.415102 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:56:00 crc kubenswrapper[4754]: I1005 20:56:00.415141 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:56:00 crc kubenswrapper[4754]: I1005 20:56:00.415167 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:56:00Z","lastTransitionTime":"2025-10-05T20:56:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:56:00 crc kubenswrapper[4754]: I1005 20:56:00.518206 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:56:00 crc kubenswrapper[4754]: I1005 20:56:00.518274 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:56:00 crc kubenswrapper[4754]: I1005 20:56:00.518293 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:56:00 crc kubenswrapper[4754]: I1005 20:56:00.518321 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:56:00 crc kubenswrapper[4754]: I1005 20:56:00.518340 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:56:00Z","lastTransitionTime":"2025-10-05T20:56:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:56:00 crc kubenswrapper[4754]: I1005 20:56:00.622672 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:56:00 crc kubenswrapper[4754]: I1005 20:56:00.622734 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:56:00 crc kubenswrapper[4754]: I1005 20:56:00.622753 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:56:00 crc kubenswrapper[4754]: I1005 20:56:00.622779 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:56:00 crc kubenswrapper[4754]: I1005 20:56:00.622800 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:56:00Z","lastTransitionTime":"2025-10-05T20:56:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:56:00 crc kubenswrapper[4754]: I1005 20:56:00.727323 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:56:00 crc kubenswrapper[4754]: I1005 20:56:00.727399 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:56:00 crc kubenswrapper[4754]: I1005 20:56:00.727421 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:56:00 crc kubenswrapper[4754]: I1005 20:56:00.727453 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:56:00 crc kubenswrapper[4754]: I1005 20:56:00.727473 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:56:00Z","lastTransitionTime":"2025-10-05T20:56:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:56:00 crc kubenswrapper[4754]: I1005 20:56:00.831559 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:56:00 crc kubenswrapper[4754]: I1005 20:56:00.831659 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:56:00 crc kubenswrapper[4754]: I1005 20:56:00.831688 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:56:00 crc kubenswrapper[4754]: I1005 20:56:00.831726 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:56:00 crc kubenswrapper[4754]: I1005 20:56:00.831777 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:56:00Z","lastTransitionTime":"2025-10-05T20:56:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:56:00 crc kubenswrapper[4754]: I1005 20:56:00.836820 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 20:56:00 crc kubenswrapper[4754]: E1005 20:56:00.837033 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 20:56:00 crc kubenswrapper[4754]: I1005 20:56:00.837346 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 20:56:00 crc kubenswrapper[4754]: E1005 20:56:00.837450 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 20:56:00 crc kubenswrapper[4754]: I1005 20:56:00.837691 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 20:56:00 crc kubenswrapper[4754]: E1005 20:56:00.837792 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 20:56:00 crc kubenswrapper[4754]: I1005 20:56:00.936397 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:56:00 crc kubenswrapper[4754]: I1005 20:56:00.937364 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:56:00 crc kubenswrapper[4754]: I1005 20:56:00.937539 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:56:00 crc kubenswrapper[4754]: I1005 20:56:00.937682 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:56:00 crc kubenswrapper[4754]: I1005 20:56:00.937833 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:56:00Z","lastTransitionTime":"2025-10-05T20:56:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:56:01 crc kubenswrapper[4754]: I1005 20:56:01.041486 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:56:01 crc kubenswrapper[4754]: I1005 20:56:01.041634 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:56:01 crc kubenswrapper[4754]: I1005 20:56:01.041663 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:56:01 crc kubenswrapper[4754]: I1005 20:56:01.041703 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:56:01 crc kubenswrapper[4754]: I1005 20:56:01.041727 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:56:01Z","lastTransitionTime":"2025-10-05T20:56:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:56:01 crc kubenswrapper[4754]: I1005 20:56:01.145391 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:56:01 crc kubenswrapper[4754]: I1005 20:56:01.145461 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:56:01 crc kubenswrapper[4754]: I1005 20:56:01.145481 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:56:01 crc kubenswrapper[4754]: I1005 20:56:01.145548 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:56:01 crc kubenswrapper[4754]: I1005 20:56:01.145568 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:56:01Z","lastTransitionTime":"2025-10-05T20:56:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:56:01 crc kubenswrapper[4754]: I1005 20:56:01.249071 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:56:01 crc kubenswrapper[4754]: I1005 20:56:01.249151 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:56:01 crc kubenswrapper[4754]: I1005 20:56:01.249170 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:56:01 crc kubenswrapper[4754]: I1005 20:56:01.249198 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:56:01 crc kubenswrapper[4754]: I1005 20:56:01.249220 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:56:01Z","lastTransitionTime":"2025-10-05T20:56:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:56:01 crc kubenswrapper[4754]: I1005 20:56:01.352246 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:56:01 crc kubenswrapper[4754]: I1005 20:56:01.352316 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:56:01 crc kubenswrapper[4754]: I1005 20:56:01.352333 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:56:01 crc kubenswrapper[4754]: I1005 20:56:01.352366 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:56:01 crc kubenswrapper[4754]: I1005 20:56:01.352384 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:56:01Z","lastTransitionTime":"2025-10-05T20:56:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:56:01 crc kubenswrapper[4754]: I1005 20:56:01.454892 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:56:01 crc kubenswrapper[4754]: I1005 20:56:01.454957 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:56:01 crc kubenswrapper[4754]: I1005 20:56:01.454978 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:56:01 crc kubenswrapper[4754]: I1005 20:56:01.455008 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:56:01 crc kubenswrapper[4754]: I1005 20:56:01.455029 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:56:01Z","lastTransitionTime":"2025-10-05T20:56:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:56:01 crc kubenswrapper[4754]: I1005 20:56:01.557937 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:56:01 crc kubenswrapper[4754]: I1005 20:56:01.558000 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:56:01 crc kubenswrapper[4754]: I1005 20:56:01.558019 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:56:01 crc kubenswrapper[4754]: I1005 20:56:01.558048 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:56:01 crc kubenswrapper[4754]: I1005 20:56:01.558067 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:56:01Z","lastTransitionTime":"2025-10-05T20:56:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:56:01 crc kubenswrapper[4754]: I1005 20:56:01.662057 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:56:01 crc kubenswrapper[4754]: I1005 20:56:01.662134 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:56:01 crc kubenswrapper[4754]: I1005 20:56:01.662153 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:56:01 crc kubenswrapper[4754]: I1005 20:56:01.662183 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:56:01 crc kubenswrapper[4754]: I1005 20:56:01.662202 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:56:01Z","lastTransitionTime":"2025-10-05T20:56:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:56:01 crc kubenswrapper[4754]: I1005 20:56:01.766854 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:56:01 crc kubenswrapper[4754]: I1005 20:56:01.766958 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:56:01 crc kubenswrapper[4754]: I1005 20:56:01.766987 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:56:01 crc kubenswrapper[4754]: I1005 20:56:01.767023 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:56:01 crc kubenswrapper[4754]: I1005 20:56:01.767049 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:56:01Z","lastTransitionTime":"2025-10-05T20:56:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:56:01 crc kubenswrapper[4754]: I1005 20:56:01.836788 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nwrnt" Oct 05 20:56:01 crc kubenswrapper[4754]: E1005 20:56:01.837180 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nwrnt" podUID="7dc728ea-2601-44f0-bcab-2913f034007d" Oct 05 20:56:01 crc kubenswrapper[4754]: I1005 20:56:01.870448 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:56:01 crc kubenswrapper[4754]: I1005 20:56:01.870589 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:56:01 crc kubenswrapper[4754]: I1005 20:56:01.870620 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:56:01 crc kubenswrapper[4754]: I1005 20:56:01.870660 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:56:01 crc kubenswrapper[4754]: I1005 20:56:01.870694 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:56:01Z","lastTransitionTime":"2025-10-05T20:56:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:56:01 crc kubenswrapper[4754]: I1005 20:56:01.974795 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:56:01 crc kubenswrapper[4754]: I1005 20:56:01.974866 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:56:01 crc kubenswrapper[4754]: I1005 20:56:01.974883 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:56:01 crc kubenswrapper[4754]: I1005 20:56:01.974909 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:56:01 crc kubenswrapper[4754]: I1005 20:56:01.974931 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:56:01Z","lastTransitionTime":"2025-10-05T20:56:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:56:02 crc kubenswrapper[4754]: I1005 20:56:02.078836 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:56:02 crc kubenswrapper[4754]: I1005 20:56:02.078934 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:56:02 crc kubenswrapper[4754]: I1005 20:56:02.078955 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:56:02 crc kubenswrapper[4754]: I1005 20:56:02.079012 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:56:02 crc kubenswrapper[4754]: I1005 20:56:02.079031 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:56:02Z","lastTransitionTime":"2025-10-05T20:56:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:56:02 crc kubenswrapper[4754]: I1005 20:56:02.183096 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:56:02 crc kubenswrapper[4754]: I1005 20:56:02.183198 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:56:02 crc kubenswrapper[4754]: I1005 20:56:02.183227 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:56:02 crc kubenswrapper[4754]: I1005 20:56:02.183260 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:56:02 crc kubenswrapper[4754]: I1005 20:56:02.183283 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:56:02Z","lastTransitionTime":"2025-10-05T20:56:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:56:02 crc kubenswrapper[4754]: I1005 20:56:02.286898 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:56:02 crc kubenswrapper[4754]: I1005 20:56:02.286978 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:56:02 crc kubenswrapper[4754]: I1005 20:56:02.286997 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:56:02 crc kubenswrapper[4754]: I1005 20:56:02.287026 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:56:02 crc kubenswrapper[4754]: I1005 20:56:02.287045 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:56:02Z","lastTransitionTime":"2025-10-05T20:56:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:56:02 crc kubenswrapper[4754]: I1005 20:56:02.391624 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:56:02 crc kubenswrapper[4754]: I1005 20:56:02.391706 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:56:02 crc kubenswrapper[4754]: I1005 20:56:02.391731 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:56:02 crc kubenswrapper[4754]: I1005 20:56:02.391764 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:56:02 crc kubenswrapper[4754]: I1005 20:56:02.391791 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:56:02Z","lastTransitionTime":"2025-10-05T20:56:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:56:02 crc kubenswrapper[4754]: I1005 20:56:02.495185 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:56:02 crc kubenswrapper[4754]: I1005 20:56:02.495256 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:56:02 crc kubenswrapper[4754]: I1005 20:56:02.495273 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:56:02 crc kubenswrapper[4754]: I1005 20:56:02.495300 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:56:02 crc kubenswrapper[4754]: I1005 20:56:02.495320 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:56:02Z","lastTransitionTime":"2025-10-05T20:56:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:56:02 crc kubenswrapper[4754]: I1005 20:56:02.514958 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 20:56:02 crc kubenswrapper[4754]: E1005 20:56:02.515225 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 20:57:06.515186006 +0000 UTC m=+150.419304746 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:56:02 crc kubenswrapper[4754]: I1005 20:56:02.599471 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:56:02 crc kubenswrapper[4754]: I1005 20:56:02.600092 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:56:02 crc kubenswrapper[4754]: I1005 20:56:02.600260 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:56:02 crc kubenswrapper[4754]: I1005 20:56:02.600474 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:56:02 crc kubenswrapper[4754]: I1005 20:56:02.600773 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:56:02Z","lastTransitionTime":"2025-10-05T20:56:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:56:02 crc kubenswrapper[4754]: I1005 20:56:02.616304 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 20:56:02 crc kubenswrapper[4754]: I1005 20:56:02.616389 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 20:56:02 crc kubenswrapper[4754]: I1005 20:56:02.616452 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 20:56:02 crc kubenswrapper[4754]: E1005 20:56:02.616473 4754 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 05 20:56:02 crc kubenswrapper[4754]: I1005 20:56:02.616553 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 20:56:02 crc kubenswrapper[4754]: E1005 20:56:02.616621 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-05 20:57:06.616585616 +0000 UTC m=+150.520704366 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 05 20:56:02 crc kubenswrapper[4754]: E1005 20:56:02.616787 4754 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 05 20:56:02 crc kubenswrapper[4754]: E1005 20:56:02.616941 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-05 20:57:06.616883183 +0000 UTC m=+150.521001923 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 05 20:56:02 crc kubenswrapper[4754]: E1005 20:56:02.616791 4754 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 05 20:56:02 crc kubenswrapper[4754]: E1005 20:56:02.617033 4754 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 05 20:56:02 crc kubenswrapper[4754]: E1005 20:56:02.617066 4754 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 05 20:56:02 crc kubenswrapper[4754]: E1005 20:56:02.616791 4754 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 05 20:56:02 crc kubenswrapper[4754]: E1005 20:56:02.617199 4754 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 05 20:56:02 crc kubenswrapper[4754]: E1005 20:56:02.617133 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-05 20:57:06.617113309 +0000 UTC m=+150.521232059 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 05 20:56:02 crc kubenswrapper[4754]: E1005 20:56:02.617228 4754 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 05 20:56:02 crc kubenswrapper[4754]: E1005 20:56:02.617326 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-05 20:57:06.617301514 +0000 UTC m=+150.521420474 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 05 20:56:02 crc kubenswrapper[4754]: I1005 20:56:02.707427 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:56:02 crc kubenswrapper[4754]: I1005 20:56:02.708219 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:56:02 crc kubenswrapper[4754]: I1005 20:56:02.708247 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:56:02 crc kubenswrapper[4754]: I1005 20:56:02.708324 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:56:02 crc kubenswrapper[4754]: I1005 20:56:02.708350 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:56:02Z","lastTransitionTime":"2025-10-05T20:56:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:56:02 crc kubenswrapper[4754]: I1005 20:56:02.813390 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:56:02 crc kubenswrapper[4754]: I1005 20:56:02.813467 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:56:02 crc kubenswrapper[4754]: I1005 20:56:02.813487 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:56:02 crc kubenswrapper[4754]: I1005 20:56:02.813558 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:56:02 crc kubenswrapper[4754]: I1005 20:56:02.813580 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:56:02Z","lastTransitionTime":"2025-10-05T20:56:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:56:02 crc kubenswrapper[4754]: I1005 20:56:02.836332 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 20:56:02 crc kubenswrapper[4754]: I1005 20:56:02.836347 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 20:56:02 crc kubenswrapper[4754]: I1005 20:56:02.836353 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 20:56:02 crc kubenswrapper[4754]: E1005 20:56:02.836641 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 20:56:02 crc kubenswrapper[4754]: E1005 20:56:02.836783 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 20:56:02 crc kubenswrapper[4754]: E1005 20:56:02.837066 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 20:56:02 crc kubenswrapper[4754]: I1005 20:56:02.916879 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:56:02 crc kubenswrapper[4754]: I1005 20:56:02.916936 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:56:02 crc kubenswrapper[4754]: I1005 20:56:02.916954 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:56:02 crc kubenswrapper[4754]: I1005 20:56:02.916981 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:56:02 crc kubenswrapper[4754]: I1005 20:56:02.917003 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:56:02Z","lastTransitionTime":"2025-10-05T20:56:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:56:03 crc kubenswrapper[4754]: I1005 20:56:03.021577 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:56:03 crc kubenswrapper[4754]: I1005 20:56:03.021644 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:56:03 crc kubenswrapper[4754]: I1005 20:56:03.021666 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:56:03 crc kubenswrapper[4754]: I1005 20:56:03.021695 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:56:03 crc kubenswrapper[4754]: I1005 20:56:03.021712 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:56:03Z","lastTransitionTime":"2025-10-05T20:56:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:56:03 crc kubenswrapper[4754]: I1005 20:56:03.124915 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:56:03 crc kubenswrapper[4754]: I1005 20:56:03.124986 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:56:03 crc kubenswrapper[4754]: I1005 20:56:03.125005 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:56:03 crc kubenswrapper[4754]: I1005 20:56:03.125033 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:56:03 crc kubenswrapper[4754]: I1005 20:56:03.125052 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:56:03Z","lastTransitionTime":"2025-10-05T20:56:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:56:03 crc kubenswrapper[4754]: I1005 20:56:03.227444 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:56:03 crc kubenswrapper[4754]: I1005 20:56:03.227536 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:56:03 crc kubenswrapper[4754]: I1005 20:56:03.227556 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:56:03 crc kubenswrapper[4754]: I1005 20:56:03.227583 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:56:03 crc kubenswrapper[4754]: I1005 20:56:03.227603 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:56:03Z","lastTransitionTime":"2025-10-05T20:56:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:56:03 crc kubenswrapper[4754]: I1005 20:56:03.331176 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:56:03 crc kubenswrapper[4754]: I1005 20:56:03.331245 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:56:03 crc kubenswrapper[4754]: I1005 20:56:03.331264 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:56:03 crc kubenswrapper[4754]: I1005 20:56:03.331293 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:56:03 crc kubenswrapper[4754]: I1005 20:56:03.331316 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:56:03Z","lastTransitionTime":"2025-10-05T20:56:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:56:03 crc kubenswrapper[4754]: I1005 20:56:03.434745 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:56:03 crc kubenswrapper[4754]: I1005 20:56:03.434817 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:56:03 crc kubenswrapper[4754]: I1005 20:56:03.434836 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:56:03 crc kubenswrapper[4754]: I1005 20:56:03.434863 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:56:03 crc kubenswrapper[4754]: I1005 20:56:03.434881 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:56:03Z","lastTransitionTime":"2025-10-05T20:56:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:56:03 crc kubenswrapper[4754]: I1005 20:56:03.537602 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:56:03 crc kubenswrapper[4754]: I1005 20:56:03.537679 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:56:03 crc kubenswrapper[4754]: I1005 20:56:03.537698 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:56:03 crc kubenswrapper[4754]: I1005 20:56:03.537725 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:56:03 crc kubenswrapper[4754]: I1005 20:56:03.537746 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:56:03Z","lastTransitionTime":"2025-10-05T20:56:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:56:03 crc kubenswrapper[4754]: I1005 20:56:03.641171 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:56:03 crc kubenswrapper[4754]: I1005 20:56:03.641233 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:56:03 crc kubenswrapper[4754]: I1005 20:56:03.641291 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:56:03 crc kubenswrapper[4754]: I1005 20:56:03.641321 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:56:03 crc kubenswrapper[4754]: I1005 20:56:03.641370 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:56:03Z","lastTransitionTime":"2025-10-05T20:56:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:56:03 crc kubenswrapper[4754]: I1005 20:56:03.746240 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:56:03 crc kubenswrapper[4754]: I1005 20:56:03.746339 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:56:03 crc kubenswrapper[4754]: I1005 20:56:03.746361 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:56:03 crc kubenswrapper[4754]: I1005 20:56:03.746399 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:56:03 crc kubenswrapper[4754]: I1005 20:56:03.746427 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:56:03Z","lastTransitionTime":"2025-10-05T20:56:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:56:03 crc kubenswrapper[4754]: I1005 20:56:03.836626 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nwrnt" Oct 05 20:56:03 crc kubenswrapper[4754]: E1005 20:56:03.836895 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nwrnt" podUID="7dc728ea-2601-44f0-bcab-2913f034007d" Oct 05 20:56:03 crc kubenswrapper[4754]: I1005 20:56:03.850792 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:56:03 crc kubenswrapper[4754]: I1005 20:56:03.850859 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:56:03 crc kubenswrapper[4754]: I1005 20:56:03.850876 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:56:03 crc kubenswrapper[4754]: I1005 20:56:03.850905 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:56:03 crc kubenswrapper[4754]: I1005 20:56:03.850925 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:56:03Z","lastTransitionTime":"2025-10-05T20:56:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:56:03 crc kubenswrapper[4754]: I1005 20:56:03.956314 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:56:03 crc kubenswrapper[4754]: I1005 20:56:03.956450 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:56:03 crc kubenswrapper[4754]: I1005 20:56:03.956470 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:56:03 crc kubenswrapper[4754]: I1005 20:56:03.956528 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:56:03 crc kubenswrapper[4754]: I1005 20:56:03.956550 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:56:03Z","lastTransitionTime":"2025-10-05T20:56:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:56:04 crc kubenswrapper[4754]: I1005 20:56:04.059827 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:56:04 crc kubenswrapper[4754]: I1005 20:56:04.059919 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:56:04 crc kubenswrapper[4754]: I1005 20:56:04.059937 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:56:04 crc kubenswrapper[4754]: I1005 20:56:04.059999 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:56:04 crc kubenswrapper[4754]: I1005 20:56:04.060023 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:56:04Z","lastTransitionTime":"2025-10-05T20:56:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:56:04 crc kubenswrapper[4754]: I1005 20:56:04.164534 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:56:04 crc kubenswrapper[4754]: I1005 20:56:04.164629 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:56:04 crc kubenswrapper[4754]: I1005 20:56:04.164652 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:56:04 crc kubenswrapper[4754]: I1005 20:56:04.164681 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:56:04 crc kubenswrapper[4754]: I1005 20:56:04.164700 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:56:04Z","lastTransitionTime":"2025-10-05T20:56:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:56:04 crc kubenswrapper[4754]: I1005 20:56:04.269126 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:56:04 crc kubenswrapper[4754]: I1005 20:56:04.269224 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:56:04 crc kubenswrapper[4754]: I1005 20:56:04.269247 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:56:04 crc kubenswrapper[4754]: I1005 20:56:04.269275 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:56:04 crc kubenswrapper[4754]: I1005 20:56:04.269295 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:56:04Z","lastTransitionTime":"2025-10-05T20:56:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:56:04 crc kubenswrapper[4754]: I1005 20:56:04.373129 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:56:04 crc kubenswrapper[4754]: I1005 20:56:04.373197 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:56:04 crc kubenswrapper[4754]: I1005 20:56:04.373214 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:56:04 crc kubenswrapper[4754]: I1005 20:56:04.373240 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:56:04 crc kubenswrapper[4754]: I1005 20:56:04.373258 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:56:04Z","lastTransitionTime":"2025-10-05T20:56:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:56:04 crc kubenswrapper[4754]: I1005 20:56:04.476919 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:56:04 crc kubenswrapper[4754]: I1005 20:56:04.476995 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:56:04 crc kubenswrapper[4754]: I1005 20:56:04.477014 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:56:04 crc kubenswrapper[4754]: I1005 20:56:04.477046 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:56:04 crc kubenswrapper[4754]: I1005 20:56:04.477070 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:56:04Z","lastTransitionTime":"2025-10-05T20:56:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:56:04 crc kubenswrapper[4754]: I1005 20:56:04.580989 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:56:04 crc kubenswrapper[4754]: I1005 20:56:04.581072 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:56:04 crc kubenswrapper[4754]: I1005 20:56:04.581102 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:56:04 crc kubenswrapper[4754]: I1005 20:56:04.581136 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:56:04 crc kubenswrapper[4754]: I1005 20:56:04.581162 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:56:04Z","lastTransitionTime":"2025-10-05T20:56:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:56:04 crc kubenswrapper[4754]: I1005 20:56:04.684603 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:56:04 crc kubenswrapper[4754]: I1005 20:56:04.684677 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:56:04 crc kubenswrapper[4754]: I1005 20:56:04.684697 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:56:04 crc kubenswrapper[4754]: I1005 20:56:04.684725 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:56:04 crc kubenswrapper[4754]: I1005 20:56:04.684743 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:56:04Z","lastTransitionTime":"2025-10-05T20:56:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:56:04 crc kubenswrapper[4754]: I1005 20:56:04.788565 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:56:04 crc kubenswrapper[4754]: I1005 20:56:04.788716 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:56:04 crc kubenswrapper[4754]: I1005 20:56:04.788737 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:56:04 crc kubenswrapper[4754]: I1005 20:56:04.788771 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:56:04 crc kubenswrapper[4754]: I1005 20:56:04.788794 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:56:04Z","lastTransitionTime":"2025-10-05T20:56:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:56:04 crc kubenswrapper[4754]: I1005 20:56:04.837194 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 20:56:04 crc kubenswrapper[4754]: I1005 20:56:04.837280 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 20:56:04 crc kubenswrapper[4754]: E1005 20:56:04.837433 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 20:56:04 crc kubenswrapper[4754]: I1005 20:56:04.837194 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 20:56:04 crc kubenswrapper[4754]: E1005 20:56:04.837656 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 20:56:04 crc kubenswrapper[4754]: E1005 20:56:04.837841 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 20:56:04 crc kubenswrapper[4754]: I1005 20:56:04.892345 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:56:04 crc kubenswrapper[4754]: I1005 20:56:04.892428 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:56:04 crc kubenswrapper[4754]: I1005 20:56:04.892451 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:56:04 crc kubenswrapper[4754]: I1005 20:56:04.892482 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:56:04 crc kubenswrapper[4754]: I1005 20:56:04.892543 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:56:04Z","lastTransitionTime":"2025-10-05T20:56:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:56:04 crc kubenswrapper[4754]: I1005 20:56:04.996776 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:56:04 crc kubenswrapper[4754]: I1005 20:56:04.996853 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:56:04 crc kubenswrapper[4754]: I1005 20:56:04.996878 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:56:04 crc kubenswrapper[4754]: I1005 20:56:04.996911 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:56:04 crc kubenswrapper[4754]: I1005 20:56:04.996935 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:56:04Z","lastTransitionTime":"2025-10-05T20:56:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:56:05 crc kubenswrapper[4754]: I1005 20:56:05.100177 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:56:05 crc kubenswrapper[4754]: I1005 20:56:05.100249 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:56:05 crc kubenswrapper[4754]: I1005 20:56:05.100271 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:56:05 crc kubenswrapper[4754]: I1005 20:56:05.100305 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:56:05 crc kubenswrapper[4754]: I1005 20:56:05.100330 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:56:05Z","lastTransitionTime":"2025-10-05T20:56:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:56:05 crc kubenswrapper[4754]: I1005 20:56:05.204103 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:56:05 crc kubenswrapper[4754]: I1005 20:56:05.204183 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:56:05 crc kubenswrapper[4754]: I1005 20:56:05.204201 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:56:05 crc kubenswrapper[4754]: I1005 20:56:05.204229 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:56:05 crc kubenswrapper[4754]: I1005 20:56:05.204250 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:56:05Z","lastTransitionTime":"2025-10-05T20:56:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:56:05 crc kubenswrapper[4754]: I1005 20:56:05.308906 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:56:05 crc kubenswrapper[4754]: I1005 20:56:05.308959 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:56:05 crc kubenswrapper[4754]: I1005 20:56:05.308968 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:56:05 crc kubenswrapper[4754]: I1005 20:56:05.308988 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:56:05 crc kubenswrapper[4754]: I1005 20:56:05.309000 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:56:05Z","lastTransitionTime":"2025-10-05T20:56:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:56:05 crc kubenswrapper[4754]: I1005 20:56:05.412411 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:56:05 crc kubenswrapper[4754]: I1005 20:56:05.412535 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:56:05 crc kubenswrapper[4754]: I1005 20:56:05.412562 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:56:05 crc kubenswrapper[4754]: I1005 20:56:05.412587 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:56:05 crc kubenswrapper[4754]: I1005 20:56:05.412638 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:56:05Z","lastTransitionTime":"2025-10-05T20:56:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:56:05 crc kubenswrapper[4754]: I1005 20:56:05.516272 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:56:05 crc kubenswrapper[4754]: I1005 20:56:05.516363 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:56:05 crc kubenswrapper[4754]: I1005 20:56:05.516382 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:56:05 crc kubenswrapper[4754]: I1005 20:56:05.516404 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:56:05 crc kubenswrapper[4754]: I1005 20:56:05.516420 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:56:05Z","lastTransitionTime":"2025-10-05T20:56:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:56:05 crc kubenswrapper[4754]: I1005 20:56:05.619476 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:56:05 crc kubenswrapper[4754]: I1005 20:56:05.619580 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:56:05 crc kubenswrapper[4754]: I1005 20:56:05.619599 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:56:05 crc kubenswrapper[4754]: I1005 20:56:05.619622 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:56:05 crc kubenswrapper[4754]: I1005 20:56:05.619641 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:56:05Z","lastTransitionTime":"2025-10-05T20:56:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:56:05 crc kubenswrapper[4754]: I1005 20:56:05.723201 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:56:05 crc kubenswrapper[4754]: I1005 20:56:05.723291 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:56:05 crc kubenswrapper[4754]: I1005 20:56:05.723314 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:56:05 crc kubenswrapper[4754]: I1005 20:56:05.723344 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:56:05 crc kubenswrapper[4754]: I1005 20:56:05.723366 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:56:05Z","lastTransitionTime":"2025-10-05T20:56:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:56:05 crc kubenswrapper[4754]: I1005 20:56:05.826076 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:56:05 crc kubenswrapper[4754]: I1005 20:56:05.826143 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:56:05 crc kubenswrapper[4754]: I1005 20:56:05.826162 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:56:05 crc kubenswrapper[4754]: I1005 20:56:05.826191 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:56:05 crc kubenswrapper[4754]: I1005 20:56:05.826210 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:56:05Z","lastTransitionTime":"2025-10-05T20:56:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:56:05 crc kubenswrapper[4754]: I1005 20:56:05.836241 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nwrnt" Oct 05 20:56:05 crc kubenswrapper[4754]: E1005 20:56:05.836449 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nwrnt" podUID="7dc728ea-2601-44f0-bcab-2913f034007d" Oct 05 20:56:05 crc kubenswrapper[4754]: I1005 20:56:05.928706 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:56:05 crc kubenswrapper[4754]: I1005 20:56:05.928762 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:56:05 crc kubenswrapper[4754]: I1005 20:56:05.928775 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:56:05 crc kubenswrapper[4754]: I1005 20:56:05.928799 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:56:05 crc kubenswrapper[4754]: I1005 20:56:05.928817 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:56:05Z","lastTransitionTime":"2025-10-05T20:56:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:56:06 crc kubenswrapper[4754]: I1005 20:56:06.032273 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:56:06 crc kubenswrapper[4754]: I1005 20:56:06.032356 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:56:06 crc kubenswrapper[4754]: I1005 20:56:06.032376 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:56:06 crc kubenswrapper[4754]: I1005 20:56:06.032407 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:56:06 crc kubenswrapper[4754]: I1005 20:56:06.032428 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:56:06Z","lastTransitionTime":"2025-10-05T20:56:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:56:06 crc kubenswrapper[4754]: I1005 20:56:06.135390 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:56:06 crc kubenswrapper[4754]: I1005 20:56:06.135478 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:56:06 crc kubenswrapper[4754]: I1005 20:56:06.135521 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:56:06 crc kubenswrapper[4754]: I1005 20:56:06.135548 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:56:06 crc kubenswrapper[4754]: I1005 20:56:06.135567 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:56:06Z","lastTransitionTime":"2025-10-05T20:56:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:56:06 crc kubenswrapper[4754]: I1005 20:56:06.238431 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:56:06 crc kubenswrapper[4754]: I1005 20:56:06.238562 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:56:06 crc kubenswrapper[4754]: I1005 20:56:06.238585 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:56:06 crc kubenswrapper[4754]: I1005 20:56:06.238613 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:56:06 crc kubenswrapper[4754]: I1005 20:56:06.238633 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:56:06Z","lastTransitionTime":"2025-10-05T20:56:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:56:06 crc kubenswrapper[4754]: I1005 20:56:06.260980 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 20:56:06 crc kubenswrapper[4754]: I1005 20:56:06.261211 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 20:56:06 crc kubenswrapper[4754]: I1005 20:56:06.261630 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 20:56:06 crc kubenswrapper[4754]: I1005 20:56:06.261877 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 20:56:06 crc kubenswrapper[4754]: I1005 20:56:06.262094 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T20:56:06Z","lastTransitionTime":"2025-10-05T20:56:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 20:56:06 crc kubenswrapper[4754]: I1005 20:56:06.335082 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-8hm4h"] Oct 05 20:56:06 crc kubenswrapper[4754]: I1005 20:56:06.335816 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8hm4h" Oct 05 20:56:06 crc kubenswrapper[4754]: I1005 20:56:06.339845 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Oct 05 20:56:06 crc kubenswrapper[4754]: I1005 20:56:06.340111 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Oct 05 20:56:06 crc kubenswrapper[4754]: I1005 20:56:06.340911 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Oct 05 20:56:06 crc kubenswrapper[4754]: I1005 20:56:06.341366 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Oct 05 20:56:06 crc kubenswrapper[4754]: I1005 20:56:06.382577 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=13.382535561 podStartE2EDuration="13.382535561s" podCreationTimestamp="2025-10-05 20:55:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 20:56:06.359788382 +0000 UTC m=+90.263907122" watchObservedRunningTime="2025-10-05 20:56:06.382535561 +0000 UTC m=+90.286654321" Oct 05 20:56:06 crc kubenswrapper[4754]: I1005 20:56:06.405707 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=33.405645981 podStartE2EDuration="33.405645981s" podCreationTimestamp="2025-10-05 20:55:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 20:56:06.383413254 +0000 UTC m=+90.287532004" watchObservedRunningTime="2025-10-05 20:56:06.405645981 +0000 UTC m=+90.309764731" Oct 05 20:56:06 crc kubenswrapper[4754]: I1005 20:56:06.469545 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/6b0132bd-4db1-4d43-994f-0adc3d96d596-service-ca\") pod \"cluster-version-operator-5c965bbfc6-8hm4h\" (UID: \"6b0132bd-4db1-4d43-994f-0adc3d96d596\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8hm4h" Oct 05 20:56:06 crc kubenswrapper[4754]: I1005 20:56:06.469620 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6b0132bd-4db1-4d43-994f-0adc3d96d596-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-8hm4h\" (UID: \"6b0132bd-4db1-4d43-994f-0adc3d96d596\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8hm4h" Oct 05 20:56:06 crc kubenswrapper[4754]: I1005 20:56:06.469658 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/6b0132bd-4db1-4d43-994f-0adc3d96d596-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-8hm4h\" (UID: \"6b0132bd-4db1-4d43-994f-0adc3d96d596\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8hm4h" Oct 05 20:56:06 crc kubenswrapper[4754]: I1005 20:56:06.469752 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/6b0132bd-4db1-4d43-994f-0adc3d96d596-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-8hm4h\" (UID: \"6b0132bd-4db1-4d43-994f-0adc3d96d596\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8hm4h" Oct 05 20:56:06 crc kubenswrapper[4754]: I1005 20:56:06.469788 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6b0132bd-4db1-4d43-994f-0adc3d96d596-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-8hm4h\" (UID: \"6b0132bd-4db1-4d43-994f-0adc3d96d596\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8hm4h" Oct 05 20:56:06 crc kubenswrapper[4754]: I1005 20:56:06.507092 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=66.507062251 podStartE2EDuration="1m6.507062251s" podCreationTimestamp="2025-10-05 20:55:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 20:56:06.506861465 +0000 UTC m=+90.410980245" watchObservedRunningTime="2025-10-05 20:56:06.507062251 +0000 UTC m=+90.411180991" Oct 05 20:56:06 crc kubenswrapper[4754]: I1005 20:56:06.571113 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/6b0132bd-4db1-4d43-994f-0adc3d96d596-service-ca\") pod \"cluster-version-operator-5c965bbfc6-8hm4h\" (UID: \"6b0132bd-4db1-4d43-994f-0adc3d96d596\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8hm4h" Oct 05 20:56:06 crc kubenswrapper[4754]: I1005 20:56:06.571176 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6b0132bd-4db1-4d43-994f-0adc3d96d596-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-8hm4h\" (UID: \"6b0132bd-4db1-4d43-994f-0adc3d96d596\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8hm4h" Oct 05 20:56:06 crc kubenswrapper[4754]: I1005 20:56:06.571213 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/6b0132bd-4db1-4d43-994f-0adc3d96d596-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-8hm4h\" (UID: \"6b0132bd-4db1-4d43-994f-0adc3d96d596\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8hm4h" Oct 05 20:56:06 crc kubenswrapper[4754]: I1005 20:56:06.571289 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/6b0132bd-4db1-4d43-994f-0adc3d96d596-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-8hm4h\" (UID: \"6b0132bd-4db1-4d43-994f-0adc3d96d596\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8hm4h" Oct 05 20:56:06 crc kubenswrapper[4754]: I1005 20:56:06.571324 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6b0132bd-4db1-4d43-994f-0adc3d96d596-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-8hm4h\" (UID: \"6b0132bd-4db1-4d43-994f-0adc3d96d596\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8hm4h" Oct 05 20:56:06 crc kubenswrapper[4754]: I1005 20:56:06.571811 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/6b0132bd-4db1-4d43-994f-0adc3d96d596-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-8hm4h\" (UID: \"6b0132bd-4db1-4d43-994f-0adc3d96d596\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8hm4h" Oct 05 20:56:06 crc kubenswrapper[4754]: I1005 20:56:06.571857 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/6b0132bd-4db1-4d43-994f-0adc3d96d596-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-8hm4h\" (UID: \"6b0132bd-4db1-4d43-994f-0adc3d96d596\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8hm4h" Oct 05 20:56:06 crc kubenswrapper[4754]: I1005 20:56:06.573901 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/6b0132bd-4db1-4d43-994f-0adc3d96d596-service-ca\") pod \"cluster-version-operator-5c965bbfc6-8hm4h\" (UID: \"6b0132bd-4db1-4d43-994f-0adc3d96d596\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8hm4h" Oct 05 20:56:06 crc kubenswrapper[4754]: I1005 20:56:06.578739 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6b0132bd-4db1-4d43-994f-0adc3d96d596-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-8hm4h\" (UID: \"6b0132bd-4db1-4d43-994f-0adc3d96d596\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8hm4h" Oct 05 20:56:06 crc kubenswrapper[4754]: I1005 20:56:06.613221 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=67.613194633 podStartE2EDuration="1m7.613194633s" podCreationTimestamp="2025-10-05 20:54:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 20:56:06.589069287 +0000 UTC m=+90.493188067" watchObservedRunningTime="2025-10-05 20:56:06.613194633 +0000 UTC m=+90.517313343" Oct 05 20:56:06 crc kubenswrapper[4754]: I1005 20:56:06.613753 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6b0132bd-4db1-4d43-994f-0adc3d96d596-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-8hm4h\" (UID: \"6b0132bd-4db1-4d43-994f-0adc3d96d596\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8hm4h" Oct 05 20:56:06 crc kubenswrapper[4754]: I1005 20:56:06.663386 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8hm4h" Oct 05 20:56:06 crc kubenswrapper[4754]: I1005 20:56:06.683903 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-57sr4" podStartSLOduration=69.683869815 podStartE2EDuration="1m9.683869815s" podCreationTimestamp="2025-10-05 20:54:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 20:56:06.653684693 +0000 UTC m=+90.557803443" watchObservedRunningTime="2025-10-05 20:56:06.683869815 +0000 UTC m=+90.587988565" Oct 05 20:56:06 crc kubenswrapper[4754]: W1005 20:56:06.706347 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6b0132bd_4db1_4d43_994f_0adc3d96d596.slice/crio-0c3eb849c2ed00c3b887097d007bc3d6b93f9739e64314a5e585376f322e0839 WatchSource:0}: Error finding container 0c3eb849c2ed00c3b887097d007bc3d6b93f9739e64314a5e585376f322e0839: Status 404 returned error can't find the container with id 0c3eb849c2ed00c3b887097d007bc3d6b93f9739e64314a5e585376f322e0839 Oct 05 20:56:06 crc kubenswrapper[4754]: I1005 20:56:06.709385 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-ss7lm" podStartSLOduration=69.709365247 podStartE2EDuration="1m9.709365247s" podCreationTimestamp="2025-10-05 20:54:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 20:56:06.708905855 +0000 UTC m=+90.613024575" watchObservedRunningTime="2025-10-05 20:56:06.709365247 +0000 UTC m=+90.613483967" Oct 05 20:56:06 crc kubenswrapper[4754]: I1005 20:56:06.710021 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-cn76s" podStartSLOduration=68.710014783 podStartE2EDuration="1m8.710014783s" podCreationTimestamp="2025-10-05 20:54:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 20:56:06.685222741 +0000 UTC m=+90.589341501" watchObservedRunningTime="2025-10-05 20:56:06.710014783 +0000 UTC m=+90.614133503" Oct 05 20:56:06 crc kubenswrapper[4754]: I1005 20:56:06.808399 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=67.808379904 podStartE2EDuration="1m7.808379904s" podCreationTimestamp="2025-10-05 20:54:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 20:56:06.780264855 +0000 UTC m=+90.684383565" watchObservedRunningTime="2025-10-05 20:56:06.808379904 +0000 UTC m=+90.712498614" Oct 05 20:56:06 crc kubenswrapper[4754]: I1005 20:56:06.831160 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-tskt7" podStartSLOduration=68.831146525 podStartE2EDuration="1m8.831146525s" podCreationTimestamp="2025-10-05 20:54:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 20:56:06.831090473 +0000 UTC m=+90.735209183" watchObservedRunningTime="2025-10-05 20:56:06.831146525 +0000 UTC m=+90.735265235" Oct 05 20:56:06 crc kubenswrapper[4754]: I1005 20:56:06.838643 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 20:56:06 crc kubenswrapper[4754]: E1005 20:56:06.838746 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 20:56:06 crc kubenswrapper[4754]: I1005 20:56:06.838908 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 20:56:06 crc kubenswrapper[4754]: E1005 20:56:06.838970 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 20:56:06 crc kubenswrapper[4754]: I1005 20:56:06.839077 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 20:56:06 crc kubenswrapper[4754]: E1005 20:56:06.839125 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 20:56:06 crc kubenswrapper[4754]: I1005 20:56:06.855713 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podStartSLOduration=68.855695551 podStartE2EDuration="1m8.855695551s" podCreationTimestamp="2025-10-05 20:54:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 20:56:06.84485038 +0000 UTC m=+90.748969090" watchObservedRunningTime="2025-10-05 20:56:06.855695551 +0000 UTC m=+90.759814261" Oct 05 20:56:06 crc kubenswrapper[4754]: I1005 20:56:06.856084 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rddkl" podStartSLOduration=68.856080491 podStartE2EDuration="1m8.856080491s" podCreationTimestamp="2025-10-05 20:54:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 20:56:06.855381613 +0000 UTC m=+90.759500323" watchObservedRunningTime="2025-10-05 20:56:06.856080491 +0000 UTC m=+90.760199201" Oct 05 20:56:07 crc kubenswrapper[4754]: I1005 20:56:07.522842 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8hm4h" event={"ID":"6b0132bd-4db1-4d43-994f-0adc3d96d596","Type":"ContainerStarted","Data":"445b5c0c81bb4310c9339ebb4b99826138c605e4f398e8d96d2e88c040feefe4"} Oct 05 20:56:07 crc kubenswrapper[4754]: I1005 20:56:07.522930 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8hm4h" event={"ID":"6b0132bd-4db1-4d43-994f-0adc3d96d596","Type":"ContainerStarted","Data":"0c3eb849c2ed00c3b887097d007bc3d6b93f9739e64314a5e585376f322e0839"} Oct 05 20:56:07 crc kubenswrapper[4754]: I1005 20:56:07.836997 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nwrnt" Oct 05 20:56:07 crc kubenswrapper[4754]: E1005 20:56:07.837424 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nwrnt" podUID="7dc728ea-2601-44f0-bcab-2913f034007d" Oct 05 20:56:08 crc kubenswrapper[4754]: I1005 20:56:08.836947 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 20:56:08 crc kubenswrapper[4754]: I1005 20:56:08.837031 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 20:56:08 crc kubenswrapper[4754]: E1005 20:56:08.837131 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 20:56:08 crc kubenswrapper[4754]: E1005 20:56:08.837323 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 20:56:08 crc kubenswrapper[4754]: I1005 20:56:08.837681 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 20:56:08 crc kubenswrapper[4754]: E1005 20:56:08.837853 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 20:56:09 crc kubenswrapper[4754]: I1005 20:56:09.836704 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nwrnt" Oct 05 20:56:09 crc kubenswrapper[4754]: E1005 20:56:09.836935 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nwrnt" podUID="7dc728ea-2601-44f0-bcab-2913f034007d" Oct 05 20:56:10 crc kubenswrapper[4754]: I1005 20:56:10.837281 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 20:56:10 crc kubenswrapper[4754]: I1005 20:56:10.837521 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 20:56:10 crc kubenswrapper[4754]: E1005 20:56:10.837488 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 20:56:10 crc kubenswrapper[4754]: I1005 20:56:10.837650 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 20:56:10 crc kubenswrapper[4754]: E1005 20:56:10.838286 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 20:56:10 crc kubenswrapper[4754]: E1005 20:56:10.838451 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 20:56:10 crc kubenswrapper[4754]: I1005 20:56:10.838718 4754 scope.go:117] "RemoveContainer" containerID="7c990164cb7213726a9086a630586e38da00d62fdb57aed8418bbec265a54705" Oct 05 20:56:10 crc kubenswrapper[4754]: E1005 20:56:10.838962 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-7rhps_openshift-ovn-kubernetes(b44790fb-fecb-4ec3-9816-8e23dd72ab33)\"" pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" podUID="b44790fb-fecb-4ec3-9816-8e23dd72ab33" Oct 05 20:56:11 crc kubenswrapper[4754]: I1005 20:56:11.836640 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nwrnt" Oct 05 20:56:11 crc kubenswrapper[4754]: E1005 20:56:11.837045 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nwrnt" podUID="7dc728ea-2601-44f0-bcab-2913f034007d" Oct 05 20:56:12 crc kubenswrapper[4754]: I1005 20:56:12.837767 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 20:56:12 crc kubenswrapper[4754]: I1005 20:56:12.837805 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 20:56:12 crc kubenswrapper[4754]: E1005 20:56:12.838310 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 20:56:12 crc kubenswrapper[4754]: I1005 20:56:12.837997 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 20:56:12 crc kubenswrapper[4754]: E1005 20:56:12.838517 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 20:56:12 crc kubenswrapper[4754]: E1005 20:56:12.838779 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 20:56:13 crc kubenswrapper[4754]: I1005 20:56:13.836451 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nwrnt" Oct 05 20:56:13 crc kubenswrapper[4754]: E1005 20:56:13.836717 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nwrnt" podUID="7dc728ea-2601-44f0-bcab-2913f034007d" Oct 05 20:56:14 crc kubenswrapper[4754]: I1005 20:56:14.836563 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 20:56:14 crc kubenswrapper[4754]: E1005 20:56:14.836790 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 20:56:14 crc kubenswrapper[4754]: I1005 20:56:14.836879 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 20:56:14 crc kubenswrapper[4754]: E1005 20:56:14.837046 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 20:56:14 crc kubenswrapper[4754]: I1005 20:56:14.837391 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 20:56:14 crc kubenswrapper[4754]: E1005 20:56:14.837618 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 20:56:15 crc kubenswrapper[4754]: I1005 20:56:15.837100 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nwrnt" Oct 05 20:56:15 crc kubenswrapper[4754]: E1005 20:56:15.837361 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nwrnt" podUID="7dc728ea-2601-44f0-bcab-2913f034007d" Oct 05 20:56:16 crc kubenswrapper[4754]: I1005 20:56:16.836886 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 20:56:16 crc kubenswrapper[4754]: I1005 20:56:16.836910 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 20:56:16 crc kubenswrapper[4754]: E1005 20:56:16.839007 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 20:56:16 crc kubenswrapper[4754]: I1005 20:56:16.839060 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 20:56:16 crc kubenswrapper[4754]: E1005 20:56:16.839247 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 20:56:16 crc kubenswrapper[4754]: E1005 20:56:16.839408 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 20:56:17 crc kubenswrapper[4754]: I1005 20:56:17.634165 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7dc728ea-2601-44f0-bcab-2913f034007d-metrics-certs\") pod \"network-metrics-daemon-nwrnt\" (UID: \"7dc728ea-2601-44f0-bcab-2913f034007d\") " pod="openshift-multus/network-metrics-daemon-nwrnt" Oct 05 20:56:17 crc kubenswrapper[4754]: E1005 20:56:17.634531 4754 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 05 20:56:17 crc kubenswrapper[4754]: E1005 20:56:17.634679 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7dc728ea-2601-44f0-bcab-2913f034007d-metrics-certs podName:7dc728ea-2601-44f0-bcab-2913f034007d nodeName:}" failed. No retries permitted until 2025-10-05 20:57:21.63464037 +0000 UTC m=+165.538759120 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/7dc728ea-2601-44f0-bcab-2913f034007d-metrics-certs") pod "network-metrics-daemon-nwrnt" (UID: "7dc728ea-2601-44f0-bcab-2913f034007d") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 05 20:56:17 crc kubenswrapper[4754]: I1005 20:56:17.836815 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nwrnt" Oct 05 20:56:17 crc kubenswrapper[4754]: E1005 20:56:17.837056 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nwrnt" podUID="7dc728ea-2601-44f0-bcab-2913f034007d" Oct 05 20:56:18 crc kubenswrapper[4754]: I1005 20:56:18.837177 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 20:56:18 crc kubenswrapper[4754]: I1005 20:56:18.837277 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 20:56:18 crc kubenswrapper[4754]: I1005 20:56:18.837205 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 20:56:18 crc kubenswrapper[4754]: E1005 20:56:18.837388 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 20:56:18 crc kubenswrapper[4754]: E1005 20:56:18.837609 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 20:56:18 crc kubenswrapper[4754]: E1005 20:56:18.837729 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 20:56:19 crc kubenswrapper[4754]: I1005 20:56:19.837716 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nwrnt" Oct 05 20:56:19 crc kubenswrapper[4754]: E1005 20:56:19.838180 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nwrnt" podUID="7dc728ea-2601-44f0-bcab-2913f034007d" Oct 05 20:56:20 crc kubenswrapper[4754]: I1005 20:56:20.836369 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 20:56:20 crc kubenswrapper[4754]: E1005 20:56:20.836627 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 20:56:20 crc kubenswrapper[4754]: I1005 20:56:20.836711 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 20:56:20 crc kubenswrapper[4754]: E1005 20:56:20.836878 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 20:56:20 crc kubenswrapper[4754]: I1005 20:56:20.836411 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 20:56:20 crc kubenswrapper[4754]: E1005 20:56:20.837041 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 20:56:21 crc kubenswrapper[4754]: I1005 20:56:21.836997 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nwrnt" Oct 05 20:56:21 crc kubenswrapper[4754]: E1005 20:56:21.837776 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nwrnt" podUID="7dc728ea-2601-44f0-bcab-2913f034007d" Oct 05 20:56:22 crc kubenswrapper[4754]: I1005 20:56:22.837313 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 20:56:22 crc kubenswrapper[4754]: I1005 20:56:22.837372 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 20:56:22 crc kubenswrapper[4754]: I1005 20:56:22.837634 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 20:56:22 crc kubenswrapper[4754]: E1005 20:56:22.837774 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 20:56:22 crc kubenswrapper[4754]: E1005 20:56:22.837925 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 20:56:22 crc kubenswrapper[4754]: E1005 20:56:22.838065 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 20:56:23 crc kubenswrapper[4754]: I1005 20:56:23.836924 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nwrnt" Oct 05 20:56:23 crc kubenswrapper[4754]: E1005 20:56:23.837143 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nwrnt" podUID="7dc728ea-2601-44f0-bcab-2913f034007d" Oct 05 20:56:24 crc kubenswrapper[4754]: I1005 20:56:24.836889 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 20:56:24 crc kubenswrapper[4754]: I1005 20:56:24.837669 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 20:56:24 crc kubenswrapper[4754]: E1005 20:56:24.837905 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 20:56:24 crc kubenswrapper[4754]: I1005 20:56:24.838408 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 20:56:24 crc kubenswrapper[4754]: E1005 20:56:24.838554 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 20:56:24 crc kubenswrapper[4754]: E1005 20:56:24.838738 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 20:56:24 crc kubenswrapper[4754]: I1005 20:56:24.838752 4754 scope.go:117] "RemoveContainer" containerID="7c990164cb7213726a9086a630586e38da00d62fdb57aed8418bbec265a54705" Oct 05 20:56:24 crc kubenswrapper[4754]: E1005 20:56:24.839110 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-7rhps_openshift-ovn-kubernetes(b44790fb-fecb-4ec3-9816-8e23dd72ab33)\"" pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" podUID="b44790fb-fecb-4ec3-9816-8e23dd72ab33" Oct 05 20:56:25 crc kubenswrapper[4754]: I1005 20:56:25.837267 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nwrnt" Oct 05 20:56:25 crc kubenswrapper[4754]: E1005 20:56:25.837949 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nwrnt" podUID="7dc728ea-2601-44f0-bcab-2913f034007d" Oct 05 20:56:26 crc kubenswrapper[4754]: I1005 20:56:26.836612 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 20:56:26 crc kubenswrapper[4754]: I1005 20:56:26.836775 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 20:56:26 crc kubenswrapper[4754]: I1005 20:56:26.836856 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 20:56:26 crc kubenswrapper[4754]: E1005 20:56:26.838792 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 20:56:26 crc kubenswrapper[4754]: E1005 20:56:26.839783 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 20:56:26 crc kubenswrapper[4754]: E1005 20:56:26.839855 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 20:56:27 crc kubenswrapper[4754]: I1005 20:56:27.836931 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nwrnt" Oct 05 20:56:27 crc kubenswrapper[4754]: E1005 20:56:27.837605 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nwrnt" podUID="7dc728ea-2601-44f0-bcab-2913f034007d" Oct 05 20:56:28 crc kubenswrapper[4754]: I1005 20:56:28.837019 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 20:56:28 crc kubenswrapper[4754]: I1005 20:56:28.837185 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 20:56:28 crc kubenswrapper[4754]: E1005 20:56:28.837441 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 20:56:28 crc kubenswrapper[4754]: I1005 20:56:28.837561 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 20:56:28 crc kubenswrapper[4754]: E1005 20:56:28.837961 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 20:56:28 crc kubenswrapper[4754]: E1005 20:56:28.838062 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 20:56:29 crc kubenswrapper[4754]: I1005 20:56:29.836795 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nwrnt" Oct 05 20:56:29 crc kubenswrapper[4754]: E1005 20:56:29.836985 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nwrnt" podUID="7dc728ea-2601-44f0-bcab-2913f034007d" Oct 05 20:56:30 crc kubenswrapper[4754]: I1005 20:56:30.836966 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 20:56:30 crc kubenswrapper[4754]: I1005 20:56:30.836966 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 20:56:30 crc kubenswrapper[4754]: E1005 20:56:30.837209 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 20:56:30 crc kubenswrapper[4754]: I1005 20:56:30.836984 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 20:56:30 crc kubenswrapper[4754]: E1005 20:56:30.837305 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 20:56:30 crc kubenswrapper[4754]: E1005 20:56:30.837346 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 20:56:31 crc kubenswrapper[4754]: I1005 20:56:31.836236 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nwrnt" Oct 05 20:56:31 crc kubenswrapper[4754]: E1005 20:56:31.836421 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nwrnt" podUID="7dc728ea-2601-44f0-bcab-2913f034007d" Oct 05 20:56:32 crc kubenswrapper[4754]: I1005 20:56:32.837325 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 20:56:32 crc kubenswrapper[4754]: I1005 20:56:32.837358 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 20:56:32 crc kubenswrapper[4754]: E1005 20:56:32.837541 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 20:56:32 crc kubenswrapper[4754]: I1005 20:56:32.837615 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 20:56:32 crc kubenswrapper[4754]: E1005 20:56:32.837819 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 20:56:32 crc kubenswrapper[4754]: E1005 20:56:32.838427 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 20:56:33 crc kubenswrapper[4754]: I1005 20:56:33.630872 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-cn76s_02d5e3f9-73c2-4496-9aca-0787184aef19/kube-multus/1.log" Oct 05 20:56:33 crc kubenswrapper[4754]: I1005 20:56:33.631705 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-cn76s_02d5e3f9-73c2-4496-9aca-0787184aef19/kube-multus/0.log" Oct 05 20:56:33 crc kubenswrapper[4754]: I1005 20:56:33.631804 4754 generic.go:334] "Generic (PLEG): container finished" podID="02d5e3f9-73c2-4496-9aca-0787184aef19" containerID="5ed23b9cd9b57bb169a8ebbf16b7ff8d196eb14330df0edf2b9fcd2a8c9e2c88" exitCode=1 Oct 05 20:56:33 crc kubenswrapper[4754]: I1005 20:56:33.631862 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-cn76s" event={"ID":"02d5e3f9-73c2-4496-9aca-0787184aef19","Type":"ContainerDied","Data":"5ed23b9cd9b57bb169a8ebbf16b7ff8d196eb14330df0edf2b9fcd2a8c9e2c88"} Oct 05 20:56:33 crc kubenswrapper[4754]: I1005 20:56:33.631932 4754 scope.go:117] "RemoveContainer" containerID="f7ba2b076839f4a5f586d2e6c4f4d7188254128b782bfd2180e22bf3ea5ffdd7" Oct 05 20:56:33 crc kubenswrapper[4754]: I1005 20:56:33.632554 4754 scope.go:117] "RemoveContainer" containerID="5ed23b9cd9b57bb169a8ebbf16b7ff8d196eb14330df0edf2b9fcd2a8c9e2c88" Oct 05 20:56:33 crc kubenswrapper[4754]: E1005 20:56:33.632873 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-cn76s_openshift-multus(02d5e3f9-73c2-4496-9aca-0787184aef19)\"" pod="openshift-multus/multus-cn76s" podUID="02d5e3f9-73c2-4496-9aca-0787184aef19" Oct 05 20:56:33 crc kubenswrapper[4754]: I1005 20:56:33.665048 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8hm4h" podStartSLOduration=95.665015753 podStartE2EDuration="1m35.665015753s" podCreationTimestamp="2025-10-05 20:54:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 20:56:07.542574302 +0000 UTC m=+91.446693042" watchObservedRunningTime="2025-10-05 20:56:33.665015753 +0000 UTC m=+117.569134503" Oct 05 20:56:33 crc kubenswrapper[4754]: I1005 20:56:33.836653 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nwrnt" Oct 05 20:56:33 crc kubenswrapper[4754]: E1005 20:56:33.836864 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nwrnt" podUID="7dc728ea-2601-44f0-bcab-2913f034007d" Oct 05 20:56:34 crc kubenswrapper[4754]: I1005 20:56:34.639071 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-cn76s_02d5e3f9-73c2-4496-9aca-0787184aef19/kube-multus/1.log" Oct 05 20:56:34 crc kubenswrapper[4754]: I1005 20:56:34.836452 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 20:56:34 crc kubenswrapper[4754]: I1005 20:56:34.836520 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 20:56:34 crc kubenswrapper[4754]: I1005 20:56:34.836521 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 20:56:34 crc kubenswrapper[4754]: E1005 20:56:34.836726 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 20:56:34 crc kubenswrapper[4754]: E1005 20:56:34.837611 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 20:56:34 crc kubenswrapper[4754]: E1005 20:56:34.838011 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 20:56:35 crc kubenswrapper[4754]: I1005 20:56:35.836875 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nwrnt" Oct 05 20:56:35 crc kubenswrapper[4754]: E1005 20:56:35.837265 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nwrnt" podUID="7dc728ea-2601-44f0-bcab-2913f034007d" Oct 05 20:56:36 crc kubenswrapper[4754]: E1005 20:56:36.826707 4754 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Oct 05 20:56:36 crc kubenswrapper[4754]: I1005 20:56:36.836585 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 20:56:36 crc kubenswrapper[4754]: I1005 20:56:36.836674 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 20:56:36 crc kubenswrapper[4754]: I1005 20:56:36.836775 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 20:56:36 crc kubenswrapper[4754]: E1005 20:56:36.838374 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 20:56:36 crc kubenswrapper[4754]: E1005 20:56:36.839285 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 20:56:36 crc kubenswrapper[4754]: E1005 20:56:36.839452 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 20:56:36 crc kubenswrapper[4754]: I1005 20:56:36.842667 4754 scope.go:117] "RemoveContainer" containerID="7c990164cb7213726a9086a630586e38da00d62fdb57aed8418bbec265a54705" Oct 05 20:56:36 crc kubenswrapper[4754]: E1005 20:56:36.939310 4754 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 05 20:56:37 crc kubenswrapper[4754]: I1005 20:56:37.652701 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7rhps_b44790fb-fecb-4ec3-9816-8e23dd72ab33/ovnkube-controller/3.log" Oct 05 20:56:37 crc kubenswrapper[4754]: I1005 20:56:37.655337 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" event={"ID":"b44790fb-fecb-4ec3-9816-8e23dd72ab33","Type":"ContainerStarted","Data":"40eef1aa41324f51a533acc8eb81eb76ae6eea135681050b6b79cef39fd3446f"} Oct 05 20:56:37 crc kubenswrapper[4754]: I1005 20:56:37.655891 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" Oct 05 20:56:37 crc kubenswrapper[4754]: I1005 20:56:37.696132 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" podStartSLOduration=99.696093174 podStartE2EDuration="1m39.696093174s" podCreationTimestamp="2025-10-05 20:54:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 20:56:37.694441501 +0000 UTC m=+121.598560231" watchObservedRunningTime="2025-10-05 20:56:37.696093174 +0000 UTC m=+121.600211924" Oct 05 20:56:37 crc kubenswrapper[4754]: I1005 20:56:37.836780 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nwrnt" Oct 05 20:56:37 crc kubenswrapper[4754]: E1005 20:56:37.837017 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nwrnt" podUID="7dc728ea-2601-44f0-bcab-2913f034007d" Oct 05 20:56:37 crc kubenswrapper[4754]: I1005 20:56:37.900006 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-nwrnt"] Oct 05 20:56:38 crc kubenswrapper[4754]: I1005 20:56:38.659266 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nwrnt" Oct 05 20:56:38 crc kubenswrapper[4754]: E1005 20:56:38.660239 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nwrnt" podUID="7dc728ea-2601-44f0-bcab-2913f034007d" Oct 05 20:56:38 crc kubenswrapper[4754]: I1005 20:56:38.837242 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 20:56:38 crc kubenswrapper[4754]: I1005 20:56:38.837375 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 20:56:38 crc kubenswrapper[4754]: E1005 20:56:38.837450 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 20:56:38 crc kubenswrapper[4754]: I1005 20:56:38.837582 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 20:56:38 crc kubenswrapper[4754]: E1005 20:56:38.837817 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 20:56:38 crc kubenswrapper[4754]: E1005 20:56:38.838152 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 20:56:39 crc kubenswrapper[4754]: I1005 20:56:39.837104 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nwrnt" Oct 05 20:56:39 crc kubenswrapper[4754]: E1005 20:56:39.838611 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nwrnt" podUID="7dc728ea-2601-44f0-bcab-2913f034007d" Oct 05 20:56:40 crc kubenswrapper[4754]: I1005 20:56:40.836719 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 20:56:40 crc kubenswrapper[4754]: I1005 20:56:40.836732 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 20:56:40 crc kubenswrapper[4754]: I1005 20:56:40.836795 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 20:56:40 crc kubenswrapper[4754]: E1005 20:56:40.837427 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 20:56:40 crc kubenswrapper[4754]: E1005 20:56:40.838337 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 20:56:40 crc kubenswrapper[4754]: E1005 20:56:40.838595 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 20:56:41 crc kubenswrapper[4754]: I1005 20:56:41.837709 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nwrnt" Oct 05 20:56:41 crc kubenswrapper[4754]: E1005 20:56:41.837935 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nwrnt" podUID="7dc728ea-2601-44f0-bcab-2913f034007d" Oct 05 20:56:41 crc kubenswrapper[4754]: E1005 20:56:41.940929 4754 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 05 20:56:42 crc kubenswrapper[4754]: I1005 20:56:42.836979 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 20:56:42 crc kubenswrapper[4754]: I1005 20:56:42.837095 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 20:56:42 crc kubenswrapper[4754]: E1005 20:56:42.837208 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 20:56:42 crc kubenswrapper[4754]: E1005 20:56:42.837302 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 20:56:42 crc kubenswrapper[4754]: I1005 20:56:42.837405 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 20:56:42 crc kubenswrapper[4754]: E1005 20:56:42.837565 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 20:56:43 crc kubenswrapper[4754]: I1005 20:56:43.837370 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nwrnt" Oct 05 20:56:43 crc kubenswrapper[4754]: E1005 20:56:43.837653 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nwrnt" podUID="7dc728ea-2601-44f0-bcab-2913f034007d" Oct 05 20:56:44 crc kubenswrapper[4754]: I1005 20:56:44.836711 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 20:56:44 crc kubenswrapper[4754]: I1005 20:56:44.836775 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 20:56:44 crc kubenswrapper[4754]: I1005 20:56:44.836819 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 20:56:44 crc kubenswrapper[4754]: E1005 20:56:44.836939 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 20:56:44 crc kubenswrapper[4754]: E1005 20:56:44.837096 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 20:56:44 crc kubenswrapper[4754]: E1005 20:56:44.837419 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 20:56:44 crc kubenswrapper[4754]: I1005 20:56:44.838137 4754 scope.go:117] "RemoveContainer" containerID="5ed23b9cd9b57bb169a8ebbf16b7ff8d196eb14330df0edf2b9fcd2a8c9e2c88" Oct 05 20:56:45 crc kubenswrapper[4754]: I1005 20:56:45.699446 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-cn76s_02d5e3f9-73c2-4496-9aca-0787184aef19/kube-multus/1.log" Oct 05 20:56:45 crc kubenswrapper[4754]: I1005 20:56:45.700040 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-cn76s" event={"ID":"02d5e3f9-73c2-4496-9aca-0787184aef19","Type":"ContainerStarted","Data":"b31d3403864365b964b38f903edb9d1f9d1567917a71eafd5a7120faf14feb99"} Oct 05 20:56:45 crc kubenswrapper[4754]: I1005 20:56:45.836806 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nwrnt" Oct 05 20:56:45 crc kubenswrapper[4754]: E1005 20:56:45.837028 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nwrnt" podUID="7dc728ea-2601-44f0-bcab-2913f034007d" Oct 05 20:56:46 crc kubenswrapper[4754]: I1005 20:56:46.836818 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 20:56:46 crc kubenswrapper[4754]: I1005 20:56:46.836851 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 20:56:46 crc kubenswrapper[4754]: E1005 20:56:46.837019 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 20:56:46 crc kubenswrapper[4754]: E1005 20:56:46.838582 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 20:56:46 crc kubenswrapper[4754]: I1005 20:56:46.838687 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 20:56:46 crc kubenswrapper[4754]: E1005 20:56:46.838819 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 20:56:47 crc kubenswrapper[4754]: I1005 20:56:47.836252 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nwrnt" Oct 05 20:56:47 crc kubenswrapper[4754]: I1005 20:56:47.839021 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Oct 05 20:56:47 crc kubenswrapper[4754]: I1005 20:56:47.840407 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Oct 05 20:56:48 crc kubenswrapper[4754]: I1005 20:56:48.836564 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 20:56:48 crc kubenswrapper[4754]: I1005 20:56:48.836807 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 20:56:48 crc kubenswrapper[4754]: I1005 20:56:48.837594 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 20:56:48 crc kubenswrapper[4754]: I1005 20:56:48.840551 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Oct 05 20:56:48 crc kubenswrapper[4754]: I1005 20:56:48.840604 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Oct 05 20:56:48 crc kubenswrapper[4754]: I1005 20:56:48.840671 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Oct 05 20:56:48 crc kubenswrapper[4754]: I1005 20:56:48.840812 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.753257 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.813285 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-k8qff"] Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.814219 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-k8qff" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.817119 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-rvsbj"] Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.818064 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-rvsbj" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.818097 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-hm4ch"] Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.818877 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hm4ch" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.828752 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tcjmf"] Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.829348 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-9chlx"] Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.829773 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-p5wqw"] Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.830391 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-p5wqw" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.830979 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tcjmf" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.831343 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-9chlx" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.848129 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.852982 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.860618 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.861093 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.862525 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-5sffm"] Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.862905 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-5sffm" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.863226 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.864319 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.864339 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.864574 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.864598 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.864722 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.864788 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.864858 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.864953 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.865056 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.865169 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.865264 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.865355 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.865462 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.865684 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.865795 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.865874 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.865915 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.867611 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.867850 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.870570 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.871718 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.872523 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.873014 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.873791 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.873842 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-lxp7v"] Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.865874 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.886178 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lxp7v" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.899486 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.899597 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.899855 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.899904 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.900031 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.900040 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.900240 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.900510 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-cwtlx"] Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.901124 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-cwtlx" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.902126 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.902245 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.902451 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-pnhf9"] Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.902735 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-pnhf9" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.906104 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.906141 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.906302 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.906410 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.906466 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.906518 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.906583 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.906610 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.906687 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.906694 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.906764 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.906777 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.906864 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.906890 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.906699 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.906997 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.907083 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-ksczt"] Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.909976 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.910575 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.910791 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.911008 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-ksczt" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.911221 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.911323 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.911426 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.911475 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.911681 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.911943 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.912582 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.913504 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.918955 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.919849 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.924901 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-5scsb"] Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.925716 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-5scsb" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.927670 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9nms8"] Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.928291 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9nms8" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.930354 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-gchnw"] Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.930887 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-gchnw" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.931328 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.931529 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.931684 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.931741 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.932566 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-22htd"] Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.932939 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-q2jq4"] Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.933080 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.933279 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.933337 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-q2jq4" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.933449 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-22htd" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.934696 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.934843 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-5sqj9"] Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.935409 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-5sqj9" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.949837 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wx9l2"] Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.950466 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-2qmwc"] Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.951161 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-2qmwc" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.951456 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wx9l2" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.951642 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-99kwp\" (UniqueName: \"kubernetes.io/projected/ef8c2ff8-db1e-44ce-b489-daa0832720c3-kube-api-access-99kwp\") pod \"console-operator-58897d9998-9chlx\" (UID: \"ef8c2ff8-db1e-44ce-b489-daa0832720c3\") " pod="openshift-console-operator/console-operator-58897d9998-9chlx" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.951670 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/c56f5eb8-3a12-428c-b347-78369dc19008-service-ca\") pod \"console-f9d7485db-5sffm\" (UID: \"c56f5eb8-3a12-428c-b347-78369dc19008\") " pod="openshift-console/console-f9d7485db-5sffm" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.951713 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b1286ed6-4e15-46e0-b4c3-3b7e3cad6057-serving-cert\") pod \"controller-manager-879f6c89f-rvsbj\" (UID: \"b1286ed6-4e15-46e0-b4c3-3b7e3cad6057\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rvsbj" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.951737 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/20a5d042-ffe2-4244-b25d-700b26534a04-node-pullsecrets\") pod \"apiserver-76f77b778f-k8qff\" (UID: \"20a5d042-ffe2-4244-b25d-700b26534a04\") " pod="openshift-apiserver/apiserver-76f77b778f-k8qff" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.951768 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/c56f5eb8-3a12-428c-b347-78369dc19008-console-serving-cert\") pod \"console-f9d7485db-5sffm\" (UID: \"c56f5eb8-3a12-428c-b347-78369dc19008\") " pod="openshift-console/console-f9d7485db-5sffm" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.951792 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/c463939d-d2d1-4049-b438-3754a118ff4b-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-p5wqw\" (UID: \"c463939d-d2d1-4049-b438-3754a118ff4b\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-p5wqw" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.951864 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/20a5d042-ffe2-4244-b25d-700b26534a04-audit\") pod \"apiserver-76f77b778f-k8qff\" (UID: \"20a5d042-ffe2-4244-b25d-700b26534a04\") " pod="openshift-apiserver/apiserver-76f77b778f-k8qff" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.951887 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/20a5d042-ffe2-4244-b25d-700b26534a04-audit-dir\") pod \"apiserver-76f77b778f-k8qff\" (UID: \"20a5d042-ffe2-4244-b25d-700b26534a04\") " pod="openshift-apiserver/apiserver-76f77b778f-k8qff" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.951908 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/c56f5eb8-3a12-428c-b347-78369dc19008-console-oauth-config\") pod \"console-f9d7485db-5sffm\" (UID: \"c56f5eb8-3a12-428c-b347-78369dc19008\") " pod="openshift-console/console-f9d7485db-5sffm" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.951935 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/c56f5eb8-3a12-428c-b347-78369dc19008-oauth-serving-cert\") pod \"console-f9d7485db-5sffm\" (UID: \"c56f5eb8-3a12-428c-b347-78369dc19008\") " pod="openshift-console/console-f9d7485db-5sffm" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.951968 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nnwxh\" (UniqueName: \"kubernetes.io/projected/b1286ed6-4e15-46e0-b4c3-3b7e3cad6057-kube-api-access-nnwxh\") pod \"controller-manager-879f6c89f-rvsbj\" (UID: \"b1286ed6-4e15-46e0-b4c3-3b7e3cad6057\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rvsbj" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.951989 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/c463939d-d2d1-4049-b438-3754a118ff4b-images\") pod \"machine-api-operator-5694c8668f-p5wqw\" (UID: \"c463939d-d2d1-4049-b438-3754a118ff4b\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-p5wqw" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.952011 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3d4ca0c2-4e63-425f-bdb6-89da1a2e04c5-config\") pod \"machine-approver-56656f9798-hm4ch\" (UID: \"3d4ca0c2-4e63-425f-bdb6-89da1a2e04c5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hm4ch" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.952033 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/20a5d042-ffe2-4244-b25d-700b26534a04-encryption-config\") pod \"apiserver-76f77b778f-k8qff\" (UID: \"20a5d042-ffe2-4244-b25d-700b26534a04\") " pod="openshift-apiserver/apiserver-76f77b778f-k8qff" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.952062 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m66bm\" (UniqueName: \"kubernetes.io/projected/c56f5eb8-3a12-428c-b347-78369dc19008-kube-api-access-m66bm\") pod \"console-f9d7485db-5sffm\" (UID: \"c56f5eb8-3a12-428c-b347-78369dc19008\") " pod="openshift-console/console-f9d7485db-5sffm" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.952084 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/20a5d042-ffe2-4244-b25d-700b26534a04-etcd-serving-ca\") pod \"apiserver-76f77b778f-k8qff\" (UID: \"20a5d042-ffe2-4244-b25d-700b26534a04\") " pod="openshift-apiserver/apiserver-76f77b778f-k8qff" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.952102 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cpf6q\" (UniqueName: \"kubernetes.io/projected/20a5d042-ffe2-4244-b25d-700b26534a04-kube-api-access-cpf6q\") pod \"apiserver-76f77b778f-k8qff\" (UID: \"20a5d042-ffe2-4244-b25d-700b26534a04\") " pod="openshift-apiserver/apiserver-76f77b778f-k8qff" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.952124 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ef8c2ff8-db1e-44ce-b489-daa0832720c3-trusted-ca\") pod \"console-operator-58897d9998-9chlx\" (UID: \"ef8c2ff8-db1e-44ce-b489-daa0832720c3\") " pod="openshift-console-operator/console-operator-58897d9998-9chlx" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.952269 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/20a5d042-ffe2-4244-b25d-700b26534a04-trusted-ca-bundle\") pod \"apiserver-76f77b778f-k8qff\" (UID: \"20a5d042-ffe2-4244-b25d-700b26534a04\") " pod="openshift-apiserver/apiserver-76f77b778f-k8qff" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.952294 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ef8c2ff8-db1e-44ce-b489-daa0832720c3-config\") pod \"console-operator-58897d9998-9chlx\" (UID: \"ef8c2ff8-db1e-44ce-b489-daa0832720c3\") " pod="openshift-console-operator/console-operator-58897d9998-9chlx" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.952314 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b1286ed6-4e15-46e0-b4c3-3b7e3cad6057-client-ca\") pod \"controller-manager-879f6c89f-rvsbj\" (UID: \"b1286ed6-4e15-46e0-b4c3-3b7e3cad6057\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rvsbj" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.952340 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ef8c2ff8-db1e-44ce-b489-daa0832720c3-serving-cert\") pod \"console-operator-58897d9998-9chlx\" (UID: \"ef8c2ff8-db1e-44ce-b489-daa0832720c3\") " pod="openshift-console-operator/console-operator-58897d9998-9chlx" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.952362 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b1286ed6-4e15-46e0-b4c3-3b7e3cad6057-config\") pod \"controller-manager-879f6c89f-rvsbj\" (UID: \"b1286ed6-4e15-46e0-b4c3-3b7e3cad6057\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rvsbj" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.952385 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/20a5d042-ffe2-4244-b25d-700b26534a04-image-import-ca\") pod \"apiserver-76f77b778f-k8qff\" (UID: \"20a5d042-ffe2-4244-b25d-700b26534a04\") " pod="openshift-apiserver/apiserver-76f77b778f-k8qff" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.952405 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c56f5eb8-3a12-428c-b347-78369dc19008-trusted-ca-bundle\") pod \"console-f9d7485db-5sffm\" (UID: \"c56f5eb8-3a12-428c-b347-78369dc19008\") " pod="openshift-console/console-f9d7485db-5sffm" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.952422 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/20a5d042-ffe2-4244-b25d-700b26534a04-serving-cert\") pod \"apiserver-76f77b778f-k8qff\" (UID: \"20a5d042-ffe2-4244-b25d-700b26534a04\") " pod="openshift-apiserver/apiserver-76f77b778f-k8qff" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.952442 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/c56f5eb8-3a12-428c-b347-78369dc19008-console-config\") pod \"console-f9d7485db-5sffm\" (UID: \"c56f5eb8-3a12-428c-b347-78369dc19008\") " pod="openshift-console/console-f9d7485db-5sffm" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.952467 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/20a5d042-ffe2-4244-b25d-700b26534a04-etcd-client\") pod \"apiserver-76f77b778f-k8qff\" (UID: \"20a5d042-ffe2-4244-b25d-700b26534a04\") " pod="openshift-apiserver/apiserver-76f77b778f-k8qff" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.952516 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/20a5d042-ffe2-4244-b25d-700b26534a04-config\") pod \"apiserver-76f77b778f-k8qff\" (UID: \"20a5d042-ffe2-4244-b25d-700b26534a04\") " pod="openshift-apiserver/apiserver-76f77b778f-k8qff" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.952538 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/3d4ca0c2-4e63-425f-bdb6-89da1a2e04c5-auth-proxy-config\") pod \"machine-approver-56656f9798-hm4ch\" (UID: \"3d4ca0c2-4e63-425f-bdb6-89da1a2e04c5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hm4ch" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.952579 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/afbf61d3-201e-4326-8463-4c660057a865-config\") pod \"openshift-apiserver-operator-796bbdcf4f-tcjmf\" (UID: \"afbf61d3-201e-4326-8463-4c660057a865\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tcjmf" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.952613 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kzbj2\" (UniqueName: \"kubernetes.io/projected/afbf61d3-201e-4326-8463-4c660057a865-kube-api-access-kzbj2\") pod \"openshift-apiserver-operator-796bbdcf4f-tcjmf\" (UID: \"afbf61d3-201e-4326-8463-4c660057a865\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tcjmf" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.952638 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b1286ed6-4e15-46e0-b4c3-3b7e3cad6057-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-rvsbj\" (UID: \"b1286ed6-4e15-46e0-b4c3-3b7e3cad6057\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rvsbj" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.952665 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pk76j\" (UniqueName: \"kubernetes.io/projected/c463939d-d2d1-4049-b438-3754a118ff4b-kube-api-access-pk76j\") pod \"machine-api-operator-5694c8668f-p5wqw\" (UID: \"c463939d-d2d1-4049-b438-3754a118ff4b\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-p5wqw" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.952873 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/afbf61d3-201e-4326-8463-4c660057a865-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-tcjmf\" (UID: \"afbf61d3-201e-4326-8463-4c660057a865\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tcjmf" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.952906 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9b4kw\" (UniqueName: \"kubernetes.io/projected/3d4ca0c2-4e63-425f-bdb6-89da1a2e04c5-kube-api-access-9b4kw\") pod \"machine-approver-56656f9798-hm4ch\" (UID: \"3d4ca0c2-4e63-425f-bdb6-89da1a2e04c5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hm4ch" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.952949 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/3d4ca0c2-4e63-425f-bdb6-89da1a2e04c5-machine-approver-tls\") pod \"machine-approver-56656f9798-hm4ch\" (UID: \"3d4ca0c2-4e63-425f-bdb6-89da1a2e04c5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hm4ch" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.952980 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c463939d-d2d1-4049-b438-3754a118ff4b-config\") pod \"machine-api-operator-5694c8668f-p5wqw\" (UID: \"c463939d-d2d1-4049-b438-3754a118ff4b\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-p5wqw" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.957696 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-k8qff"] Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.959626 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.960527 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.961334 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.975602 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.976844 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.977019 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.977065 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.977113 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.977076 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.977406 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.978611 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.978702 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.978885 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.979032 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.979078 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.979106 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.979228 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.979286 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.979385 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.983568 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.983906 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.984321 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.985096 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.986744 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.993618 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.997672 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-8qm2l"] Oct 05 20:56:56 crc kubenswrapper[4754]: I1005 20:56:56.998448 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-8qm2l" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.002206 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.005785 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.006576 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6wnmn"] Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.007224 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-w6krf"] Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.007613 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-9vndv"] Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.008062 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.009765 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6wnmn" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.011725 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-hjpzm"] Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.011726 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.011927 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-w6krf" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.012085 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-9vndv" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.012706 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-hjpzm" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.013367 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.016743 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-rvsbj"] Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.028279 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.039050 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-wz2dg"] Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.048619 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.051369 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-ll5jq"] Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.051769 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-682zv"] Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.051915 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-wz2dg" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.051992 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-ll5jq" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.052504 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rl5n7"] Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.052979 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rl5n7" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.053973 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-682zv" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.055386 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/20a5d042-ffe2-4244-b25d-700b26534a04-config\") pod \"apiserver-76f77b778f-k8qff\" (UID: \"20a5d042-ffe2-4244-b25d-700b26534a04\") " pod="openshift-apiserver/apiserver-76f77b778f-k8qff" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.055424 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/9e7db3dd-7fae-469a-9ca5-eba1d965cbe8-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-9nms8\" (UID: \"9e7db3dd-7fae-469a-9ca5-eba1d965cbe8\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9nms8" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.055456 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/3d4ca0c2-4e63-425f-bdb6-89da1a2e04c5-auth-proxy-config\") pod \"machine-approver-56656f9798-hm4ch\" (UID: \"3d4ca0c2-4e63-425f-bdb6-89da1a2e04c5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hm4ch" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.055476 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/afbf61d3-201e-4326-8463-4c660057a865-config\") pod \"openshift-apiserver-operator-796bbdcf4f-tcjmf\" (UID: \"afbf61d3-201e-4326-8463-4c660057a865\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tcjmf" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.055512 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kzbj2\" (UniqueName: \"kubernetes.io/projected/afbf61d3-201e-4326-8463-4c660057a865-kube-api-access-kzbj2\") pod \"openshift-apiserver-operator-796bbdcf4f-tcjmf\" (UID: \"afbf61d3-201e-4326-8463-4c660057a865\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tcjmf" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.055539 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jlsgz\" (UniqueName: \"kubernetes.io/projected/5a98537a-186b-4150-bd41-c2627ec772f9-kube-api-access-jlsgz\") pod \"cluster-samples-operator-665b6dd947-5scsb\" (UID: \"5a98537a-186b-4150-bd41-c2627ec772f9\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-5scsb" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.055559 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7n2wn\" (UniqueName: \"kubernetes.io/projected/811694fe-dcfd-4fe3-876c-964d09321d2f-kube-api-access-7n2wn\") pod \"authentication-operator-69f744f599-22htd\" (UID: \"811694fe-dcfd-4fe3-876c-964d09321d2f\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-22htd" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.055582 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b1286ed6-4e15-46e0-b4c3-3b7e3cad6057-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-rvsbj\" (UID: \"b1286ed6-4e15-46e0-b4c3-3b7e3cad6057\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rvsbj" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.059243 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pk76j\" (UniqueName: \"kubernetes.io/projected/c463939d-d2d1-4049-b438-3754a118ff4b-kube-api-access-pk76j\") pod \"machine-api-operator-5694c8668f-p5wqw\" (UID: \"c463939d-d2d1-4049-b438-3754a118ff4b\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-p5wqw" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.059299 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/afbf61d3-201e-4326-8463-4c660057a865-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-tcjmf\" (UID: \"afbf61d3-201e-4326-8463-4c660057a865\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tcjmf" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.059324 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9b4kw\" (UniqueName: \"kubernetes.io/projected/3d4ca0c2-4e63-425f-bdb6-89da1a2e04c5-kube-api-access-9b4kw\") pod \"machine-approver-56656f9798-hm4ch\" (UID: \"3d4ca0c2-4e63-425f-bdb6-89da1a2e04c5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hm4ch" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.059411 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/3d4ca0c2-4e63-425f-bdb6-89da1a2e04c5-machine-approver-tls\") pod \"machine-approver-56656f9798-hm4ch\" (UID: \"3d4ca0c2-4e63-425f-bdb6-89da1a2e04c5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hm4ch" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.059456 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c463939d-d2d1-4049-b438-3754a118ff4b-config\") pod \"machine-api-operator-5694c8668f-p5wqw\" (UID: \"c463939d-d2d1-4049-b438-3754a118ff4b\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-p5wqw" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.059484 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-99kwp\" (UniqueName: \"kubernetes.io/projected/ef8c2ff8-db1e-44ce-b489-daa0832720c3-kube-api-access-99kwp\") pod \"console-operator-58897d9998-9chlx\" (UID: \"ef8c2ff8-db1e-44ce-b489-daa0832720c3\") " pod="openshift-console-operator/console-operator-58897d9998-9chlx" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.059714 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-ll9lb"] Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.060429 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b1286ed6-4e15-46e0-b4c3-3b7e3cad6057-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-rvsbj\" (UID: \"b1286ed6-4e15-46e0-b4c3-3b7e3cad6057\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rvsbj" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.060850 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.062453 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/20a5d042-ffe2-4244-b25d-700b26534a04-config\") pod \"apiserver-76f77b778f-k8qff\" (UID: \"20a5d042-ffe2-4244-b25d-700b26534a04\") " pod="openshift-apiserver/apiserver-76f77b778f-k8qff" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.067028 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c463939d-d2d1-4049-b438-3754a118ff4b-config\") pod \"machine-api-operator-5694c8668f-p5wqw\" (UID: \"c463939d-d2d1-4049-b438-3754a118ff4b\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-p5wqw" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.067630 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/3d4ca0c2-4e63-425f-bdb6-89da1a2e04c5-auth-proxy-config\") pod \"machine-approver-56656f9798-hm4ch\" (UID: \"3d4ca0c2-4e63-425f-bdb6-89da1a2e04c5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hm4ch" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.069507 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/c56f5eb8-3a12-428c-b347-78369dc19008-service-ca\") pod \"console-f9d7485db-5sffm\" (UID: \"c56f5eb8-3a12-428c-b347-78369dc19008\") " pod="openshift-console/console-f9d7485db-5sffm" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.069604 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/5a98537a-186b-4150-bd41-c2627ec772f9-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-5scsb\" (UID: \"5a98537a-186b-4150-bd41-c2627ec772f9\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-5scsb" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.069661 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pwqsk\" (UniqueName: \"kubernetes.io/projected/9e7db3dd-7fae-469a-9ca5-eba1d965cbe8-kube-api-access-pwqsk\") pod \"cluster-image-registry-operator-dc59b4c8b-9nms8\" (UID: \"9e7db3dd-7fae-469a-9ca5-eba1d965cbe8\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9nms8" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.069709 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b1286ed6-4e15-46e0-b4c3-3b7e3cad6057-serving-cert\") pod \"controller-manager-879f6c89f-rvsbj\" (UID: \"b1286ed6-4e15-46e0-b4c3-3b7e3cad6057\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rvsbj" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.069737 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/20a5d042-ffe2-4244-b25d-700b26534a04-node-pullsecrets\") pod \"apiserver-76f77b778f-k8qff\" (UID: \"20a5d042-ffe2-4244-b25d-700b26534a04\") " pod="openshift-apiserver/apiserver-76f77b778f-k8qff" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.069769 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/c56f5eb8-3a12-428c-b347-78369dc19008-console-serving-cert\") pod \"console-f9d7485db-5sffm\" (UID: \"c56f5eb8-3a12-428c-b347-78369dc19008\") " pod="openshift-console/console-f9d7485db-5sffm" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.069789 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/afbf61d3-201e-4326-8463-4c660057a865-config\") pod \"openshift-apiserver-operator-796bbdcf4f-tcjmf\" (UID: \"afbf61d3-201e-4326-8463-4c660057a865\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tcjmf" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.069799 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/c463939d-d2d1-4049-b438-3754a118ff4b-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-p5wqw\" (UID: \"c463939d-d2d1-4049-b438-3754a118ff4b\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-p5wqw" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.069866 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/20a5d042-ffe2-4244-b25d-700b26534a04-audit\") pod \"apiserver-76f77b778f-k8qff\" (UID: \"20a5d042-ffe2-4244-b25d-700b26534a04\") " pod="openshift-apiserver/apiserver-76f77b778f-k8qff" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.069903 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/20a5d042-ffe2-4244-b25d-700b26534a04-audit-dir\") pod \"apiserver-76f77b778f-k8qff\" (UID: \"20a5d042-ffe2-4244-b25d-700b26534a04\") " pod="openshift-apiserver/apiserver-76f77b778f-k8qff" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.069924 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/c56f5eb8-3a12-428c-b347-78369dc19008-console-oauth-config\") pod \"console-f9d7485db-5sffm\" (UID: \"c56f5eb8-3a12-428c-b347-78369dc19008\") " pod="openshift-console/console-f9d7485db-5sffm" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.069953 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/c56f5eb8-3a12-428c-b347-78369dc19008-oauth-serving-cert\") pod \"console-f9d7485db-5sffm\" (UID: \"c56f5eb8-3a12-428c-b347-78369dc19008\") " pod="openshift-console/console-f9d7485db-5sffm" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.070006 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nnwxh\" (UniqueName: \"kubernetes.io/projected/b1286ed6-4e15-46e0-b4c3-3b7e3cad6057-kube-api-access-nnwxh\") pod \"controller-manager-879f6c89f-rvsbj\" (UID: \"b1286ed6-4e15-46e0-b4c3-3b7e3cad6057\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rvsbj" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.070028 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/c463939d-d2d1-4049-b438-3754a118ff4b-images\") pod \"machine-api-operator-5694c8668f-p5wqw\" (UID: \"c463939d-d2d1-4049-b438-3754a118ff4b\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-p5wqw" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.070072 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/811694fe-dcfd-4fe3-876c-964d09321d2f-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-22htd\" (UID: \"811694fe-dcfd-4fe3-876c-964d09321d2f\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-22htd" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.070106 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3d4ca0c2-4e63-425f-bdb6-89da1a2e04c5-config\") pod \"machine-approver-56656f9798-hm4ch\" (UID: \"3d4ca0c2-4e63-425f-bdb6-89da1a2e04c5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hm4ch" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.070138 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/20a5d042-ffe2-4244-b25d-700b26534a04-encryption-config\") pod \"apiserver-76f77b778f-k8qff\" (UID: \"20a5d042-ffe2-4244-b25d-700b26534a04\") " pod="openshift-apiserver/apiserver-76f77b778f-k8qff" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.070174 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m66bm\" (UniqueName: \"kubernetes.io/projected/c56f5eb8-3a12-428c-b347-78369dc19008-kube-api-access-m66bm\") pod \"console-f9d7485db-5sffm\" (UID: \"c56f5eb8-3a12-428c-b347-78369dc19008\") " pod="openshift-console/console-f9d7485db-5sffm" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.070201 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/811694fe-dcfd-4fe3-876c-964d09321d2f-config\") pod \"authentication-operator-69f744f599-22htd\" (UID: \"811694fe-dcfd-4fe3-876c-964d09321d2f\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-22htd" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.070226 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9e7db3dd-7fae-469a-9ca5-eba1d965cbe8-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-9nms8\" (UID: \"9e7db3dd-7fae-469a-9ca5-eba1d965cbe8\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9nms8" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.070257 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/20a5d042-ffe2-4244-b25d-700b26534a04-etcd-serving-ca\") pod \"apiserver-76f77b778f-k8qff\" (UID: \"20a5d042-ffe2-4244-b25d-700b26534a04\") " pod="openshift-apiserver/apiserver-76f77b778f-k8qff" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.070285 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cpf6q\" (UniqueName: \"kubernetes.io/projected/20a5d042-ffe2-4244-b25d-700b26534a04-kube-api-access-cpf6q\") pod \"apiserver-76f77b778f-k8qff\" (UID: \"20a5d042-ffe2-4244-b25d-700b26534a04\") " pod="openshift-apiserver/apiserver-76f77b778f-k8qff" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.070313 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ef8c2ff8-db1e-44ce-b489-daa0832720c3-trusted-ca\") pod \"console-operator-58897d9998-9chlx\" (UID: \"ef8c2ff8-db1e-44ce-b489-daa0832720c3\") " pod="openshift-console-operator/console-operator-58897d9998-9chlx" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.070345 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/20a5d042-ffe2-4244-b25d-700b26534a04-trusted-ca-bundle\") pod \"apiserver-76f77b778f-k8qff\" (UID: \"20a5d042-ffe2-4244-b25d-700b26534a04\") " pod="openshift-apiserver/apiserver-76f77b778f-k8qff" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.070373 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9e7db3dd-7fae-469a-9ca5-eba1d965cbe8-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-9nms8\" (UID: \"9e7db3dd-7fae-469a-9ca5-eba1d965cbe8\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9nms8" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.070403 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/811694fe-dcfd-4fe3-876c-964d09321d2f-serving-cert\") pod \"authentication-operator-69f744f599-22htd\" (UID: \"811694fe-dcfd-4fe3-876c-964d09321d2f\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-22htd" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.070425 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ef8c2ff8-db1e-44ce-b489-daa0832720c3-config\") pod \"console-operator-58897d9998-9chlx\" (UID: \"ef8c2ff8-db1e-44ce-b489-daa0832720c3\") " pod="openshift-console-operator/console-operator-58897d9998-9chlx" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.070459 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b1286ed6-4e15-46e0-b4c3-3b7e3cad6057-client-ca\") pod \"controller-manager-879f6c89f-rvsbj\" (UID: \"b1286ed6-4e15-46e0-b4c3-3b7e3cad6057\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rvsbj" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.070484 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/811694fe-dcfd-4fe3-876c-964d09321d2f-service-ca-bundle\") pod \"authentication-operator-69f744f599-22htd\" (UID: \"811694fe-dcfd-4fe3-876c-964d09321d2f\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-22htd" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.070535 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ef8c2ff8-db1e-44ce-b489-daa0832720c3-serving-cert\") pod \"console-operator-58897d9998-9chlx\" (UID: \"ef8c2ff8-db1e-44ce-b489-daa0832720c3\") " pod="openshift-console-operator/console-operator-58897d9998-9chlx" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.070563 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b1286ed6-4e15-46e0-b4c3-3b7e3cad6057-config\") pod \"controller-manager-879f6c89f-rvsbj\" (UID: \"b1286ed6-4e15-46e0-b4c3-3b7e3cad6057\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rvsbj" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.070589 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/20a5d042-ffe2-4244-b25d-700b26534a04-image-import-ca\") pod \"apiserver-76f77b778f-k8qff\" (UID: \"20a5d042-ffe2-4244-b25d-700b26534a04\") " pod="openshift-apiserver/apiserver-76f77b778f-k8qff" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.070644 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c56f5eb8-3a12-428c-b347-78369dc19008-trusted-ca-bundle\") pod \"console-f9d7485db-5sffm\" (UID: \"c56f5eb8-3a12-428c-b347-78369dc19008\") " pod="openshift-console/console-f9d7485db-5sffm" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.070671 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/20a5d042-ffe2-4244-b25d-700b26534a04-serving-cert\") pod \"apiserver-76f77b778f-k8qff\" (UID: \"20a5d042-ffe2-4244-b25d-700b26534a04\") " pod="openshift-apiserver/apiserver-76f77b778f-k8qff" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.070698 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/c56f5eb8-3a12-428c-b347-78369dc19008-console-config\") pod \"console-f9d7485db-5sffm\" (UID: \"c56f5eb8-3a12-428c-b347-78369dc19008\") " pod="openshift-console/console-f9d7485db-5sffm" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.070718 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/20a5d042-ffe2-4244-b25d-700b26534a04-etcd-client\") pod \"apiserver-76f77b778f-k8qff\" (UID: \"20a5d042-ffe2-4244-b25d-700b26534a04\") " pod="openshift-apiserver/apiserver-76f77b778f-k8qff" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.071253 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/c56f5eb8-3a12-428c-b347-78369dc19008-service-ca\") pod \"console-f9d7485db-5sffm\" (UID: \"c56f5eb8-3a12-428c-b347-78369dc19008\") " pod="openshift-console/console-f9d7485db-5sffm" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.072295 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tcjmf"] Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.072332 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-5wx56"] Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.072988 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3d4ca0c2-4e63-425f-bdb6-89da1a2e04c5-config\") pod \"machine-approver-56656f9798-hm4ch\" (UID: \"3d4ca0c2-4e63-425f-bdb6-89da1a2e04c5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hm4ch" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.073728 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-ll9lb" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.074101 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-5wx56" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.074980 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nrd6b"] Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.075603 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/20a5d042-ffe2-4244-b25d-700b26534a04-node-pullsecrets\") pod \"apiserver-76f77b778f-k8qff\" (UID: \"20a5d042-ffe2-4244-b25d-700b26534a04\") " pod="openshift-apiserver/apiserver-76f77b778f-k8qff" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.077976 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nrd6b" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.078943 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.084427 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/c56f5eb8-3a12-428c-b347-78369dc19008-oauth-serving-cert\") pod \"console-f9d7485db-5sffm\" (UID: \"c56f5eb8-3a12-428c-b347-78369dc19008\") " pod="openshift-console/console-f9d7485db-5sffm" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.085834 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b1286ed6-4e15-46e0-b4c3-3b7e3cad6057-client-ca\") pod \"controller-manager-879f6c89f-rvsbj\" (UID: \"b1286ed6-4e15-46e0-b4c3-3b7e3cad6057\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rvsbj" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.087989 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-67bjd"] Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.095905 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/20a5d042-ffe2-4244-b25d-700b26534a04-etcd-serving-ca\") pod \"apiserver-76f77b778f-k8qff\" (UID: \"20a5d042-ffe2-4244-b25d-700b26534a04\") " pod="openshift-apiserver/apiserver-76f77b778f-k8qff" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.100023 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ef8c2ff8-db1e-44ce-b489-daa0832720c3-config\") pod \"console-operator-58897d9998-9chlx\" (UID: \"ef8c2ff8-db1e-44ce-b489-daa0832720c3\") " pod="openshift-console-operator/console-operator-58897d9998-9chlx" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.100786 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-9chlx"] Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.100820 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jpbfr"] Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.101708 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b1286ed6-4e15-46e0-b4c3-3b7e3cad6057-config\") pod \"controller-manager-879f6c89f-rvsbj\" (UID: \"b1286ed6-4e15-46e0-b4c3-3b7e3cad6057\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rvsbj" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.101822 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ef8c2ff8-db1e-44ce-b489-daa0832720c3-trusted-ca\") pod \"console-operator-58897d9998-9chlx\" (UID: \"ef8c2ff8-db1e-44ce-b489-daa0832720c3\") " pod="openshift-console-operator/console-operator-58897d9998-9chlx" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.102278 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jpbfr" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.103205 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-67bjd" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.103899 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ef8c2ff8-db1e-44ce-b489-daa0832720c3-serving-cert\") pod \"console-operator-58897d9998-9chlx\" (UID: \"ef8c2ff8-db1e-44ce-b489-daa0832720c3\") " pod="openshift-console-operator/console-operator-58897d9998-9chlx" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.104333 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/20a5d042-ffe2-4244-b25d-700b26534a04-image-import-ca\") pod \"apiserver-76f77b778f-k8qff\" (UID: \"20a5d042-ffe2-4244-b25d-700b26534a04\") " pod="openshift-apiserver/apiserver-76f77b778f-k8qff" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.104428 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29328285-mrswx"] Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.104669 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/20a5d042-ffe2-4244-b25d-700b26534a04-trusted-ca-bundle\") pod \"apiserver-76f77b778f-k8qff\" (UID: \"20a5d042-ffe2-4244-b25d-700b26534a04\") " pod="openshift-apiserver/apiserver-76f77b778f-k8qff" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.106553 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/c56f5eb8-3a12-428c-b347-78369dc19008-console-config\") pod \"console-f9d7485db-5sffm\" (UID: \"c56f5eb8-3a12-428c-b347-78369dc19008\") " pod="openshift-console/console-f9d7485db-5sffm" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.107230 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/c56f5eb8-3a12-428c-b347-78369dc19008-console-serving-cert\") pod \"console-f9d7485db-5sffm\" (UID: \"c56f5eb8-3a12-428c-b347-78369dc19008\") " pod="openshift-console/console-f9d7485db-5sffm" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.107664 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/c463939d-d2d1-4049-b438-3754a118ff4b-images\") pod \"machine-api-operator-5694c8668f-p5wqw\" (UID: \"c463939d-d2d1-4049-b438-3754a118ff4b\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-p5wqw" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.108109 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/afbf61d3-201e-4326-8463-4c660057a865-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-tcjmf\" (UID: \"afbf61d3-201e-4326-8463-4c660057a865\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tcjmf" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.108308 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/20a5d042-ffe2-4244-b25d-700b26534a04-encryption-config\") pod \"apiserver-76f77b778f-k8qff\" (UID: \"20a5d042-ffe2-4244-b25d-700b26534a04\") " pod="openshift-apiserver/apiserver-76f77b778f-k8qff" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.108675 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.108943 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29328285-mrswx" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.109620 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c56f5eb8-3a12-428c-b347-78369dc19008-trusted-ca-bundle\") pod \"console-f9d7485db-5sffm\" (UID: \"c56f5eb8-3a12-428c-b347-78369dc19008\") " pod="openshift-console/console-f9d7485db-5sffm" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.109708 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/20a5d042-ffe2-4244-b25d-700b26534a04-audit-dir\") pod \"apiserver-76f77b778f-k8qff\" (UID: \"20a5d042-ffe2-4244-b25d-700b26534a04\") " pod="openshift-apiserver/apiserver-76f77b778f-k8qff" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.110751 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/20a5d042-ffe2-4244-b25d-700b26534a04-audit\") pod \"apiserver-76f77b778f-k8qff\" (UID: \"20a5d042-ffe2-4244-b25d-700b26534a04\") " pod="openshift-apiserver/apiserver-76f77b778f-k8qff" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.111998 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-nrgbt"] Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.112546 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/20a5d042-ffe2-4244-b25d-700b26534a04-etcd-client\") pod \"apiserver-76f77b778f-k8qff\" (UID: \"20a5d042-ffe2-4244-b25d-700b26534a04\") " pod="openshift-apiserver/apiserver-76f77b778f-k8qff" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.114987 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/c56f5eb8-3a12-428c-b347-78369dc19008-console-oauth-config\") pod \"console-f9d7485db-5sffm\" (UID: \"c56f5eb8-3a12-428c-b347-78369dc19008\") " pod="openshift-console/console-f9d7485db-5sffm" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.115160 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/3d4ca0c2-4e63-425f-bdb6-89da1a2e04c5-machine-approver-tls\") pod \"machine-approver-56656f9798-hm4ch\" (UID: \"3d4ca0c2-4e63-425f-bdb6-89da1a2e04c5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hm4ch" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.115514 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/20a5d042-ffe2-4244-b25d-700b26534a04-serving-cert\") pod \"apiserver-76f77b778f-k8qff\" (UID: \"20a5d042-ffe2-4244-b25d-700b26534a04\") " pod="openshift-apiserver/apiserver-76f77b778f-k8qff" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.116184 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-nrgbt" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.121875 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.122581 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-nrt9t"] Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.123169 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/c463939d-d2d1-4049-b438-3754a118ff4b-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-p5wqw\" (UID: \"c463939d-d2d1-4049-b438-3754a118ff4b\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-p5wqw" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.125924 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-nrt9t" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.126817 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-chv7x"] Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.128529 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-chv7x" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.136030 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-lv842"] Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.137151 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.138831 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-5sffm"] Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.138956 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-p5wqw"] Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.139027 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-lv842" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.140539 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b1286ed6-4e15-46e0-b4c3-3b7e3cad6057-serving-cert\") pod \"controller-manager-879f6c89f-rvsbj\" (UID: \"b1286ed6-4e15-46e0-b4c3-3b7e3cad6057\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rvsbj" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.141230 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6wnmn"] Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.142837 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-cwtlx"] Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.144695 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-2qmwc"] Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.146631 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-5wx56"] Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.147786 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-ksczt"] Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.149432 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-s4zph"] Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.150705 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-s4zph" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.154120 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-bndqf"] Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.155173 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-6xm7j"] Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.156283 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.156623 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-bndqf" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.157578 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-22htd"] Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.158366 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-6xm7j" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.160741 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wx9l2"] Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.162744 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-ll5jq"] Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.163620 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-5scsb"] Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.165005 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-lxp7v"] Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.167062 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-682zv"] Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.168603 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-8qm2l"] Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.170561 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-gchnw"] Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.171740 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/9e7db3dd-7fae-469a-9ca5-eba1d965cbe8-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-9nms8\" (UID: \"9e7db3dd-7fae-469a-9ca5-eba1d965cbe8\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9nms8" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.171790 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jlsgz\" (UniqueName: \"kubernetes.io/projected/5a98537a-186b-4150-bd41-c2627ec772f9-kube-api-access-jlsgz\") pod \"cluster-samples-operator-665b6dd947-5scsb\" (UID: \"5a98537a-186b-4150-bd41-c2627ec772f9\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-5scsb" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.171818 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7n2wn\" (UniqueName: \"kubernetes.io/projected/811694fe-dcfd-4fe3-876c-964d09321d2f-kube-api-access-7n2wn\") pod \"authentication-operator-69f744f599-22htd\" (UID: \"811694fe-dcfd-4fe3-876c-964d09321d2f\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-22htd" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.171886 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/5a98537a-186b-4150-bd41-c2627ec772f9-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-5scsb\" (UID: \"5a98537a-186b-4150-bd41-c2627ec772f9\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-5scsb" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.171909 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pwqsk\" (UniqueName: \"kubernetes.io/projected/9e7db3dd-7fae-469a-9ca5-eba1d965cbe8-kube-api-access-pwqsk\") pod \"cluster-image-registry-operator-dc59b4c8b-9nms8\" (UID: \"9e7db3dd-7fae-469a-9ca5-eba1d965cbe8\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9nms8" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.171966 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/811694fe-dcfd-4fe3-876c-964d09321d2f-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-22htd\" (UID: \"811694fe-dcfd-4fe3-876c-964d09321d2f\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-22htd" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.172008 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/811694fe-dcfd-4fe3-876c-964d09321d2f-config\") pod \"authentication-operator-69f744f599-22htd\" (UID: \"811694fe-dcfd-4fe3-876c-964d09321d2f\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-22htd" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.172026 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9e7db3dd-7fae-469a-9ca5-eba1d965cbe8-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-9nms8\" (UID: \"9e7db3dd-7fae-469a-9ca5-eba1d965cbe8\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9nms8" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.172046 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/811694fe-dcfd-4fe3-876c-964d09321d2f-serving-cert\") pod \"authentication-operator-69f744f599-22htd\" (UID: \"811694fe-dcfd-4fe3-876c-964d09321d2f\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-22htd" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.172067 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9e7db3dd-7fae-469a-9ca5-eba1d965cbe8-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-9nms8\" (UID: \"9e7db3dd-7fae-469a-9ca5-eba1d965cbe8\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9nms8" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.172088 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/811694fe-dcfd-4fe3-876c-964d09321d2f-service-ca-bundle\") pod \"authentication-operator-69f744f599-22htd\" (UID: \"811694fe-dcfd-4fe3-876c-964d09321d2f\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-22htd" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.172964 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/811694fe-dcfd-4fe3-876c-964d09321d2f-service-ca-bundle\") pod \"authentication-operator-69f744f599-22htd\" (UID: \"811694fe-dcfd-4fe3-876c-964d09321d2f\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-22htd" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.172995 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/811694fe-dcfd-4fe3-876c-964d09321d2f-config\") pod \"authentication-operator-69f744f599-22htd\" (UID: \"811694fe-dcfd-4fe3-876c-964d09321d2f\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-22htd" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.173901 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/811694fe-dcfd-4fe3-876c-964d09321d2f-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-22htd\" (UID: \"811694fe-dcfd-4fe3-876c-964d09321d2f\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-22htd" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.175652 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9e7db3dd-7fae-469a-9ca5-eba1d965cbe8-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-9nms8\" (UID: \"9e7db3dd-7fae-469a-9ca5-eba1d965cbe8\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9nms8" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.176846 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/5a98537a-186b-4150-bd41-c2627ec772f9-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-5scsb\" (UID: \"5a98537a-186b-4150-bd41-c2627ec772f9\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-5scsb" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.177922 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/9e7db3dd-7fae-469a-9ca5-eba1d965cbe8-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-9nms8\" (UID: \"9e7db3dd-7fae-469a-9ca5-eba1d965cbe8\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9nms8" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.178093 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/811694fe-dcfd-4fe3-876c-964d09321d2f-serving-cert\") pod \"authentication-operator-69f744f599-22htd\" (UID: \"811694fe-dcfd-4fe3-876c-964d09321d2f\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-22htd" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.178237 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-5sqj9"] Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.178286 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-pnhf9"] Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.178307 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nrd6b"] Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.178321 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.178528 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-w6krf"] Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.179950 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-q2jq4"] Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.181521 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-hjpzm"] Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.183446 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rl5n7"] Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.185147 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-67bjd"] Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.186205 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-9vndv"] Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.187890 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-s4zph"] Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.188810 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9nms8"] Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.190260 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29328285-mrswx"] Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.191089 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jpbfr"] Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.192739 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-wz2dg"] Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.194387 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-chv7x"] Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.196652 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.197668 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-nrgbt"] Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.199041 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-nrt9t"] Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.201134 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-6xm7j"] Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.203649 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-7g427"] Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.204383 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-7g427" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.204831 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-lv842"] Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.206188 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-7g427"] Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.216832 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.238156 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.256572 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.276826 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.316915 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.337452 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.357248 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.376230 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.398294 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.417284 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.436855 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.457615 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.497653 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.517535 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.537397 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.557597 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.577385 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.597611 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.618722 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.638531 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.666813 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.677127 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.697382 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.716661 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.737378 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.757335 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.778462 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.799172 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.816478 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.847952 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.858029 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.876446 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.897649 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.919053 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.958174 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.959353 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-99kwp\" (UniqueName: \"kubernetes.io/projected/ef8c2ff8-db1e-44ce-b489-daa0832720c3-kube-api-access-99kwp\") pod \"console-operator-58897d9998-9chlx\" (UID: \"ef8c2ff8-db1e-44ce-b489-daa0832720c3\") " pod="openshift-console-operator/console-operator-58897d9998-9chlx" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.978682 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Oct 05 20:56:57 crc kubenswrapper[4754]: I1005 20:56:57.998777 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Oct 05 20:56:58 crc kubenswrapper[4754]: I1005 20:56:58.017985 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Oct 05 20:56:58 crc kubenswrapper[4754]: I1005 20:56:58.064800 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pk76j\" (UniqueName: \"kubernetes.io/projected/c463939d-d2d1-4049-b438-3754a118ff4b-kube-api-access-pk76j\") pod \"machine-api-operator-5694c8668f-p5wqw\" (UID: \"c463939d-d2d1-4049-b438-3754a118ff4b\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-p5wqw" Oct 05 20:56:58 crc kubenswrapper[4754]: I1005 20:56:58.075888 4754 request.go:700] Waited for 1.008932824s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-apiserver-operator/serviceaccounts/openshift-apiserver-operator/token Oct 05 20:56:58 crc kubenswrapper[4754]: I1005 20:56:58.082238 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-p5wqw" Oct 05 20:56:58 crc kubenswrapper[4754]: I1005 20:56:58.089788 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9b4kw\" (UniqueName: \"kubernetes.io/projected/3d4ca0c2-4e63-425f-bdb6-89da1a2e04c5-kube-api-access-9b4kw\") pod \"machine-approver-56656f9798-hm4ch\" (UID: \"3d4ca0c2-4e63-425f-bdb6-89da1a2e04c5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hm4ch" Oct 05 20:56:58 crc kubenswrapper[4754]: I1005 20:56:58.102200 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Oct 05 20:56:58 crc kubenswrapper[4754]: I1005 20:56:58.109137 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kzbj2\" (UniqueName: \"kubernetes.io/projected/afbf61d3-201e-4326-8463-4c660057a865-kube-api-access-kzbj2\") pod \"openshift-apiserver-operator-796bbdcf4f-tcjmf\" (UID: \"afbf61d3-201e-4326-8463-4c660057a865\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tcjmf" Oct 05 20:56:58 crc kubenswrapper[4754]: I1005 20:56:58.117735 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Oct 05 20:56:58 crc kubenswrapper[4754]: I1005 20:56:58.132966 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-9chlx" Oct 05 20:56:58 crc kubenswrapper[4754]: I1005 20:56:58.138296 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Oct 05 20:56:58 crc kubenswrapper[4754]: I1005 20:56:58.158011 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Oct 05 20:56:58 crc kubenswrapper[4754]: I1005 20:56:58.177376 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Oct 05 20:56:58 crc kubenswrapper[4754]: I1005 20:56:58.265425 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Oct 05 20:56:58 crc kubenswrapper[4754]: I1005 20:56:58.269429 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Oct 05 20:56:58 crc kubenswrapper[4754]: I1005 20:56:58.269912 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Oct 05 20:56:58 crc kubenswrapper[4754]: I1005 20:56:58.270388 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Oct 05 20:56:58 crc kubenswrapper[4754]: I1005 20:56:58.277942 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Oct 05 20:56:58 crc kubenswrapper[4754]: I1005 20:56:58.297391 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Oct 05 20:56:58 crc kubenswrapper[4754]: I1005 20:56:58.317242 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Oct 05 20:56:58 crc kubenswrapper[4754]: I1005 20:56:58.338592 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Oct 05 20:56:58 crc kubenswrapper[4754]: I1005 20:56:58.361183 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hm4ch" Oct 05 20:56:58 crc kubenswrapper[4754]: I1005 20:56:58.382456 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m66bm\" (UniqueName: \"kubernetes.io/projected/c56f5eb8-3a12-428c-b347-78369dc19008-kube-api-access-m66bm\") pod \"console-f9d7485db-5sffm\" (UID: \"c56f5eb8-3a12-428c-b347-78369dc19008\") " pod="openshift-console/console-f9d7485db-5sffm" Oct 05 20:56:58 crc kubenswrapper[4754]: I1005 20:56:58.394307 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tcjmf" Oct 05 20:56:58 crc kubenswrapper[4754]: I1005 20:56:58.404064 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-p5wqw"] Oct 05 20:56:58 crc kubenswrapper[4754]: I1005 20:56:58.407171 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nnwxh\" (UniqueName: \"kubernetes.io/projected/b1286ed6-4e15-46e0-b4c3-3b7e3cad6057-kube-api-access-nnwxh\") pod \"controller-manager-879f6c89f-rvsbj\" (UID: \"b1286ed6-4e15-46e0-b4c3-3b7e3cad6057\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rvsbj" Oct 05 20:56:58 crc kubenswrapper[4754]: I1005 20:56:58.416956 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Oct 05 20:56:58 crc kubenswrapper[4754]: I1005 20:56:58.419614 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cpf6q\" (UniqueName: \"kubernetes.io/projected/20a5d042-ffe2-4244-b25d-700b26534a04-kube-api-access-cpf6q\") pod \"apiserver-76f77b778f-k8qff\" (UID: \"20a5d042-ffe2-4244-b25d-700b26534a04\") " pod="openshift-apiserver/apiserver-76f77b778f-k8qff" Oct 05 20:56:58 crc kubenswrapper[4754]: W1005 20:56:58.425320 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc463939d_d2d1_4049_b438_3754a118ff4b.slice/crio-4515eeb4a2995769bfed55b46bf67807214d8d8ba0fe748651a374005f68a62e WatchSource:0}: Error finding container 4515eeb4a2995769bfed55b46bf67807214d8d8ba0fe748651a374005f68a62e: Status 404 returned error can't find the container with id 4515eeb4a2995769bfed55b46bf67807214d8d8ba0fe748651a374005f68a62e Oct 05 20:56:58 crc kubenswrapper[4754]: I1005 20:56:58.438041 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Oct 05 20:56:58 crc kubenswrapper[4754]: I1005 20:56:58.451463 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-9chlx"] Oct 05 20:56:58 crc kubenswrapper[4754]: I1005 20:56:58.457978 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Oct 05 20:56:58 crc kubenswrapper[4754]: I1005 20:56:58.466358 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-5sffm" Oct 05 20:56:58 crc kubenswrapper[4754]: I1005 20:56:58.476631 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Oct 05 20:56:58 crc kubenswrapper[4754]: I1005 20:56:58.497033 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Oct 05 20:56:58 crc kubenswrapper[4754]: I1005 20:56:58.517363 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 05 20:56:58 crc kubenswrapper[4754]: I1005 20:56:58.539552 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 05 20:56:58 crc kubenswrapper[4754]: I1005 20:56:58.558212 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Oct 05 20:56:58 crc kubenswrapper[4754]: I1005 20:56:58.577780 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Oct 05 20:56:58 crc kubenswrapper[4754]: I1005 20:56:58.597110 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Oct 05 20:56:58 crc kubenswrapper[4754]: I1005 20:56:58.618100 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Oct 05 20:56:58 crc kubenswrapper[4754]: I1005 20:56:58.636871 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Oct 05 20:56:58 crc kubenswrapper[4754]: I1005 20:56:58.639315 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-k8qff" Oct 05 20:56:58 crc kubenswrapper[4754]: I1005 20:56:58.649580 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-rvsbj" Oct 05 20:56:58 crc kubenswrapper[4754]: I1005 20:56:58.660432 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Oct 05 20:56:58 crc kubenswrapper[4754]: I1005 20:56:58.681634 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Oct 05 20:56:58 crc kubenswrapper[4754]: I1005 20:56:58.694710 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tcjmf"] Oct 05 20:56:58 crc kubenswrapper[4754]: I1005 20:56:58.697307 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Oct 05 20:56:58 crc kubenswrapper[4754]: W1005 20:56:58.712906 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podafbf61d3_201e_4326_8463_4c660057a865.slice/crio-fb8ddde177a6e2a0389675e626119bd138c7835e02623fd3767bff05d73840ba WatchSource:0}: Error finding container fb8ddde177a6e2a0389675e626119bd138c7835e02623fd3767bff05d73840ba: Status 404 returned error can't find the container with id fb8ddde177a6e2a0389675e626119bd138c7835e02623fd3767bff05d73840ba Oct 05 20:56:58 crc kubenswrapper[4754]: I1005 20:56:58.718921 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Oct 05 20:56:58 crc kubenswrapper[4754]: I1005 20:56:58.737750 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Oct 05 20:56:58 crc kubenswrapper[4754]: I1005 20:56:58.759128 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Oct 05 20:56:58 crc kubenswrapper[4754]: I1005 20:56:58.777089 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Oct 05 20:56:58 crc kubenswrapper[4754]: I1005 20:56:58.784911 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-9chlx" event={"ID":"ef8c2ff8-db1e-44ce-b489-daa0832720c3","Type":"ContainerStarted","Data":"eee8e408b44d7c0202f2cc0ba2fe01b9d2029622b0992b5a51d4102697e1bd90"} Oct 05 20:56:58 crc kubenswrapper[4754]: I1005 20:56:58.784969 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-9chlx" event={"ID":"ef8c2ff8-db1e-44ce-b489-daa0832720c3","Type":"ContainerStarted","Data":"11d0075b40990b92762c128ca7d4afdab5f56b026433ea7977ea514f29475cca"} Oct 05 20:56:58 crc kubenswrapper[4754]: I1005 20:56:58.785605 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-9chlx" Oct 05 20:56:58 crc kubenswrapper[4754]: I1005 20:56:58.791020 4754 patch_prober.go:28] interesting pod/console-operator-58897d9998-9chlx container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.8:8443/readyz\": dial tcp 10.217.0.8:8443: connect: connection refused" start-of-body= Oct 05 20:56:58 crc kubenswrapper[4754]: I1005 20:56:58.791068 4754 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-9chlx" podUID="ef8c2ff8-db1e-44ce-b489-daa0832720c3" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.8:8443/readyz\": dial tcp 10.217.0.8:8443: connect: connection refused" Oct 05 20:56:58 crc kubenswrapper[4754]: I1005 20:56:58.791339 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-p5wqw" event={"ID":"c463939d-d2d1-4049-b438-3754a118ff4b","Type":"ContainerStarted","Data":"280caaaefcf797ad6cc7da6bee67fd8089675ba9bd8800ebb3761d3ebd2db88a"} Oct 05 20:56:58 crc kubenswrapper[4754]: I1005 20:56:58.791394 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-p5wqw" event={"ID":"c463939d-d2d1-4049-b438-3754a118ff4b","Type":"ContainerStarted","Data":"e9e0f440529b469c7e7c81ec01d54223647c0f922135e145b22752dd03171612"} Oct 05 20:56:58 crc kubenswrapper[4754]: I1005 20:56:58.791407 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-p5wqw" event={"ID":"c463939d-d2d1-4049-b438-3754a118ff4b","Type":"ContainerStarted","Data":"4515eeb4a2995769bfed55b46bf67807214d8d8ba0fe748651a374005f68a62e"} Oct 05 20:56:58 crc kubenswrapper[4754]: I1005 20:56:58.794442 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tcjmf" event={"ID":"afbf61d3-201e-4326-8463-4c660057a865","Type":"ContainerStarted","Data":"fb8ddde177a6e2a0389675e626119bd138c7835e02623fd3767bff05d73840ba"} Oct 05 20:56:58 crc kubenswrapper[4754]: I1005 20:56:58.795964 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Oct 05 20:56:58 crc kubenswrapper[4754]: I1005 20:56:58.796358 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hm4ch" event={"ID":"3d4ca0c2-4e63-425f-bdb6-89da1a2e04c5","Type":"ContainerStarted","Data":"5626739f6f01ffda50126b6f9f9af7e549e57915e32b9a30c6a10c4d75ae761d"} Oct 05 20:56:58 crc kubenswrapper[4754]: I1005 20:56:58.796445 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hm4ch" event={"ID":"3d4ca0c2-4e63-425f-bdb6-89da1a2e04c5","Type":"ContainerStarted","Data":"37df27ee1b6428fe53e0a3ae9080c1674b79f216c91bd791acedcb5216fd3924"} Oct 05 20:56:58 crc kubenswrapper[4754]: I1005 20:56:58.807887 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-5sffm"] Oct 05 20:56:58 crc kubenswrapper[4754]: I1005 20:56:58.821755 4754 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Oct 05 20:56:58 crc kubenswrapper[4754]: I1005 20:56:58.837228 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Oct 05 20:56:58 crc kubenswrapper[4754]: W1005 20:56:58.857889 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc56f5eb8_3a12_428c_b347_78369dc19008.slice/crio-1b5f2b6c650425bb321b1679f0ac14848e2a2b5706e1443dbdeda08a6b35615f WatchSource:0}: Error finding container 1b5f2b6c650425bb321b1679f0ac14848e2a2b5706e1443dbdeda08a6b35615f: Status 404 returned error can't find the container with id 1b5f2b6c650425bb321b1679f0ac14848e2a2b5706e1443dbdeda08a6b35615f Oct 05 20:56:58 crc kubenswrapper[4754]: I1005 20:56:58.860423 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Oct 05 20:56:58 crc kubenswrapper[4754]: I1005 20:56:58.877092 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Oct 05 20:56:58 crc kubenswrapper[4754]: I1005 20:56:58.897676 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-k8qff"] Oct 05 20:56:58 crc kubenswrapper[4754]: I1005 20:56:58.898554 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Oct 05 20:56:58 crc kubenswrapper[4754]: I1005 20:56:58.917003 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Oct 05 20:56:58 crc kubenswrapper[4754]: I1005 20:56:58.942432 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Oct 05 20:56:58 crc kubenswrapper[4754]: I1005 20:56:58.956176 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Oct 05 20:56:58 crc kubenswrapper[4754]: I1005 20:56:58.963090 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-rvsbj"] Oct 05 20:56:58 crc kubenswrapper[4754]: W1005 20:56:58.977976 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb1286ed6_4e15_46e0_b4c3_3b7e3cad6057.slice/crio-2fd8d2336ece00bcebac4481b957ad7bb2adf173e4f8f96087c3c4d72ea26934 WatchSource:0}: Error finding container 2fd8d2336ece00bcebac4481b957ad7bb2adf173e4f8f96087c3c4d72ea26934: Status 404 returned error can't find the container with id 2fd8d2336ece00bcebac4481b957ad7bb2adf173e4f8f96087c3c4d72ea26934 Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.000242 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7n2wn\" (UniqueName: \"kubernetes.io/projected/811694fe-dcfd-4fe3-876c-964d09321d2f-kube-api-access-7n2wn\") pod \"authentication-operator-69f744f599-22htd\" (UID: \"811694fe-dcfd-4fe3-876c-964d09321d2f\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-22htd" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.014427 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pwqsk\" (UniqueName: \"kubernetes.io/projected/9e7db3dd-7fae-469a-9ca5-eba1d965cbe8-kube-api-access-pwqsk\") pod \"cluster-image-registry-operator-dc59b4c8b-9nms8\" (UID: \"9e7db3dd-7fae-469a-9ca5-eba1d965cbe8\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9nms8" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.033655 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9e7db3dd-7fae-469a-9ca5-eba1d965cbe8-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-9nms8\" (UID: \"9e7db3dd-7fae-469a-9ca5-eba1d965cbe8\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9nms8" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.058094 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.058669 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jlsgz\" (UniqueName: \"kubernetes.io/projected/5a98537a-186b-4150-bd41-c2627ec772f9-kube-api-access-jlsgz\") pod \"cluster-samples-operator-665b6dd947-5scsb\" (UID: \"5a98537a-186b-4150-bd41-c2627ec772f9\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-5scsb" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.076721 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.094935 4754 request.go:700] Waited for 1.890226392s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-ingress-canary/secrets?fieldSelector=metadata.name%3Dcanary-serving-cert&limit=500&resourceVersion=0 Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.096095 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.117646 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.212031 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/6c0dfaff-08a7-4152-8624-1973e4a4d148-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-gchnw\" (UID: \"6c0dfaff-08a7-4152-8624-1973e4a4d148\") " pod="openshift-authentication/oauth-openshift-558db77b4-gchnw" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.212078 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6c0dfaff-08a7-4152-8624-1973e4a4d148-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-gchnw\" (UID: \"6c0dfaff-08a7-4152-8624-1973e4a4d148\") " pod="openshift-authentication/oauth-openshift-558db77b4-gchnw" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.212120 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5sqj9\" (UID: \"c64d57c5-4bd4-48b5-85fc-9d613aea29c4\") " pod="openshift-image-registry/image-registry-697d97f7c8-5sqj9" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.212150 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hmrqj\" (UniqueName: \"kubernetes.io/projected/3408670c-55e3-4bf0-9589-fef6200c3490-kube-api-access-hmrqj\") pod \"openshift-controller-manager-operator-756b6f6bc6-wx9l2\" (UID: \"3408670c-55e3-4bf0-9589-fef6200c3490\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wx9l2" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.212175 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/6c0dfaff-08a7-4152-8624-1973e4a4d148-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-gchnw\" (UID: \"6c0dfaff-08a7-4152-8624-1973e4a4d148\") " pod="openshift-authentication/oauth-openshift-558db77b4-gchnw" Oct 05 20:56:59 crc kubenswrapper[4754]: E1005 20:56:59.212585 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 20:56:59.712553182 +0000 UTC m=+143.616671882 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5sqj9" (UID: "c64d57c5-4bd4-48b5-85fc-9d613aea29c4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.212654 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/6c0dfaff-08a7-4152-8624-1973e4a4d148-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-gchnw\" (UID: \"6c0dfaff-08a7-4152-8624-1973e4a4d148\") " pod="openshift-authentication/oauth-openshift-558db77b4-gchnw" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.212680 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/c64d57c5-4bd4-48b5-85fc-9d613aea29c4-registry-tls\") pod \"image-registry-697d97f7c8-5sqj9\" (UID: \"c64d57c5-4bd4-48b5-85fc-9d613aea29c4\") " pod="openshift-image-registry/image-registry-697d97f7c8-5sqj9" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.212700 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/6c0dfaff-08a7-4152-8624-1973e4a4d148-audit-policies\") pod \"oauth-openshift-558db77b4-gchnw\" (UID: \"6c0dfaff-08a7-4152-8624-1973e4a4d148\") " pod="openshift-authentication/oauth-openshift-558db77b4-gchnw" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.212738 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/6c0dfaff-08a7-4152-8624-1973e4a4d148-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-gchnw\" (UID: \"6c0dfaff-08a7-4152-8624-1973e4a4d148\") " pod="openshift-authentication/oauth-openshift-558db77b4-gchnw" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.212761 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/6c0dfaff-08a7-4152-8624-1973e4a4d148-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-gchnw\" (UID: \"6c0dfaff-08a7-4152-8624-1973e4a4d148\") " pod="openshift-authentication/oauth-openshift-558db77b4-gchnw" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.212793 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wtph5\" (UniqueName: \"kubernetes.io/projected/bc8bf6c1-363b-4e24-830c-898e4f7c6325-kube-api-access-wtph5\") pod \"dns-operator-744455d44c-2qmwc\" (UID: \"bc8bf6c1-363b-4e24-830c-898e4f7c6325\") " pod="openshift-dns-operator/dns-operator-744455d44c-2qmwc" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.212815 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/d3c66e2a-8404-498c-b8a9-2bd2e6fd318b-audit-dir\") pod \"apiserver-7bbb656c7d-lxp7v\" (UID: \"d3c66e2a-8404-498c-b8a9-2bd2e6fd318b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lxp7v" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.212834 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-btbzr\" (UniqueName: \"kubernetes.io/projected/d3c66e2a-8404-498c-b8a9-2bd2e6fd318b-kube-api-access-btbzr\") pod \"apiserver-7bbb656c7d-lxp7v\" (UID: \"d3c66e2a-8404-498c-b8a9-2bd2e6fd318b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lxp7v" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.212851 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/6c0dfaff-08a7-4152-8624-1973e4a4d148-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-gchnw\" (UID: \"6c0dfaff-08a7-4152-8624-1973e4a4d148\") " pod="openshift-authentication/oauth-openshift-558db77b4-gchnw" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.212919 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/6c0dfaff-08a7-4152-8624-1973e4a4d148-audit-dir\") pod \"oauth-openshift-558db77b4-gchnw\" (UID: \"6c0dfaff-08a7-4152-8624-1973e4a4d148\") " pod="openshift-authentication/oauth-openshift-558db77b4-gchnw" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.212946 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/bc8bf6c1-363b-4e24-830c-898e4f7c6325-metrics-tls\") pod \"dns-operator-744455d44c-2qmwc\" (UID: \"bc8bf6c1-363b-4e24-830c-898e4f7c6325\") " pod="openshift-dns-operator/dns-operator-744455d44c-2qmwc" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.213817 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/dbe65326-0856-4b33-8894-4c9266bc1a90-etcd-ca\") pod \"etcd-operator-b45778765-q2jq4\" (UID: \"dbe65326-0856-4b33-8894-4c9266bc1a90\") " pod="openshift-etcd-operator/etcd-operator-b45778765-q2jq4" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.213848 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/dbe65326-0856-4b33-8894-4c9266bc1a90-etcd-client\") pod \"etcd-operator-b45778765-q2jq4\" (UID: \"dbe65326-0856-4b33-8894-4c9266bc1a90\") " pod="openshift-etcd-operator/etcd-operator-b45778765-q2jq4" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.213883 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mqjlq\" (UniqueName: \"kubernetes.io/projected/6c0dfaff-08a7-4152-8624-1973e4a4d148-kube-api-access-mqjlq\") pod \"oauth-openshift-558db77b4-gchnw\" (UID: \"6c0dfaff-08a7-4152-8624-1973e4a4d148\") " pod="openshift-authentication/oauth-openshift-558db77b4-gchnw" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.213945 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c64d57c5-4bd4-48b5-85fc-9d613aea29c4-trusted-ca\") pod \"image-registry-697d97f7c8-5sqj9\" (UID: \"c64d57c5-4bd4-48b5-85fc-9d613aea29c4\") " pod="openshift-image-registry/image-registry-697d97f7c8-5sqj9" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.213991 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/6c0dfaff-08a7-4152-8624-1973e4a4d148-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-gchnw\" (UID: \"6c0dfaff-08a7-4152-8624-1973e4a4d148\") " pod="openshift-authentication/oauth-openshift-558db77b4-gchnw" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.214066 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/d3c66e2a-8404-498c-b8a9-2bd2e6fd318b-audit-policies\") pod \"apiserver-7bbb656c7d-lxp7v\" (UID: \"d3c66e2a-8404-498c-b8a9-2bd2e6fd318b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lxp7v" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.214123 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/d3c66e2a-8404-498c-b8a9-2bd2e6fd318b-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-lxp7v\" (UID: \"d3c66e2a-8404-498c-b8a9-2bd2e6fd318b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lxp7v" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.214170 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3408670c-55e3-4bf0-9589-fef6200c3490-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-wx9l2\" (UID: \"3408670c-55e3-4bf0-9589-fef6200c3490\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wx9l2" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.214238 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7qtr7\" (UniqueName: \"kubernetes.io/projected/c64d57c5-4bd4-48b5-85fc-9d613aea29c4-kube-api-access-7qtr7\") pod \"image-registry-697d97f7c8-5sqj9\" (UID: \"c64d57c5-4bd4-48b5-85fc-9d613aea29c4\") " pod="openshift-image-registry/image-registry-697d97f7c8-5sqj9" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.214306 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d3c66e2a-8404-498c-b8a9-2bd2e6fd318b-serving-cert\") pod \"apiserver-7bbb656c7d-lxp7v\" (UID: \"d3c66e2a-8404-498c-b8a9-2bd2e6fd318b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lxp7v" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.214351 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/b157fa21-ae67-4c98-806c-93854b072965-available-featuregates\") pod \"openshift-config-operator-7777fb866f-cwtlx\" (UID: \"b157fa21-ae67-4c98-806c-93854b072965\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-cwtlx" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.214378 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dbe65326-0856-4b33-8894-4c9266bc1a90-config\") pod \"etcd-operator-b45778765-q2jq4\" (UID: \"dbe65326-0856-4b33-8894-4c9266bc1a90\") " pod="openshift-etcd-operator/etcd-operator-b45778765-q2jq4" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.214417 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3408670c-55e3-4bf0-9589-fef6200c3490-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-wx9l2\" (UID: \"3408670c-55e3-4bf0-9589-fef6200c3490\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wx9l2" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.214460 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/6c0dfaff-08a7-4152-8624-1973e4a4d148-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-gchnw\" (UID: \"6c0dfaff-08a7-4152-8624-1973e4a4d148\") " pod="openshift-authentication/oauth-openshift-558db77b4-gchnw" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.214508 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/c64d57c5-4bd4-48b5-85fc-9d613aea29c4-ca-trust-extracted\") pod \"image-registry-697d97f7c8-5sqj9\" (UID: \"c64d57c5-4bd4-48b5-85fc-9d613aea29c4\") " pod="openshift-image-registry/image-registry-697d97f7c8-5sqj9" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.214536 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1e16d79e-eb1e-4b8b-8f0a-acb370163d17-client-ca\") pod \"route-controller-manager-6576b87f9c-pnhf9\" (UID: \"1e16d79e-eb1e-4b8b-8f0a-acb370163d17\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-pnhf9" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.214561 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1e16d79e-eb1e-4b8b-8f0a-acb370163d17-serving-cert\") pod \"route-controller-manager-6576b87f9c-pnhf9\" (UID: \"1e16d79e-eb1e-4b8b-8f0a-acb370163d17\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-pnhf9" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.214584 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d3c66e2a-8404-498c-b8a9-2bd2e6fd318b-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-lxp7v\" (UID: \"d3c66e2a-8404-498c-b8a9-2bd2e6fd318b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lxp7v" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.214602 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/d3c66e2a-8404-498c-b8a9-2bd2e6fd318b-etcd-client\") pod \"apiserver-7bbb656c7d-lxp7v\" (UID: \"d3c66e2a-8404-498c-b8a9-2bd2e6fd318b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lxp7v" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.214627 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/d3c66e2a-8404-498c-b8a9-2bd2e6fd318b-encryption-config\") pod \"apiserver-7bbb656c7d-lxp7v\" (UID: \"d3c66e2a-8404-498c-b8a9-2bd2e6fd318b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lxp7v" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.214649 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b157fa21-ae67-4c98-806c-93854b072965-serving-cert\") pod \"openshift-config-operator-7777fb866f-cwtlx\" (UID: \"b157fa21-ae67-4c98-806c-93854b072965\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-cwtlx" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.214666 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dbe65326-0856-4b33-8894-4c9266bc1a90-serving-cert\") pod \"etcd-operator-b45778765-q2jq4\" (UID: \"dbe65326-0856-4b33-8894-4c9266bc1a90\") " pod="openshift-etcd-operator/etcd-operator-b45778765-q2jq4" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.214684 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1e16d79e-eb1e-4b8b-8f0a-acb370163d17-config\") pod \"route-controller-manager-6576b87f9c-pnhf9\" (UID: \"1e16d79e-eb1e-4b8b-8f0a-acb370163d17\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-pnhf9" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.214727 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/c64d57c5-4bd4-48b5-85fc-9d613aea29c4-registry-certificates\") pod \"image-registry-697d97f7c8-5sqj9\" (UID: \"c64d57c5-4bd4-48b5-85fc-9d613aea29c4\") " pod="openshift-image-registry/image-registry-697d97f7c8-5sqj9" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.214747 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9wb4b\" (UniqueName: \"kubernetes.io/projected/1e16d79e-eb1e-4b8b-8f0a-acb370163d17-kube-api-access-9wb4b\") pod \"route-controller-manager-6576b87f9c-pnhf9\" (UID: \"1e16d79e-eb1e-4b8b-8f0a-acb370163d17\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-pnhf9" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.214768 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/6c0dfaff-08a7-4152-8624-1973e4a4d148-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-gchnw\" (UID: \"6c0dfaff-08a7-4152-8624-1973e4a4d148\") " pod="openshift-authentication/oauth-openshift-558db77b4-gchnw" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.214800 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/6c0dfaff-08a7-4152-8624-1973e4a4d148-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-gchnw\" (UID: \"6c0dfaff-08a7-4152-8624-1973e4a4d148\") " pod="openshift-authentication/oauth-openshift-558db77b4-gchnw" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.214849 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/c64d57c5-4bd4-48b5-85fc-9d613aea29c4-installation-pull-secrets\") pod \"image-registry-697d97f7c8-5sqj9\" (UID: \"c64d57c5-4bd4-48b5-85fc-9d613aea29c4\") " pod="openshift-image-registry/image-registry-697d97f7c8-5sqj9" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.214866 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dw4mx\" (UniqueName: \"kubernetes.io/projected/3934dd94-b9fa-4079-a268-60a9470dab41-kube-api-access-dw4mx\") pod \"downloads-7954f5f757-ksczt\" (UID: \"3934dd94-b9fa-4079-a268-60a9470dab41\") " pod="openshift-console/downloads-7954f5f757-ksczt" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.214886 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c64d57c5-4bd4-48b5-85fc-9d613aea29c4-bound-sa-token\") pod \"image-registry-697d97f7c8-5sqj9\" (UID: \"c64d57c5-4bd4-48b5-85fc-9d613aea29c4\") " pod="openshift-image-registry/image-registry-697d97f7c8-5sqj9" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.214904 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wtxlp\" (UniqueName: \"kubernetes.io/projected/b157fa21-ae67-4c98-806c-93854b072965-kube-api-access-wtxlp\") pod \"openshift-config-operator-7777fb866f-cwtlx\" (UID: \"b157fa21-ae67-4c98-806c-93854b072965\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-cwtlx" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.214921 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/dbe65326-0856-4b33-8894-4c9266bc1a90-etcd-service-ca\") pod \"etcd-operator-b45778765-q2jq4\" (UID: \"dbe65326-0856-4b33-8894-4c9266bc1a90\") " pod="openshift-etcd-operator/etcd-operator-b45778765-q2jq4" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.214952 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7pfxf\" (UniqueName: \"kubernetes.io/projected/dbe65326-0856-4b33-8894-4c9266bc1a90-kube-api-access-7pfxf\") pod \"etcd-operator-b45778765-q2jq4\" (UID: \"dbe65326-0856-4b33-8894-4c9266bc1a90\") " pod="openshift-etcd-operator/etcd-operator-b45778765-q2jq4" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.236311 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-5scsb" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.248729 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9nms8" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.264755 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-22htd" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.316089 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 20:56:59 crc kubenswrapper[4754]: E1005 20:56:59.316306 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 20:56:59.816261209 +0000 UTC m=+143.720379919 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.316382 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/5a5b1574-7f76-4bcf-8c24-9ff7773e86f2-plugins-dir\") pod \"csi-hostpathplugin-s4zph\" (UID: \"5a5b1574-7f76-4bcf-8c24-9ff7773e86f2\") " pod="hostpath-provisioner/csi-hostpathplugin-s4zph" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.316440 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/091efc31-55eb-4026-bf38-626e09a13253-webhook-cert\") pod \"packageserver-d55dfcdfc-nrd6b\" (UID: \"091efc31-55eb-4026-bf38-626e09a13253\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nrd6b" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.316469 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gxjxv\" (UniqueName: \"kubernetes.io/projected/0b4c539c-22cf-4337-9959-afcc73e64370-kube-api-access-gxjxv\") pod \"package-server-manager-789f6589d5-nrt9t\" (UID: \"0b4c539c-22cf-4337-9959-afcc73e64370\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-nrt9t" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.316537 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/b157fa21-ae67-4c98-806c-93854b072965-available-featuregates\") pod \"openshift-config-operator-7777fb866f-cwtlx\" (UID: \"b157fa21-ae67-4c98-806c-93854b072965\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-cwtlx" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.316572 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dbe65326-0856-4b33-8894-4c9266bc1a90-config\") pod \"etcd-operator-b45778765-q2jq4\" (UID: \"dbe65326-0856-4b33-8894-4c9266bc1a90\") " pod="openshift-etcd-operator/etcd-operator-b45778765-q2jq4" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.316600 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d3c66e2a-8404-498c-b8a9-2bd2e6fd318b-serving-cert\") pod \"apiserver-7bbb656c7d-lxp7v\" (UID: \"d3c66e2a-8404-498c-b8a9-2bd2e6fd318b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lxp7v" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.316625 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/c1c75259-4cf0-46b9-aea0-2c2ad2a6928c-node-bootstrap-token\") pod \"machine-config-server-bndqf\" (UID: \"c1c75259-4cf0-46b9-aea0-2c2ad2a6928c\") " pod="openshift-machine-config-operator/machine-config-server-bndqf" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.316648 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2e5d7ee3-ea83-4445-a74d-9efa3864b3bb-config\") pod \"service-ca-operator-777779d784-nrgbt\" (UID: \"2e5d7ee3-ea83-4445-a74d-9efa3864b3bb\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-nrgbt" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.316696 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/c1c4fc78-c305-4dba-b4d5-9f05d58a8ef6-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-ll5jq\" (UID: \"c1c4fc78-c305-4dba-b4d5-9f05d58a8ef6\") " pod="openshift-marketplace/marketplace-operator-79b997595-ll5jq" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.316764 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3408670c-55e3-4bf0-9589-fef6200c3490-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-wx9l2\" (UID: \"3408670c-55e3-4bf0-9589-fef6200c3490\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wx9l2" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.316796 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/c64d57c5-4bd4-48b5-85fc-9d613aea29c4-ca-trust-extracted\") pod \"image-registry-697d97f7c8-5sqj9\" (UID: \"c64d57c5-4bd4-48b5-85fc-9d613aea29c4\") " pod="openshift-image-registry/image-registry-697d97f7c8-5sqj9" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.316820 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zqm88\" (UniqueName: \"kubernetes.io/projected/cce92b6a-69b0-49d5-bb66-36349ca96841-kube-api-access-zqm88\") pod \"machine-config-controller-84d6567774-wz2dg\" (UID: \"cce92b6a-69b0-49d5-bb66-36349ca96841\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-wz2dg" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.316851 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mv4pz\" (UniqueName: \"kubernetes.io/projected/60ff4c5c-9e45-47d3-b45d-a365e988342e-kube-api-access-mv4pz\") pod \"collect-profiles-29328285-mrswx\" (UID: \"60ff4c5c-9e45-47d3-b45d-a365e988342e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29328285-mrswx" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.316893 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b157fa21-ae67-4c98-806c-93854b072965-serving-cert\") pod \"openshift-config-operator-7777fb866f-cwtlx\" (UID: \"b157fa21-ae67-4c98-806c-93854b072965\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-cwtlx" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.316920 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/d3c66e2a-8404-498c-b8a9-2bd2e6fd318b-etcd-client\") pod \"apiserver-7bbb656c7d-lxp7v\" (UID: \"d3c66e2a-8404-498c-b8a9-2bd2e6fd318b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lxp7v" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.316933 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/b157fa21-ae67-4c98-806c-93854b072965-available-featuregates\") pod \"openshift-config-operator-7777fb866f-cwtlx\" (UID: \"b157fa21-ae67-4c98-806c-93854b072965\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-cwtlx" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.316953 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/c64d57c5-4bd4-48b5-85fc-9d613aea29c4-registry-certificates\") pod \"image-registry-697d97f7c8-5sqj9\" (UID: \"c64d57c5-4bd4-48b5-85fc-9d613aea29c4\") " pod="openshift-image-registry/image-registry-697d97f7c8-5sqj9" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.317010 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9wb4b\" (UniqueName: \"kubernetes.io/projected/1e16d79e-eb1e-4b8b-8f0a-acb370163d17-kube-api-access-9wb4b\") pod \"route-controller-manager-6576b87f9c-pnhf9\" (UID: \"1e16d79e-eb1e-4b8b-8f0a-acb370163d17\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-pnhf9" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.317037 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e542cd22-7a8b-4a4d-8205-d3dc60b2b8c6-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-w6krf\" (UID: \"e542cd22-7a8b-4a4d-8205-d3dc60b2b8c6\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-w6krf" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.317084 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-czgsx\" (UniqueName: \"kubernetes.io/projected/ebf750fb-f47c-420b-931d-3ac0a23e9e7b-kube-api-access-czgsx\") pod \"router-default-5444994796-ll9lb\" (UID: \"ebf750fb-f47c-420b-931d-3ac0a23e9e7b\") " pod="openshift-ingress/router-default-5444994796-ll9lb" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.317113 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/c64d57c5-4bd4-48b5-85fc-9d613aea29c4-installation-pull-secrets\") pod \"image-registry-697d97f7c8-5sqj9\" (UID: \"c64d57c5-4bd4-48b5-85fc-9d613aea29c4\") " pod="openshift-image-registry/image-registry-697d97f7c8-5sqj9" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.317137 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/0b4c539c-22cf-4337-9959-afcc73e64370-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-nrt9t\" (UID: \"0b4c539c-22cf-4337-9959-afcc73e64370\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-nrt9t" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.317164 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c64d57c5-4bd4-48b5-85fc-9d613aea29c4-bound-sa-token\") pod \"image-registry-697d97f7c8-5sqj9\" (UID: \"c64d57c5-4bd4-48b5-85fc-9d613aea29c4\") " pod="openshift-image-registry/image-registry-697d97f7c8-5sqj9" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.317195 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/dbe65326-0856-4b33-8894-4c9266bc1a90-etcd-service-ca\") pod \"etcd-operator-b45778765-q2jq4\" (UID: \"dbe65326-0856-4b33-8894-4c9266bc1a90\") " pod="openshift-etcd-operator/etcd-operator-b45778765-q2jq4" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.317226 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/d1b2b973-8b95-4f6a-83f1-468f84fb006d-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-5wx56\" (UID: \"d1b2b973-8b95-4f6a-83f1-468f84fb006d\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-5wx56" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.317250 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b96341ba-da55-442f-851e-5666717a1d5a-bound-sa-token\") pod \"ingress-operator-5b745b69d9-9vndv\" (UID: \"b96341ba-da55-442f-851e-5666717a1d5a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-9vndv" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.317291 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5sqj9\" (UID: \"c64d57c5-4bd4-48b5-85fc-9d613aea29c4\") " pod="openshift-image-registry/image-registry-697d97f7c8-5sqj9" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.317319 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6c0dfaff-08a7-4152-8624-1973e4a4d148-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-gchnw\" (UID: \"6c0dfaff-08a7-4152-8624-1973e4a4d148\") " pod="openshift-authentication/oauth-openshift-558db77b4-gchnw" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.317350 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hmrqj\" (UniqueName: \"kubernetes.io/projected/3408670c-55e3-4bf0-9589-fef6200c3490-kube-api-access-hmrqj\") pod \"openshift-controller-manager-operator-756b6f6bc6-wx9l2\" (UID: \"3408670c-55e3-4bf0-9589-fef6200c3490\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wx9l2" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.317374 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/6c0dfaff-08a7-4152-8624-1973e4a4d148-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-gchnw\" (UID: \"6c0dfaff-08a7-4152-8624-1973e4a4d148\") " pod="openshift-authentication/oauth-openshift-558db77b4-gchnw" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.317957 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/4da215ae-75ec-4ab1-81b0-39966d449b5f-images\") pod \"machine-config-operator-74547568cd-hjpzm\" (UID: \"4da215ae-75ec-4ab1-81b0-39966d449b5f\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-hjpzm" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.317994 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dbe65326-0856-4b33-8894-4c9266bc1a90-config\") pod \"etcd-operator-b45778765-q2jq4\" (UID: \"dbe65326-0856-4b33-8894-4c9266bc1a90\") " pod="openshift-etcd-operator/etcd-operator-b45778765-q2jq4" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.318352 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/6c0dfaff-08a7-4152-8624-1973e4a4d148-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-gchnw\" (UID: \"6c0dfaff-08a7-4152-8624-1973e4a4d148\") " pod="openshift-authentication/oauth-openshift-558db77b4-gchnw" Oct 05 20:56:59 crc kubenswrapper[4754]: E1005 20:56:59.318867 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 20:56:59.818851548 +0000 UTC m=+143.722970258 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5sqj9" (UID: "c64d57c5-4bd4-48b5-85fc-9d613aea29c4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.318881 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3408670c-55e3-4bf0-9589-fef6200c3490-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-wx9l2\" (UID: \"3408670c-55e3-4bf0-9589-fef6200c3490\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wx9l2" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.318998 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/dbe65326-0856-4b33-8894-4c9266bc1a90-etcd-service-ca\") pod \"etcd-operator-b45778765-q2jq4\" (UID: \"dbe65326-0856-4b33-8894-4c9266bc1a90\") " pod="openshift-etcd-operator/etcd-operator-b45778765-q2jq4" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.320073 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/80ede041-caa4-451e-bdd1-083e2e03ec29-config\") pod \"kube-apiserver-operator-766d6c64bb-8qm2l\" (UID: \"80ede041-caa4-451e-bdd1-083e2e03ec29\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-8qm2l" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.320151 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/6c0dfaff-08a7-4152-8624-1973e4a4d148-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-gchnw\" (UID: \"6c0dfaff-08a7-4152-8624-1973e4a4d148\") " pod="openshift-authentication/oauth-openshift-558db77b4-gchnw" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.320178 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/c1c75259-4cf0-46b9-aea0-2c2ad2a6928c-certs\") pod \"machine-config-server-bndqf\" (UID: \"c1c75259-4cf0-46b9-aea0-2c2ad2a6928c\") " pod="openshift-machine-config-operator/machine-config-server-bndqf" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.320199 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/be8e4dfb-5b8a-4bf4-8d54-880d51f990de-srv-cert\") pod \"olm-operator-6b444d44fb-lv842\" (UID: \"be8e4dfb-5b8a-4bf4-8d54-880d51f990de\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-lv842" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.320728 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6c0dfaff-08a7-4152-8624-1973e4a4d148-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-gchnw\" (UID: \"6c0dfaff-08a7-4152-8624-1973e4a4d148\") " pod="openshift-authentication/oauth-openshift-558db77b4-gchnw" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.321365 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/c64d57c5-4bd4-48b5-85fc-9d613aea29c4-ca-trust-extracted\") pod \"image-registry-697d97f7c8-5sqj9\" (UID: \"c64d57c5-4bd4-48b5-85fc-9d613aea29c4\") " pod="openshift-image-registry/image-registry-697d97f7c8-5sqj9" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.321835 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/c64d57c5-4bd4-48b5-85fc-9d613aea29c4-registry-certificates\") pod \"image-registry-697d97f7c8-5sqj9\" (UID: \"c64d57c5-4bd4-48b5-85fc-9d613aea29c4\") " pod="openshift-image-registry/image-registry-697d97f7c8-5sqj9" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.321893 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/d3c66e2a-8404-498c-b8a9-2bd2e6fd318b-audit-dir\") pod \"apiserver-7bbb656c7d-lxp7v\" (UID: \"d3c66e2a-8404-498c-b8a9-2bd2e6fd318b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lxp7v" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.321935 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-btbzr\" (UniqueName: \"kubernetes.io/projected/d3c66e2a-8404-498c-b8a9-2bd2e6fd318b-kube-api-access-btbzr\") pod \"apiserver-7bbb656c7d-lxp7v\" (UID: \"d3c66e2a-8404-498c-b8a9-2bd2e6fd318b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lxp7v" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.321965 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/ebf750fb-f47c-420b-931d-3ac0a23e9e7b-default-certificate\") pod \"router-default-5444994796-ll9lb\" (UID: \"ebf750fb-f47c-420b-931d-3ac0a23e9e7b\") " pod="openshift-ingress/router-default-5444994796-ll9lb" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.322118 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/d3c66e2a-8404-498c-b8a9-2bd2e6fd318b-audit-dir\") pod \"apiserver-7bbb656c7d-lxp7v\" (UID: \"d3c66e2a-8404-498c-b8a9-2bd2e6fd318b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lxp7v" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.322136 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xjxn7\" (UniqueName: \"kubernetes.io/projected/091efc31-55eb-4026-bf38-626e09a13253-kube-api-access-xjxn7\") pod \"packageserver-d55dfcdfc-nrd6b\" (UID: \"091efc31-55eb-4026-bf38-626e09a13253\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nrd6b" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.322208 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/6c0dfaff-08a7-4152-8624-1973e4a4d148-audit-dir\") pod \"oauth-openshift-558db77b4-gchnw\" (UID: \"6c0dfaff-08a7-4152-8624-1973e4a4d148\") " pod="openshift-authentication/oauth-openshift-558db77b4-gchnw" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.322240 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/6c0dfaff-08a7-4152-8624-1973e4a4d148-audit-dir\") pod \"oauth-openshift-558db77b4-gchnw\" (UID: \"6c0dfaff-08a7-4152-8624-1973e4a4d148\") " pod="openshift-authentication/oauth-openshift-558db77b4-gchnw" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.322263 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e542cd22-7a8b-4a4d-8205-d3dc60b2b8c6-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-w6krf\" (UID: \"e542cd22-7a8b-4a4d-8205-d3dc60b2b8c6\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-w6krf" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.322296 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z9jr9\" (UniqueName: \"kubernetes.io/projected/c1c4fc78-c305-4dba-b4d5-9f05d58a8ef6-kube-api-access-z9jr9\") pod \"marketplace-operator-79b997595-ll5jq\" (UID: \"c1c4fc78-c305-4dba-b4d5-9f05d58a8ef6\") " pod="openshift-marketplace/marketplace-operator-79b997595-ll5jq" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.322326 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/bc8bf6c1-363b-4e24-830c-898e4f7c6325-metrics-tls\") pod \"dns-operator-744455d44c-2qmwc\" (UID: \"bc8bf6c1-363b-4e24-830c-898e4f7c6325\") " pod="openshift-dns-operator/dns-operator-744455d44c-2qmwc" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.322344 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/4da215ae-75ec-4ab1-81b0-39966d449b5f-auth-proxy-config\") pod \"machine-config-operator-74547568cd-hjpzm\" (UID: \"4da215ae-75ec-4ab1-81b0-39966d449b5f\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-hjpzm" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.322381 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/cce92b6a-69b0-49d5-bb66-36349ca96841-proxy-tls\") pod \"machine-config-controller-84d6567774-wz2dg\" (UID: \"cce92b6a-69b0-49d5-bb66-36349ca96841\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-wz2dg" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.322404 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/60ff4c5c-9e45-47d3-b45d-a365e988342e-config-volume\") pod \"collect-profiles-29328285-mrswx\" (UID: \"60ff4c5c-9e45-47d3-b45d-a365e988342e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29328285-mrswx" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.322423 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/dbe65326-0856-4b33-8894-4c9266bc1a90-etcd-client\") pod \"etcd-operator-b45778765-q2jq4\" (UID: \"dbe65326-0856-4b33-8894-4c9266bc1a90\") " pod="openshift-etcd-operator/etcd-operator-b45778765-q2jq4" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.322454 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mqjlq\" (UniqueName: \"kubernetes.io/projected/6c0dfaff-08a7-4152-8624-1973e4a4d148-kube-api-access-mqjlq\") pod \"oauth-openshift-558db77b4-gchnw\" (UID: \"6c0dfaff-08a7-4152-8624-1973e4a4d148\") " pod="openshift-authentication/oauth-openshift-558db77b4-gchnw" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.322473 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9263c4f4-1540-4190-9e85-b1d8e2ca6c2f-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-6wnmn\" (UID: \"9263c4f4-1540-4190-9e85-b1d8e2ca6c2f\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6wnmn" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.322541 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/dbe65326-0856-4b33-8894-4c9266bc1a90-etcd-ca\") pod \"etcd-operator-b45778765-q2jq4\" (UID: \"dbe65326-0856-4b33-8894-4c9266bc1a90\") " pod="openshift-etcd-operator/etcd-operator-b45778765-q2jq4" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.322589 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x42wt\" (UniqueName: \"kubernetes.io/projected/230b53bd-7c59-42f5-9604-0c5414fcdf80-kube-api-access-x42wt\") pod \"ingress-canary-7g427\" (UID: \"230b53bd-7c59-42f5-9604-0c5414fcdf80\") " pod="openshift-ingress-canary/ingress-canary-7g427" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.322609 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/be8e4dfb-5b8a-4bf4-8d54-880d51f990de-profile-collector-cert\") pod \"olm-operator-6b444d44fb-lv842\" (UID: \"be8e4dfb-5b8a-4bf4-8d54-880d51f990de\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-lv842" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.322626 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/5a5b1574-7f76-4bcf-8c24-9ff7773e86f2-registration-dir\") pod \"csi-hostpathplugin-s4zph\" (UID: \"5a5b1574-7f76-4bcf-8c24-9ff7773e86f2\") " pod="hostpath-provisioner/csi-hostpathplugin-s4zph" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.322649 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/6c0dfaff-08a7-4152-8624-1973e4a4d148-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-gchnw\" (UID: \"6c0dfaff-08a7-4152-8624-1973e4a4d148\") " pod="openshift-authentication/oauth-openshift-558db77b4-gchnw" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.322667 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/091efc31-55eb-4026-bf38-626e09a13253-tmpfs\") pod \"packageserver-d55dfcdfc-nrd6b\" (UID: \"091efc31-55eb-4026-bf38-626e09a13253\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nrd6b" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.322685 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c1c4fc78-c305-4dba-b4d5-9f05d58a8ef6-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-ll5jq\" (UID: \"c1c4fc78-c305-4dba-b4d5-9f05d58a8ef6\") " pod="openshift-marketplace/marketplace-operator-79b997595-ll5jq" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.322794 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/d3c66e2a-8404-498c-b8a9-2bd2e6fd318b-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-lxp7v\" (UID: \"d3c66e2a-8404-498c-b8a9-2bd2e6fd318b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lxp7v" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.322821 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3408670c-55e3-4bf0-9589-fef6200c3490-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-wx9l2\" (UID: \"3408670c-55e3-4bf0-9589-fef6200c3490\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wx9l2" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.322845 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cjn98\" (UniqueName: \"kubernetes.io/projected/11048ba4-fef7-4675-b3c3-2671af4a4e27-kube-api-access-cjn98\") pod \"kube-storage-version-migrator-operator-b67b599dd-rl5n7\" (UID: \"11048ba4-fef7-4675-b3c3-2671af4a4e27\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rl5n7" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.322867 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/5a5b1574-7f76-4bcf-8c24-9ff7773e86f2-mountpoint-dir\") pod \"csi-hostpathplugin-s4zph\" (UID: \"5a5b1574-7f76-4bcf-8c24-9ff7773e86f2\") " pod="hostpath-provisioner/csi-hostpathplugin-s4zph" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.322897 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e542cd22-7a8b-4a4d-8205-d3dc60b2b8c6-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-w6krf\" (UID: \"e542cd22-7a8b-4a4d-8205-d3dc60b2b8c6\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-w6krf" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.323377 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/dbe65326-0856-4b33-8894-4c9266bc1a90-etcd-ca\") pod \"etcd-operator-b45778765-q2jq4\" (UID: \"dbe65326-0856-4b33-8894-4c9266bc1a90\") " pod="openshift-etcd-operator/etcd-operator-b45778765-q2jq4" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.323978 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/d3c66e2a-8404-498c-b8a9-2bd2e6fd318b-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-lxp7v\" (UID: \"d3c66e2a-8404-498c-b8a9-2bd2e6fd318b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lxp7v" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.324453 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/6c0dfaff-08a7-4152-8624-1973e4a4d148-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-gchnw\" (UID: \"6c0dfaff-08a7-4152-8624-1973e4a4d148\") " pod="openshift-authentication/oauth-openshift-558db77b4-gchnw" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.324483 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/830a0a27-fab1-4efc-aeb3-e38a962ffcfa-config-volume\") pod \"dns-default-6xm7j\" (UID: \"830a0a27-fab1-4efc-aeb3-e38a962ffcfa\") " pod="openshift-dns/dns-default-6xm7j" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.324519 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ebf750fb-f47c-420b-931d-3ac0a23e9e7b-metrics-certs\") pod \"router-default-5444994796-ll9lb\" (UID: \"ebf750fb-f47c-420b-931d-3ac0a23e9e7b\") " pod="openshift-ingress/router-default-5444994796-ll9lb" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.324569 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1e16d79e-eb1e-4b8b-8f0a-acb370163d17-client-ca\") pod \"route-controller-manager-6576b87f9c-pnhf9\" (UID: \"1e16d79e-eb1e-4b8b-8f0a-acb370163d17\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-pnhf9" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.324632 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1e16d79e-eb1e-4b8b-8f0a-acb370163d17-serving-cert\") pod \"route-controller-manager-6576b87f9c-pnhf9\" (UID: \"1e16d79e-eb1e-4b8b-8f0a-acb370163d17\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-pnhf9" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.324655 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gb999\" (UniqueName: \"kubernetes.io/projected/be8e4dfb-5b8a-4bf4-8d54-880d51f990de-kube-api-access-gb999\") pod \"olm-operator-6b444d44fb-lv842\" (UID: \"be8e4dfb-5b8a-4bf4-8d54-880d51f990de\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-lv842" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.324676 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d3c66e2a-8404-498c-b8a9-2bd2e6fd318b-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-lxp7v\" (UID: \"d3c66e2a-8404-498c-b8a9-2bd2e6fd318b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lxp7v" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.324698 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2e5d7ee3-ea83-4445-a74d-9efa3864b3bb-serving-cert\") pod \"service-ca-operator-777779d784-nrgbt\" (UID: \"2e5d7ee3-ea83-4445-a74d-9efa3864b3bb\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-nrgbt" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.325675 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/d3c66e2a-8404-498c-b8a9-2bd2e6fd318b-encryption-config\") pod \"apiserver-7bbb656c7d-lxp7v\" (UID: \"d3c66e2a-8404-498c-b8a9-2bd2e6fd318b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lxp7v" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.325729 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dbe65326-0856-4b33-8894-4c9266bc1a90-serving-cert\") pod \"etcd-operator-b45778765-q2jq4\" (UID: \"dbe65326-0856-4b33-8894-4c9266bc1a90\") " pod="openshift-etcd-operator/etcd-operator-b45778765-q2jq4" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.325751 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1e16d79e-eb1e-4b8b-8f0a-acb370163d17-config\") pod \"route-controller-manager-6576b87f9c-pnhf9\" (UID: \"1e16d79e-eb1e-4b8b-8f0a-acb370163d17\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-pnhf9" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.325820 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9263c4f4-1540-4190-9e85-b1d8e2ca6c2f-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-6wnmn\" (UID: \"9263c4f4-1540-4190-9e85-b1d8e2ca6c2f\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6wnmn" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.325856 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/091efc31-55eb-4026-bf38-626e09a13253-apiservice-cert\") pod \"packageserver-d55dfcdfc-nrd6b\" (UID: \"091efc31-55eb-4026-bf38-626e09a13253\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nrd6b" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.325882 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9263c4f4-1540-4190-9e85-b1d8e2ca6c2f-config\") pod \"kube-controller-manager-operator-78b949d7b-6wnmn\" (UID: \"9263c4f4-1540-4190-9e85-b1d8e2ca6c2f\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6wnmn" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.325901 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/b96341ba-da55-442f-851e-5666717a1d5a-metrics-tls\") pod \"ingress-operator-5b745b69d9-9vndv\" (UID: \"b96341ba-da55-442f-851e-5666717a1d5a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-9vndv" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.325927 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d3c66e2a-8404-498c-b8a9-2bd2e6fd318b-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-lxp7v\" (UID: \"d3c66e2a-8404-498c-b8a9-2bd2e6fd318b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lxp7v" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.326198 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/6c0dfaff-08a7-4152-8624-1973e4a4d148-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-gchnw\" (UID: \"6c0dfaff-08a7-4152-8624-1973e4a4d148\") " pod="openshift-authentication/oauth-openshift-558db77b4-gchnw" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.326235 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/6c0dfaff-08a7-4152-8624-1973e4a4d148-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-gchnw\" (UID: \"6c0dfaff-08a7-4152-8624-1973e4a4d148\") " pod="openshift-authentication/oauth-openshift-558db77b4-gchnw" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.326771 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/6c0dfaff-08a7-4152-8624-1973e4a4d148-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-gchnw\" (UID: \"6c0dfaff-08a7-4152-8624-1973e4a4d148\") " pod="openshift-authentication/oauth-openshift-558db77b4-gchnw" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.329110 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b157fa21-ae67-4c98-806c-93854b072965-serving-cert\") pod \"openshift-config-operator-7777fb866f-cwtlx\" (UID: \"b157fa21-ae67-4c98-806c-93854b072965\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-cwtlx" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.329566 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1e16d79e-eb1e-4b8b-8f0a-acb370163d17-client-ca\") pod \"route-controller-manager-6576b87f9c-pnhf9\" (UID: \"1e16d79e-eb1e-4b8b-8f0a-acb370163d17\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-pnhf9" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.329614 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11048ba4-fef7-4675-b3c3-2671af4a4e27-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-rl5n7\" (UID: \"11048ba4-fef7-4675-b3c3-2671af4a4e27\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rl5n7" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.329643 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b96341ba-da55-442f-851e-5666717a1d5a-trusted-ca\") pod \"ingress-operator-5b745b69d9-9vndv\" (UID: \"b96341ba-da55-442f-851e-5666717a1d5a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-9vndv" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.329661 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/4da215ae-75ec-4ab1-81b0-39966d449b5f-proxy-tls\") pod \"machine-config-operator-74547568cd-hjpzm\" (UID: \"4da215ae-75ec-4ab1-81b0-39966d449b5f\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-hjpzm" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.329733 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wtxlp\" (UniqueName: \"kubernetes.io/projected/b157fa21-ae67-4c98-806c-93854b072965-kube-api-access-wtxlp\") pod \"openshift-config-operator-7777fb866f-cwtlx\" (UID: \"b157fa21-ae67-4c98-806c-93854b072965\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-cwtlx" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.329869 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dw4mx\" (UniqueName: \"kubernetes.io/projected/3934dd94-b9fa-4079-a268-60a9470dab41-kube-api-access-dw4mx\") pod \"downloads-7954f5f757-ksczt\" (UID: \"3934dd94-b9fa-4079-a268-60a9470dab41\") " pod="openshift-console/downloads-7954f5f757-ksczt" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.329909 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s6vzh\" (UniqueName: \"kubernetes.io/projected/4da215ae-75ec-4ab1-81b0-39966d449b5f-kube-api-access-s6vzh\") pod \"machine-config-operator-74547568cd-hjpzm\" (UID: \"4da215ae-75ec-4ab1-81b0-39966d449b5f\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-hjpzm" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.329937 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bmkt9\" (UniqueName: \"kubernetes.io/projected/c2c0ebd0-5cd9-49f4-9cd6-df5995cfcd74-kube-api-access-bmkt9\") pod \"catalog-operator-68c6474976-jpbfr\" (UID: \"c2c0ebd0-5cd9-49f4-9cd6-df5995cfcd74\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jpbfr" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.329967 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7pfxf\" (UniqueName: \"kubernetes.io/projected/dbe65326-0856-4b33-8894-4c9266bc1a90-kube-api-access-7pfxf\") pod \"etcd-operator-b45778765-q2jq4\" (UID: \"dbe65326-0856-4b33-8894-4c9266bc1a90\") " pod="openshift-etcd-operator/etcd-operator-b45778765-q2jq4" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.329989 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p4pmn\" (UniqueName: \"kubernetes.io/projected/b96341ba-da55-442f-851e-5666717a1d5a-kube-api-access-p4pmn\") pod \"ingress-operator-5b745b69d9-9vndv\" (UID: \"b96341ba-da55-442f-851e-5666717a1d5a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-9vndv" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.330011 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/6c0dfaff-08a7-4152-8624-1973e4a4d148-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-gchnw\" (UID: \"6c0dfaff-08a7-4152-8624-1973e4a4d148\") " pod="openshift-authentication/oauth-openshift-558db77b4-gchnw" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.330082 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-944m8\" (UniqueName: \"kubernetes.io/projected/51547f65-61cc-4504-8399-5d540160fea1-kube-api-access-944m8\") pod \"multus-admission-controller-857f4d67dd-682zv\" (UID: \"51547f65-61cc-4504-8399-5d540160fea1\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-682zv" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.331064 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/6c0dfaff-08a7-4152-8624-1973e4a4d148-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-gchnw\" (UID: \"6c0dfaff-08a7-4152-8624-1973e4a4d148\") " pod="openshift-authentication/oauth-openshift-558db77b4-gchnw" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.331218 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/5a5b1574-7f76-4bcf-8c24-9ff7773e86f2-socket-dir\") pod \"csi-hostpathplugin-s4zph\" (UID: \"5a5b1574-7f76-4bcf-8c24-9ff7773e86f2\") " pod="hostpath-provisioner/csi-hostpathplugin-s4zph" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.331546 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/80ede041-caa4-451e-bdd1-083e2e03ec29-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-8qm2l\" (UID: \"80ede041-caa4-451e-bdd1-083e2e03ec29\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-8qm2l" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.331608 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xpdzs\" (UniqueName: \"kubernetes.io/projected/d1b2b973-8b95-4f6a-83f1-468f84fb006d-kube-api-access-xpdzs\") pod \"control-plane-machine-set-operator-78cbb6b69f-5wx56\" (UID: \"d1b2b973-8b95-4f6a-83f1-468f84fb006d\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-5wx56" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.331629 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/ebf750fb-f47c-420b-931d-3ac0a23e9e7b-stats-auth\") pod \"router-default-5444994796-ll9lb\" (UID: \"ebf750fb-f47c-420b-931d-3ac0a23e9e7b\") " pod="openshift-ingress/router-default-5444994796-ll9lb" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.331679 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/6c0dfaff-08a7-4152-8624-1973e4a4d148-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-gchnw\" (UID: \"6c0dfaff-08a7-4152-8624-1973e4a4d148\") " pod="openshift-authentication/oauth-openshift-558db77b4-gchnw" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.332285 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/6c0dfaff-08a7-4152-8624-1973e4a4d148-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-gchnw\" (UID: \"6c0dfaff-08a7-4152-8624-1973e4a4d148\") " pod="openshift-authentication/oauth-openshift-558db77b4-gchnw" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.332409 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/60ff4c5c-9e45-47d3-b45d-a365e988342e-secret-volume\") pod \"collect-profiles-29328285-mrswx\" (UID: \"60ff4c5c-9e45-47d3-b45d-a365e988342e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29328285-mrswx" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.332452 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/c64d57c5-4bd4-48b5-85fc-9d613aea29c4-registry-tls\") pod \"image-registry-697d97f7c8-5sqj9\" (UID: \"c64d57c5-4bd4-48b5-85fc-9d613aea29c4\") " pod="openshift-image-registry/image-registry-697d97f7c8-5sqj9" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.332515 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/6c0dfaff-08a7-4152-8624-1973e4a4d148-audit-policies\") pod \"oauth-openshift-558db77b4-gchnw\" (UID: \"6c0dfaff-08a7-4152-8624-1973e4a4d148\") " pod="openshift-authentication/oauth-openshift-558db77b4-gchnw" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.332544 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jkkxq\" (UniqueName: \"kubernetes.io/projected/c1c75259-4cf0-46b9-aea0-2c2ad2a6928c-kube-api-access-jkkxq\") pod \"machine-config-server-bndqf\" (UID: \"c1c75259-4cf0-46b9-aea0-2c2ad2a6928c\") " pod="openshift-machine-config-operator/machine-config-server-bndqf" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.332566 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/c2c0ebd0-5cd9-49f4-9cd6-df5995cfcd74-srv-cert\") pod \"catalog-operator-68c6474976-jpbfr\" (UID: \"c2c0ebd0-5cd9-49f4-9cd6-df5995cfcd74\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jpbfr" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.332603 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/c2c0ebd0-5cd9-49f4-9cd6-df5995cfcd74-profile-collector-cert\") pod \"catalog-operator-68c6474976-jpbfr\" (UID: \"c2c0ebd0-5cd9-49f4-9cd6-df5995cfcd74\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jpbfr" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.332662 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gv99q\" (UniqueName: \"kubernetes.io/projected/5a5b1574-7f76-4bcf-8c24-9ff7773e86f2-kube-api-access-gv99q\") pod \"csi-hostpathplugin-s4zph\" (UID: \"5a5b1574-7f76-4bcf-8c24-9ff7773e86f2\") " pod="hostpath-provisioner/csi-hostpathplugin-s4zph" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.332688 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/6c0dfaff-08a7-4152-8624-1973e4a4d148-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-gchnw\" (UID: \"6c0dfaff-08a7-4152-8624-1973e4a4d148\") " pod="openshift-authentication/oauth-openshift-558db77b4-gchnw" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.332713 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/11048ba4-fef7-4675-b3c3-2671af4a4e27-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-rl5n7\" (UID: \"11048ba4-fef7-4675-b3c3-2671af4a4e27\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rl5n7" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.332746 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/5a5b1574-7f76-4bcf-8c24-9ff7773e86f2-csi-data-dir\") pod \"csi-hostpathplugin-s4zph\" (UID: \"5a5b1574-7f76-4bcf-8c24-9ff7773e86f2\") " pod="hostpath-provisioner/csi-hostpathplugin-s4zph" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.332790 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wtph5\" (UniqueName: \"kubernetes.io/projected/bc8bf6c1-363b-4e24-830c-898e4f7c6325-kube-api-access-wtph5\") pod \"dns-operator-744455d44c-2qmwc\" (UID: \"bc8bf6c1-363b-4e24-830c-898e4f7c6325\") " pod="openshift-dns-operator/dns-operator-744455d44c-2qmwc" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.332865 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/51547f65-61cc-4504-8399-5d540160fea1-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-682zv\" (UID: \"51547f65-61cc-4504-8399-5d540160fea1\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-682zv" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.332905 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ebf750fb-f47c-420b-931d-3ac0a23e9e7b-service-ca-bundle\") pod \"router-default-5444994796-ll9lb\" (UID: \"ebf750fb-f47c-420b-931d-3ac0a23e9e7b\") " pod="openshift-ingress/router-default-5444994796-ll9lb" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.332945 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/6c0dfaff-08a7-4152-8624-1973e4a4d148-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-gchnw\" (UID: \"6c0dfaff-08a7-4152-8624-1973e4a4d148\") " pod="openshift-authentication/oauth-openshift-558db77b4-gchnw" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.332984 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/cce92b6a-69b0-49d5-bb66-36349ca96841-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-wz2dg\" (UID: \"cce92b6a-69b0-49d5-bb66-36349ca96841\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-wz2dg" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.333030 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7p9rv\" (UniqueName: \"kubernetes.io/projected/0593c91b-48f4-47aa-a57f-f9b50d7f385e-kube-api-access-7p9rv\") pod \"migrator-59844c95c7-67bjd\" (UID: \"0593c91b-48f4-47aa-a57f-f9b50d7f385e\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-67bjd" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.337082 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/6c0dfaff-08a7-4152-8624-1973e4a4d148-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-gchnw\" (UID: \"6c0dfaff-08a7-4152-8624-1973e4a4d148\") " pod="openshift-authentication/oauth-openshift-558db77b4-gchnw" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.338201 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dbe65326-0856-4b33-8894-4c9266bc1a90-serving-cert\") pod \"etcd-operator-b45778765-q2jq4\" (UID: \"dbe65326-0856-4b33-8894-4c9266bc1a90\") " pod="openshift-etcd-operator/etcd-operator-b45778765-q2jq4" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.339070 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/d3c66e2a-8404-498c-b8a9-2bd2e6fd318b-encryption-config\") pod \"apiserver-7bbb656c7d-lxp7v\" (UID: \"d3c66e2a-8404-498c-b8a9-2bd2e6fd318b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lxp7v" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.339434 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/6c0dfaff-08a7-4152-8624-1973e4a4d148-audit-policies\") pod \"oauth-openshift-558db77b4-gchnw\" (UID: \"6c0dfaff-08a7-4152-8624-1973e4a4d148\") " pod="openshift-authentication/oauth-openshift-558db77b4-gchnw" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.339489 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/6c0dfaff-08a7-4152-8624-1973e4a4d148-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-gchnw\" (UID: \"6c0dfaff-08a7-4152-8624-1973e4a4d148\") " pod="openshift-authentication/oauth-openshift-558db77b4-gchnw" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.339745 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8zqxw\" (UniqueName: \"kubernetes.io/projected/2e5d7ee3-ea83-4445-a74d-9efa3864b3bb-kube-api-access-8zqxw\") pod \"service-ca-operator-777779d784-nrgbt\" (UID: \"2e5d7ee3-ea83-4445-a74d-9efa3864b3bb\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-nrgbt" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.339834 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cjb58\" (UniqueName: \"kubernetes.io/projected/830a0a27-fab1-4efc-aeb3-e38a962ffcfa-kube-api-access-cjb58\") pod \"dns-default-6xm7j\" (UID: \"830a0a27-fab1-4efc-aeb3-e38a962ffcfa\") " pod="openshift-dns/dns-default-6xm7j" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.339883 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/230b53bd-7c59-42f5-9604-0c5414fcdf80-cert\") pod \"ingress-canary-7g427\" (UID: \"230b53bd-7c59-42f5-9604-0c5414fcdf80\") " pod="openshift-ingress-canary/ingress-canary-7g427" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.340190 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/830a0a27-fab1-4efc-aeb3-e38a962ffcfa-metrics-tls\") pod \"dns-default-6xm7j\" (UID: \"830a0a27-fab1-4efc-aeb3-e38a962ffcfa\") " pod="openshift-dns/dns-default-6xm7j" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.340252 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rfjtr\" (UniqueName: \"kubernetes.io/projected/8797a357-3b55-4d63-9215-4c64bb676a3d-kube-api-access-rfjtr\") pod \"service-ca-9c57cc56f-chv7x\" (UID: \"8797a357-3b55-4d63-9215-4c64bb676a3d\") " pod="openshift-service-ca/service-ca-9c57cc56f-chv7x" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.340451 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c64d57c5-4bd4-48b5-85fc-9d613aea29c4-trusted-ca\") pod \"image-registry-697d97f7c8-5sqj9\" (UID: \"c64d57c5-4bd4-48b5-85fc-9d613aea29c4\") " pod="openshift-image-registry/image-registry-697d97f7c8-5sqj9" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.340538 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/8797a357-3b55-4d63-9215-4c64bb676a3d-signing-cabundle\") pod \"service-ca-9c57cc56f-chv7x\" (UID: \"8797a357-3b55-4d63-9215-4c64bb676a3d\") " pod="openshift-service-ca/service-ca-9c57cc56f-chv7x" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.340642 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/d3c66e2a-8404-498c-b8a9-2bd2e6fd318b-audit-policies\") pod \"apiserver-7bbb656c7d-lxp7v\" (UID: \"d3c66e2a-8404-498c-b8a9-2bd2e6fd318b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lxp7v" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.340665 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/80ede041-caa4-451e-bdd1-083e2e03ec29-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-8qm2l\" (UID: \"80ede041-caa4-451e-bdd1-083e2e03ec29\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-8qm2l" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.344079 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/c64d57c5-4bd4-48b5-85fc-9d613aea29c4-registry-tls\") pod \"image-registry-697d97f7c8-5sqj9\" (UID: \"c64d57c5-4bd4-48b5-85fc-9d613aea29c4\") " pod="openshift-image-registry/image-registry-697d97f7c8-5sqj9" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.344188 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1e16d79e-eb1e-4b8b-8f0a-acb370163d17-config\") pod \"route-controller-manager-6576b87f9c-pnhf9\" (UID: \"1e16d79e-eb1e-4b8b-8f0a-acb370163d17\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-pnhf9" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.344229 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7qtr7\" (UniqueName: \"kubernetes.io/projected/c64d57c5-4bd4-48b5-85fc-9d613aea29c4-kube-api-access-7qtr7\") pod \"image-registry-697d97f7c8-5sqj9\" (UID: \"c64d57c5-4bd4-48b5-85fc-9d613aea29c4\") " pod="openshift-image-registry/image-registry-697d97f7c8-5sqj9" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.344290 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/8797a357-3b55-4d63-9215-4c64bb676a3d-signing-key\") pod \"service-ca-9c57cc56f-chv7x\" (UID: \"8797a357-3b55-4d63-9215-4c64bb676a3d\") " pod="openshift-service-ca/service-ca-9c57cc56f-chv7x" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.344431 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/d3c66e2a-8404-498c-b8a9-2bd2e6fd318b-audit-policies\") pod \"apiserver-7bbb656c7d-lxp7v\" (UID: \"d3c66e2a-8404-498c-b8a9-2bd2e6fd318b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lxp7v" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.345374 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/6c0dfaff-08a7-4152-8624-1973e4a4d148-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-gchnw\" (UID: \"6c0dfaff-08a7-4152-8624-1973e4a4d148\") " pod="openshift-authentication/oauth-openshift-558db77b4-gchnw" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.346450 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/dbe65326-0856-4b33-8894-4c9266bc1a90-etcd-client\") pod \"etcd-operator-b45778765-q2jq4\" (UID: \"dbe65326-0856-4b33-8894-4c9266bc1a90\") " pod="openshift-etcd-operator/etcd-operator-b45778765-q2jq4" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.347515 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d3c66e2a-8404-498c-b8a9-2bd2e6fd318b-serving-cert\") pod \"apiserver-7bbb656c7d-lxp7v\" (UID: \"d3c66e2a-8404-498c-b8a9-2bd2e6fd318b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lxp7v" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.348011 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/c64d57c5-4bd4-48b5-85fc-9d613aea29c4-installation-pull-secrets\") pod \"image-registry-697d97f7c8-5sqj9\" (UID: \"c64d57c5-4bd4-48b5-85fc-9d613aea29c4\") " pod="openshift-image-registry/image-registry-697d97f7c8-5sqj9" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.348067 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/d3c66e2a-8404-498c-b8a9-2bd2e6fd318b-etcd-client\") pod \"apiserver-7bbb656c7d-lxp7v\" (UID: \"d3c66e2a-8404-498c-b8a9-2bd2e6fd318b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lxp7v" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.349995 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3408670c-55e3-4bf0-9589-fef6200c3490-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-wx9l2\" (UID: \"3408670c-55e3-4bf0-9589-fef6200c3490\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wx9l2" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.350047 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/bc8bf6c1-363b-4e24-830c-898e4f7c6325-metrics-tls\") pod \"dns-operator-744455d44c-2qmwc\" (UID: \"bc8bf6c1-363b-4e24-830c-898e4f7c6325\") " pod="openshift-dns-operator/dns-operator-744455d44c-2qmwc" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.353262 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/6c0dfaff-08a7-4152-8624-1973e4a4d148-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-gchnw\" (UID: \"6c0dfaff-08a7-4152-8624-1973e4a4d148\") " pod="openshift-authentication/oauth-openshift-558db77b4-gchnw" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.357428 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c64d57c5-4bd4-48b5-85fc-9d613aea29c4-bound-sa-token\") pod \"image-registry-697d97f7c8-5sqj9\" (UID: \"c64d57c5-4bd4-48b5-85fc-9d613aea29c4\") " pod="openshift-image-registry/image-registry-697d97f7c8-5sqj9" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.360030 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1e16d79e-eb1e-4b8b-8f0a-acb370163d17-serving-cert\") pod \"route-controller-manager-6576b87f9c-pnhf9\" (UID: \"1e16d79e-eb1e-4b8b-8f0a-acb370163d17\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-pnhf9" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.360438 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c64d57c5-4bd4-48b5-85fc-9d613aea29c4-trusted-ca\") pod \"image-registry-697d97f7c8-5sqj9\" (UID: \"c64d57c5-4bd4-48b5-85fc-9d613aea29c4\") " pod="openshift-image-registry/image-registry-697d97f7c8-5sqj9" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.363811 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/6c0dfaff-08a7-4152-8624-1973e4a4d148-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-gchnw\" (UID: \"6c0dfaff-08a7-4152-8624-1973e4a4d148\") " pod="openshift-authentication/oauth-openshift-558db77b4-gchnw" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.363939 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/6c0dfaff-08a7-4152-8624-1973e4a4d148-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-gchnw\" (UID: \"6c0dfaff-08a7-4152-8624-1973e4a4d148\") " pod="openshift-authentication/oauth-openshift-558db77b4-gchnw" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.373075 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hmrqj\" (UniqueName: \"kubernetes.io/projected/3408670c-55e3-4bf0-9589-fef6200c3490-kube-api-access-hmrqj\") pod \"openshift-controller-manager-operator-756b6f6bc6-wx9l2\" (UID: \"3408670c-55e3-4bf0-9589-fef6200c3490\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wx9l2" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.393151 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9wb4b\" (UniqueName: \"kubernetes.io/projected/1e16d79e-eb1e-4b8b-8f0a-acb370163d17-kube-api-access-9wb4b\") pod \"route-controller-manager-6576b87f9c-pnhf9\" (UID: \"1e16d79e-eb1e-4b8b-8f0a-acb370163d17\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-pnhf9" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.416392 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-btbzr\" (UniqueName: \"kubernetes.io/projected/d3c66e2a-8404-498c-b8a9-2bd2e6fd318b-kube-api-access-btbzr\") pod \"apiserver-7bbb656c7d-lxp7v\" (UID: \"d3c66e2a-8404-498c-b8a9-2bd2e6fd318b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lxp7v" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.436037 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mqjlq\" (UniqueName: \"kubernetes.io/projected/6c0dfaff-08a7-4152-8624-1973e4a4d148-kube-api-access-mqjlq\") pod \"oauth-openshift-558db77b4-gchnw\" (UID: \"6c0dfaff-08a7-4152-8624-1973e4a4d148\") " pod="openshift-authentication/oauth-openshift-558db77b4-gchnw" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.445553 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.445868 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cjn98\" (UniqueName: \"kubernetes.io/projected/11048ba4-fef7-4675-b3c3-2671af4a4e27-kube-api-access-cjn98\") pod \"kube-storage-version-migrator-operator-b67b599dd-rl5n7\" (UID: \"11048ba4-fef7-4675-b3c3-2671af4a4e27\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rl5n7" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.445892 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/5a5b1574-7f76-4bcf-8c24-9ff7773e86f2-mountpoint-dir\") pod \"csi-hostpathplugin-s4zph\" (UID: \"5a5b1574-7f76-4bcf-8c24-9ff7773e86f2\") " pod="hostpath-provisioner/csi-hostpathplugin-s4zph" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.445913 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e542cd22-7a8b-4a4d-8205-d3dc60b2b8c6-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-w6krf\" (UID: \"e542cd22-7a8b-4a4d-8205-d3dc60b2b8c6\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-w6krf" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.445931 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/830a0a27-fab1-4efc-aeb3-e38a962ffcfa-config-volume\") pod \"dns-default-6xm7j\" (UID: \"830a0a27-fab1-4efc-aeb3-e38a962ffcfa\") " pod="openshift-dns/dns-default-6xm7j" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.445948 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ebf750fb-f47c-420b-931d-3ac0a23e9e7b-metrics-certs\") pod \"router-default-5444994796-ll9lb\" (UID: \"ebf750fb-f47c-420b-931d-3ac0a23e9e7b\") " pod="openshift-ingress/router-default-5444994796-ll9lb" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.445968 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gb999\" (UniqueName: \"kubernetes.io/projected/be8e4dfb-5b8a-4bf4-8d54-880d51f990de-kube-api-access-gb999\") pod \"olm-operator-6b444d44fb-lv842\" (UID: \"be8e4dfb-5b8a-4bf4-8d54-880d51f990de\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-lv842" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.445986 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2e5d7ee3-ea83-4445-a74d-9efa3864b3bb-serving-cert\") pod \"service-ca-operator-777779d784-nrgbt\" (UID: \"2e5d7ee3-ea83-4445-a74d-9efa3864b3bb\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-nrgbt" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.446004 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9263c4f4-1540-4190-9e85-b1d8e2ca6c2f-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-6wnmn\" (UID: \"9263c4f4-1540-4190-9e85-b1d8e2ca6c2f\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6wnmn" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.446070 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/091efc31-55eb-4026-bf38-626e09a13253-apiservice-cert\") pod \"packageserver-d55dfcdfc-nrd6b\" (UID: \"091efc31-55eb-4026-bf38-626e09a13253\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nrd6b" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.446087 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9263c4f4-1540-4190-9e85-b1d8e2ca6c2f-config\") pod \"kube-controller-manager-operator-78b949d7b-6wnmn\" (UID: \"9263c4f4-1540-4190-9e85-b1d8e2ca6c2f\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6wnmn" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.446124 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/b96341ba-da55-442f-851e-5666717a1d5a-metrics-tls\") pod \"ingress-operator-5b745b69d9-9vndv\" (UID: \"b96341ba-da55-442f-851e-5666717a1d5a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-9vndv" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.446144 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11048ba4-fef7-4675-b3c3-2671af4a4e27-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-rl5n7\" (UID: \"11048ba4-fef7-4675-b3c3-2671af4a4e27\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rl5n7" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.446160 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b96341ba-da55-442f-851e-5666717a1d5a-trusted-ca\") pod \"ingress-operator-5b745b69d9-9vndv\" (UID: \"b96341ba-da55-442f-851e-5666717a1d5a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-9vndv" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.446177 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/4da215ae-75ec-4ab1-81b0-39966d449b5f-proxy-tls\") pod \"machine-config-operator-74547568cd-hjpzm\" (UID: \"4da215ae-75ec-4ab1-81b0-39966d449b5f\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-hjpzm" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.446206 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s6vzh\" (UniqueName: \"kubernetes.io/projected/4da215ae-75ec-4ab1-81b0-39966d449b5f-kube-api-access-s6vzh\") pod \"machine-config-operator-74547568cd-hjpzm\" (UID: \"4da215ae-75ec-4ab1-81b0-39966d449b5f\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-hjpzm" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.446222 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bmkt9\" (UniqueName: \"kubernetes.io/projected/c2c0ebd0-5cd9-49f4-9cd6-df5995cfcd74-kube-api-access-bmkt9\") pod \"catalog-operator-68c6474976-jpbfr\" (UID: \"c2c0ebd0-5cd9-49f4-9cd6-df5995cfcd74\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jpbfr" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.446248 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p4pmn\" (UniqueName: \"kubernetes.io/projected/b96341ba-da55-442f-851e-5666717a1d5a-kube-api-access-p4pmn\") pod \"ingress-operator-5b745b69d9-9vndv\" (UID: \"b96341ba-da55-442f-851e-5666717a1d5a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-9vndv" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.446268 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-944m8\" (UniqueName: \"kubernetes.io/projected/51547f65-61cc-4504-8399-5d540160fea1-kube-api-access-944m8\") pod \"multus-admission-controller-857f4d67dd-682zv\" (UID: \"51547f65-61cc-4504-8399-5d540160fea1\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-682zv" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.446288 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/5a5b1574-7f76-4bcf-8c24-9ff7773e86f2-socket-dir\") pod \"csi-hostpathplugin-s4zph\" (UID: \"5a5b1574-7f76-4bcf-8c24-9ff7773e86f2\") " pod="hostpath-provisioner/csi-hostpathplugin-s4zph" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.446309 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/80ede041-caa4-451e-bdd1-083e2e03ec29-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-8qm2l\" (UID: \"80ede041-caa4-451e-bdd1-083e2e03ec29\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-8qm2l" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.446337 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xpdzs\" (UniqueName: \"kubernetes.io/projected/d1b2b973-8b95-4f6a-83f1-468f84fb006d-kube-api-access-xpdzs\") pod \"control-plane-machine-set-operator-78cbb6b69f-5wx56\" (UID: \"d1b2b973-8b95-4f6a-83f1-468f84fb006d\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-5wx56" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.446355 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/ebf750fb-f47c-420b-931d-3ac0a23e9e7b-stats-auth\") pod \"router-default-5444994796-ll9lb\" (UID: \"ebf750fb-f47c-420b-931d-3ac0a23e9e7b\") " pod="openshift-ingress/router-default-5444994796-ll9lb" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.446375 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/60ff4c5c-9e45-47d3-b45d-a365e988342e-secret-volume\") pod \"collect-profiles-29328285-mrswx\" (UID: \"60ff4c5c-9e45-47d3-b45d-a365e988342e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29328285-mrswx" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.446394 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jkkxq\" (UniqueName: \"kubernetes.io/projected/c1c75259-4cf0-46b9-aea0-2c2ad2a6928c-kube-api-access-jkkxq\") pod \"machine-config-server-bndqf\" (UID: \"c1c75259-4cf0-46b9-aea0-2c2ad2a6928c\") " pod="openshift-machine-config-operator/machine-config-server-bndqf" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.446411 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/c2c0ebd0-5cd9-49f4-9cd6-df5995cfcd74-srv-cert\") pod \"catalog-operator-68c6474976-jpbfr\" (UID: \"c2c0ebd0-5cd9-49f4-9cd6-df5995cfcd74\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jpbfr" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.446430 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/c2c0ebd0-5cd9-49f4-9cd6-df5995cfcd74-profile-collector-cert\") pod \"catalog-operator-68c6474976-jpbfr\" (UID: \"c2c0ebd0-5cd9-49f4-9cd6-df5995cfcd74\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jpbfr" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.446448 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gv99q\" (UniqueName: \"kubernetes.io/projected/5a5b1574-7f76-4bcf-8c24-9ff7773e86f2-kube-api-access-gv99q\") pod \"csi-hostpathplugin-s4zph\" (UID: \"5a5b1574-7f76-4bcf-8c24-9ff7773e86f2\") " pod="hostpath-provisioner/csi-hostpathplugin-s4zph" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.446468 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/11048ba4-fef7-4675-b3c3-2671af4a4e27-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-rl5n7\" (UID: \"11048ba4-fef7-4675-b3c3-2671af4a4e27\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rl5n7" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.446490 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/5a5b1574-7f76-4bcf-8c24-9ff7773e86f2-csi-data-dir\") pod \"csi-hostpathplugin-s4zph\" (UID: \"5a5b1574-7f76-4bcf-8c24-9ff7773e86f2\") " pod="hostpath-provisioner/csi-hostpathplugin-s4zph" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.446539 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/51547f65-61cc-4504-8399-5d540160fea1-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-682zv\" (UID: \"51547f65-61cc-4504-8399-5d540160fea1\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-682zv" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.446556 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ebf750fb-f47c-420b-931d-3ac0a23e9e7b-service-ca-bundle\") pod \"router-default-5444994796-ll9lb\" (UID: \"ebf750fb-f47c-420b-931d-3ac0a23e9e7b\") " pod="openshift-ingress/router-default-5444994796-ll9lb" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.446577 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/cce92b6a-69b0-49d5-bb66-36349ca96841-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-wz2dg\" (UID: \"cce92b6a-69b0-49d5-bb66-36349ca96841\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-wz2dg" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.446596 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7p9rv\" (UniqueName: \"kubernetes.io/projected/0593c91b-48f4-47aa-a57f-f9b50d7f385e-kube-api-access-7p9rv\") pod \"migrator-59844c95c7-67bjd\" (UID: \"0593c91b-48f4-47aa-a57f-f9b50d7f385e\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-67bjd" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.446622 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8zqxw\" (UniqueName: \"kubernetes.io/projected/2e5d7ee3-ea83-4445-a74d-9efa3864b3bb-kube-api-access-8zqxw\") pod \"service-ca-operator-777779d784-nrgbt\" (UID: \"2e5d7ee3-ea83-4445-a74d-9efa3864b3bb\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-nrgbt" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.446640 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cjb58\" (UniqueName: \"kubernetes.io/projected/830a0a27-fab1-4efc-aeb3-e38a962ffcfa-kube-api-access-cjb58\") pod \"dns-default-6xm7j\" (UID: \"830a0a27-fab1-4efc-aeb3-e38a962ffcfa\") " pod="openshift-dns/dns-default-6xm7j" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.446658 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/230b53bd-7c59-42f5-9604-0c5414fcdf80-cert\") pod \"ingress-canary-7g427\" (UID: \"230b53bd-7c59-42f5-9604-0c5414fcdf80\") " pod="openshift-ingress-canary/ingress-canary-7g427" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.446691 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/830a0a27-fab1-4efc-aeb3-e38a962ffcfa-metrics-tls\") pod \"dns-default-6xm7j\" (UID: \"830a0a27-fab1-4efc-aeb3-e38a962ffcfa\") " pod="openshift-dns/dns-default-6xm7j" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.446711 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rfjtr\" (UniqueName: \"kubernetes.io/projected/8797a357-3b55-4d63-9215-4c64bb676a3d-kube-api-access-rfjtr\") pod \"service-ca-9c57cc56f-chv7x\" (UID: \"8797a357-3b55-4d63-9215-4c64bb676a3d\") " pod="openshift-service-ca/service-ca-9c57cc56f-chv7x" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.446714 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e542cd22-7a8b-4a4d-8205-d3dc60b2b8c6-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-w6krf\" (UID: \"e542cd22-7a8b-4a4d-8205-d3dc60b2b8c6\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-w6krf" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.446739 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/8797a357-3b55-4d63-9215-4c64bb676a3d-signing-cabundle\") pod \"service-ca-9c57cc56f-chv7x\" (UID: \"8797a357-3b55-4d63-9215-4c64bb676a3d\") " pod="openshift-service-ca/service-ca-9c57cc56f-chv7x" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.446762 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/80ede041-caa4-451e-bdd1-083e2e03ec29-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-8qm2l\" (UID: \"80ede041-caa4-451e-bdd1-083e2e03ec29\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-8qm2l" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.446787 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/8797a357-3b55-4d63-9215-4c64bb676a3d-signing-key\") pod \"service-ca-9c57cc56f-chv7x\" (UID: \"8797a357-3b55-4d63-9215-4c64bb676a3d\") " pod="openshift-service-ca/service-ca-9c57cc56f-chv7x" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.446804 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/5a5b1574-7f76-4bcf-8c24-9ff7773e86f2-plugins-dir\") pod \"csi-hostpathplugin-s4zph\" (UID: \"5a5b1574-7f76-4bcf-8c24-9ff7773e86f2\") " pod="hostpath-provisioner/csi-hostpathplugin-s4zph" Oct 05 20:56:59 crc kubenswrapper[4754]: E1005 20:56:59.446826 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 20:56:59.946806715 +0000 UTC m=+143.850925425 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.446862 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/091efc31-55eb-4026-bf38-626e09a13253-webhook-cert\") pod \"packageserver-d55dfcdfc-nrd6b\" (UID: \"091efc31-55eb-4026-bf38-626e09a13253\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nrd6b" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.446895 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gxjxv\" (UniqueName: \"kubernetes.io/projected/0b4c539c-22cf-4337-9959-afcc73e64370-kube-api-access-gxjxv\") pod \"package-server-manager-789f6589d5-nrt9t\" (UID: \"0b4c539c-22cf-4337-9959-afcc73e64370\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-nrt9t" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.446919 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/c1c75259-4cf0-46b9-aea0-2c2ad2a6928c-node-bootstrap-token\") pod \"machine-config-server-bndqf\" (UID: \"c1c75259-4cf0-46b9-aea0-2c2ad2a6928c\") " pod="openshift-machine-config-operator/machine-config-server-bndqf" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.446938 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2e5d7ee3-ea83-4445-a74d-9efa3864b3bb-config\") pod \"service-ca-operator-777779d784-nrgbt\" (UID: \"2e5d7ee3-ea83-4445-a74d-9efa3864b3bb\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-nrgbt" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.446954 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/c1c4fc78-c305-4dba-b4d5-9f05d58a8ef6-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-ll5jq\" (UID: \"c1c4fc78-c305-4dba-b4d5-9f05d58a8ef6\") " pod="openshift-marketplace/marketplace-operator-79b997595-ll5jq" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.447002 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zqm88\" (UniqueName: \"kubernetes.io/projected/cce92b6a-69b0-49d5-bb66-36349ca96841-kube-api-access-zqm88\") pod \"machine-config-controller-84d6567774-wz2dg\" (UID: \"cce92b6a-69b0-49d5-bb66-36349ca96841\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-wz2dg" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.447025 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mv4pz\" (UniqueName: \"kubernetes.io/projected/60ff4c5c-9e45-47d3-b45d-a365e988342e-kube-api-access-mv4pz\") pod \"collect-profiles-29328285-mrswx\" (UID: \"60ff4c5c-9e45-47d3-b45d-a365e988342e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29328285-mrswx" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.447055 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e542cd22-7a8b-4a4d-8205-d3dc60b2b8c6-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-w6krf\" (UID: \"e542cd22-7a8b-4a4d-8205-d3dc60b2b8c6\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-w6krf" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.447078 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-czgsx\" (UniqueName: \"kubernetes.io/projected/ebf750fb-f47c-420b-931d-3ac0a23e9e7b-kube-api-access-czgsx\") pod \"router-default-5444994796-ll9lb\" (UID: \"ebf750fb-f47c-420b-931d-3ac0a23e9e7b\") " pod="openshift-ingress/router-default-5444994796-ll9lb" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.447107 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/0b4c539c-22cf-4337-9959-afcc73e64370-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-nrt9t\" (UID: \"0b4c539c-22cf-4337-9959-afcc73e64370\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-nrt9t" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.447128 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/d1b2b973-8b95-4f6a-83f1-468f84fb006d-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-5wx56\" (UID: \"d1b2b973-8b95-4f6a-83f1-468f84fb006d\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-5wx56" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.447153 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b96341ba-da55-442f-851e-5666717a1d5a-bound-sa-token\") pod \"ingress-operator-5b745b69d9-9vndv\" (UID: \"b96341ba-da55-442f-851e-5666717a1d5a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-9vndv" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.447189 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5sqj9\" (UID: \"c64d57c5-4bd4-48b5-85fc-9d613aea29c4\") " pod="openshift-image-registry/image-registry-697d97f7c8-5sqj9" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.447212 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/4da215ae-75ec-4ab1-81b0-39966d449b5f-images\") pod \"machine-config-operator-74547568cd-hjpzm\" (UID: \"4da215ae-75ec-4ab1-81b0-39966d449b5f\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-hjpzm" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.447237 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/80ede041-caa4-451e-bdd1-083e2e03ec29-config\") pod \"kube-apiserver-operator-766d6c64bb-8qm2l\" (UID: \"80ede041-caa4-451e-bdd1-083e2e03ec29\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-8qm2l" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.447258 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/c1c75259-4cf0-46b9-aea0-2c2ad2a6928c-certs\") pod \"machine-config-server-bndqf\" (UID: \"c1c75259-4cf0-46b9-aea0-2c2ad2a6928c\") " pod="openshift-machine-config-operator/machine-config-server-bndqf" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.447277 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/be8e4dfb-5b8a-4bf4-8d54-880d51f990de-srv-cert\") pod \"olm-operator-6b444d44fb-lv842\" (UID: \"be8e4dfb-5b8a-4bf4-8d54-880d51f990de\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-lv842" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.447303 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/ebf750fb-f47c-420b-931d-3ac0a23e9e7b-default-certificate\") pod \"router-default-5444994796-ll9lb\" (UID: \"ebf750fb-f47c-420b-931d-3ac0a23e9e7b\") " pod="openshift-ingress/router-default-5444994796-ll9lb" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.447325 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xjxn7\" (UniqueName: \"kubernetes.io/projected/091efc31-55eb-4026-bf38-626e09a13253-kube-api-access-xjxn7\") pod \"packageserver-d55dfcdfc-nrd6b\" (UID: \"091efc31-55eb-4026-bf38-626e09a13253\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nrd6b" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.447346 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e542cd22-7a8b-4a4d-8205-d3dc60b2b8c6-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-w6krf\" (UID: \"e542cd22-7a8b-4a4d-8205-d3dc60b2b8c6\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-w6krf" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.447365 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z9jr9\" (UniqueName: \"kubernetes.io/projected/c1c4fc78-c305-4dba-b4d5-9f05d58a8ef6-kube-api-access-z9jr9\") pod \"marketplace-operator-79b997595-ll5jq\" (UID: \"c1c4fc78-c305-4dba-b4d5-9f05d58a8ef6\") " pod="openshift-marketplace/marketplace-operator-79b997595-ll5jq" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.447385 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/4da215ae-75ec-4ab1-81b0-39966d449b5f-auth-proxy-config\") pod \"machine-config-operator-74547568cd-hjpzm\" (UID: \"4da215ae-75ec-4ab1-81b0-39966d449b5f\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-hjpzm" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.447405 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/cce92b6a-69b0-49d5-bb66-36349ca96841-proxy-tls\") pod \"machine-config-controller-84d6567774-wz2dg\" (UID: \"cce92b6a-69b0-49d5-bb66-36349ca96841\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-wz2dg" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.447429 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/60ff4c5c-9e45-47d3-b45d-a365e988342e-config-volume\") pod \"collect-profiles-29328285-mrswx\" (UID: \"60ff4c5c-9e45-47d3-b45d-a365e988342e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29328285-mrswx" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.447449 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9263c4f4-1540-4190-9e85-b1d8e2ca6c2f-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-6wnmn\" (UID: \"9263c4f4-1540-4190-9e85-b1d8e2ca6c2f\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6wnmn" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.447487 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x42wt\" (UniqueName: \"kubernetes.io/projected/230b53bd-7c59-42f5-9604-0c5414fcdf80-kube-api-access-x42wt\") pod \"ingress-canary-7g427\" (UID: \"230b53bd-7c59-42f5-9604-0c5414fcdf80\") " pod="openshift-ingress-canary/ingress-canary-7g427" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.447524 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/be8e4dfb-5b8a-4bf4-8d54-880d51f990de-profile-collector-cert\") pod \"olm-operator-6b444d44fb-lv842\" (UID: \"be8e4dfb-5b8a-4bf4-8d54-880d51f990de\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-lv842" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.447541 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/5a5b1574-7f76-4bcf-8c24-9ff7773e86f2-registration-dir\") pod \"csi-hostpathplugin-s4zph\" (UID: \"5a5b1574-7f76-4bcf-8c24-9ff7773e86f2\") " pod="hostpath-provisioner/csi-hostpathplugin-s4zph" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.447561 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/091efc31-55eb-4026-bf38-626e09a13253-tmpfs\") pod \"packageserver-d55dfcdfc-nrd6b\" (UID: \"091efc31-55eb-4026-bf38-626e09a13253\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nrd6b" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.447577 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c1c4fc78-c305-4dba-b4d5-9f05d58a8ef6-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-ll5jq\" (UID: \"c1c4fc78-c305-4dba-b4d5-9f05d58a8ef6\") " pod="openshift-marketplace/marketplace-operator-79b997595-ll5jq" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.447784 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/830a0a27-fab1-4efc-aeb3-e38a962ffcfa-config-volume\") pod \"dns-default-6xm7j\" (UID: \"830a0a27-fab1-4efc-aeb3-e38a962ffcfa\") " pod="openshift-dns/dns-default-6xm7j" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.448537 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c1c4fc78-c305-4dba-b4d5-9f05d58a8ef6-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-ll5jq\" (UID: \"c1c4fc78-c305-4dba-b4d5-9f05d58a8ef6\") " pod="openshift-marketplace/marketplace-operator-79b997595-ll5jq" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.448708 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/5a5b1574-7f76-4bcf-8c24-9ff7773e86f2-mountpoint-dir\") pod \"csi-hostpathplugin-s4zph\" (UID: \"5a5b1574-7f76-4bcf-8c24-9ff7773e86f2\") " pod="hostpath-provisioner/csi-hostpathplugin-s4zph" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.451644 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/cce92b6a-69b0-49d5-bb66-36349ca96841-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-wz2dg\" (UID: \"cce92b6a-69b0-49d5-bb66-36349ca96841\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-wz2dg" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.453685 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/4da215ae-75ec-4ab1-81b0-39966d449b5f-proxy-tls\") pod \"machine-config-operator-74547568cd-hjpzm\" (UID: \"4da215ae-75ec-4ab1-81b0-39966d449b5f\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-hjpzm" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.453815 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/80ede041-caa4-451e-bdd1-083e2e03ec29-config\") pod \"kube-apiserver-operator-766d6c64bb-8qm2l\" (UID: \"80ede041-caa4-451e-bdd1-083e2e03ec29\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-8qm2l" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.454432 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/5a5b1574-7f76-4bcf-8c24-9ff7773e86f2-socket-dir\") pod \"csi-hostpathplugin-s4zph\" (UID: \"5a5b1574-7f76-4bcf-8c24-9ff7773e86f2\") " pod="hostpath-provisioner/csi-hostpathplugin-s4zph" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.455835 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wtxlp\" (UniqueName: \"kubernetes.io/projected/b157fa21-ae67-4c98-806c-93854b072965-kube-api-access-wtxlp\") pod \"openshift-config-operator-7777fb866f-cwtlx\" (UID: \"b157fa21-ae67-4c98-806c-93854b072965\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-cwtlx" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.461913 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/8797a357-3b55-4d63-9215-4c64bb676a3d-signing-cabundle\") pod \"service-ca-9c57cc56f-chv7x\" (UID: \"8797a357-3b55-4d63-9215-4c64bb676a3d\") " pod="openshift-service-ca/service-ca-9c57cc56f-chv7x" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.463557 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/60ff4c5c-9e45-47d3-b45d-a365e988342e-secret-volume\") pod \"collect-profiles-29328285-mrswx\" (UID: \"60ff4c5c-9e45-47d3-b45d-a365e988342e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29328285-mrswx" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.447132 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/5a5b1574-7f76-4bcf-8c24-9ff7773e86f2-plugins-dir\") pod \"csi-hostpathplugin-s4zph\" (UID: \"5a5b1574-7f76-4bcf-8c24-9ff7773e86f2\") " pod="hostpath-provisioner/csi-hostpathplugin-s4zph" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.464538 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-pnhf9" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.465574 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/4da215ae-75ec-4ab1-81b0-39966d449b5f-auth-proxy-config\") pod \"machine-config-operator-74547568cd-hjpzm\" (UID: \"4da215ae-75ec-4ab1-81b0-39966d449b5f\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-hjpzm" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.466135 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9263c4f4-1540-4190-9e85-b1d8e2ca6c2f-config\") pod \"kube-controller-manager-operator-78b949d7b-6wnmn\" (UID: \"9263c4f4-1540-4190-9e85-b1d8e2ca6c2f\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6wnmn" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.466170 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/830a0a27-fab1-4efc-aeb3-e38a962ffcfa-metrics-tls\") pod \"dns-default-6xm7j\" (UID: \"830a0a27-fab1-4efc-aeb3-e38a962ffcfa\") " pod="openshift-dns/dns-default-6xm7j" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.467365 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/c2c0ebd0-5cd9-49f4-9cd6-df5995cfcd74-srv-cert\") pod \"catalog-operator-68c6474976-jpbfr\" (UID: \"c2c0ebd0-5cd9-49f4-9cd6-df5995cfcd74\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jpbfr" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.467564 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2e5d7ee3-ea83-4445-a74d-9efa3864b3bb-serving-cert\") pod \"service-ca-operator-777779d784-nrgbt\" (UID: \"2e5d7ee3-ea83-4445-a74d-9efa3864b3bb\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-nrgbt" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.467906 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2e5d7ee3-ea83-4445-a74d-9efa3864b3bb-config\") pod \"service-ca-operator-777779d784-nrgbt\" (UID: \"2e5d7ee3-ea83-4445-a74d-9efa3864b3bb\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-nrgbt" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.471585 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/c1c75259-4cf0-46b9-aea0-2c2ad2a6928c-node-bootstrap-token\") pod \"machine-config-server-bndqf\" (UID: \"c1c75259-4cf0-46b9-aea0-2c2ad2a6928c\") " pod="openshift-machine-config-operator/machine-config-server-bndqf" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.472265 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/091efc31-55eb-4026-bf38-626e09a13253-webhook-cert\") pod \"packageserver-d55dfcdfc-nrd6b\" (UID: \"091efc31-55eb-4026-bf38-626e09a13253\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nrd6b" Oct 05 20:56:59 crc kubenswrapper[4754]: E1005 20:56:59.472430 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 20:56:59.97241321 +0000 UTC m=+143.876531920 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5sqj9" (UID: "c64d57c5-4bd4-48b5-85fc-9d613aea29c4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.472691 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/091efc31-55eb-4026-bf38-626e09a13253-tmpfs\") pod \"packageserver-d55dfcdfc-nrd6b\" (UID: \"091efc31-55eb-4026-bf38-626e09a13253\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nrd6b" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.473964 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/4da215ae-75ec-4ab1-81b0-39966d449b5f-images\") pod \"machine-config-operator-74547568cd-hjpzm\" (UID: \"4da215ae-75ec-4ab1-81b0-39966d449b5f\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-hjpzm" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.475318 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ebf750fb-f47c-420b-931d-3ac0a23e9e7b-service-ca-bundle\") pod \"router-default-5444994796-ll9lb\" (UID: \"ebf750fb-f47c-420b-931d-3ac0a23e9e7b\") " pod="openshift-ingress/router-default-5444994796-ll9lb" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.477216 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/80ede041-caa4-451e-bdd1-083e2e03ec29-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-8qm2l\" (UID: \"80ede041-caa4-451e-bdd1-083e2e03ec29\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-8qm2l" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.479756 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/5a5b1574-7f76-4bcf-8c24-9ff7773e86f2-registration-dir\") pod \"csi-hostpathplugin-s4zph\" (UID: \"5a5b1574-7f76-4bcf-8c24-9ff7773e86f2\") " pod="hostpath-provisioner/csi-hostpathplugin-s4zph" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.481276 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/60ff4c5c-9e45-47d3-b45d-a365e988342e-config-volume\") pod \"collect-profiles-29328285-mrswx\" (UID: \"60ff4c5c-9e45-47d3-b45d-a365e988342e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29328285-mrswx" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.483602 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b96341ba-da55-442f-851e-5666717a1d5a-trusted-ca\") pod \"ingress-operator-5b745b69d9-9vndv\" (UID: \"b96341ba-da55-442f-851e-5666717a1d5a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-9vndv" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.483755 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/5a5b1574-7f76-4bcf-8c24-9ff7773e86f2-csi-data-dir\") pod \"csi-hostpathplugin-s4zph\" (UID: \"5a5b1574-7f76-4bcf-8c24-9ff7773e86f2\") " pod="hostpath-provisioner/csi-hostpathplugin-s4zph" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.484312 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/c2c0ebd0-5cd9-49f4-9cd6-df5995cfcd74-profile-collector-cert\") pod \"catalog-operator-68c6474976-jpbfr\" (UID: \"c2c0ebd0-5cd9-49f4-9cd6-df5995cfcd74\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jpbfr" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.484769 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/0b4c539c-22cf-4337-9959-afcc73e64370-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-nrt9t\" (UID: \"0b4c539c-22cf-4337-9959-afcc73e64370\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-nrt9t" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.485540 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11048ba4-fef7-4675-b3c3-2671af4a4e27-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-rl5n7\" (UID: \"11048ba4-fef7-4675-b3c3-2671af4a4e27\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rl5n7" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.486343 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e542cd22-7a8b-4a4d-8205-d3dc60b2b8c6-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-w6krf\" (UID: \"e542cd22-7a8b-4a4d-8205-d3dc60b2b8c6\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-w6krf" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.486605 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/ebf750fb-f47c-420b-931d-3ac0a23e9e7b-stats-auth\") pod \"router-default-5444994796-ll9lb\" (UID: \"ebf750fb-f47c-420b-931d-3ac0a23e9e7b\") " pod="openshift-ingress/router-default-5444994796-ll9lb" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.486675 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/11048ba4-fef7-4675-b3c3-2671af4a4e27-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-rl5n7\" (UID: \"11048ba4-fef7-4675-b3c3-2671af4a4e27\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rl5n7" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.486669 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ebf750fb-f47c-420b-931d-3ac0a23e9e7b-metrics-certs\") pod \"router-default-5444994796-ll9lb\" (UID: \"ebf750fb-f47c-420b-931d-3ac0a23e9e7b\") " pod="openshift-ingress/router-default-5444994796-ll9lb" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.491669 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9263c4f4-1540-4190-9e85-b1d8e2ca6c2f-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-6wnmn\" (UID: \"9263c4f4-1540-4190-9e85-b1d8e2ca6c2f\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6wnmn" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.493547 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/ebf750fb-f47c-420b-931d-3ac0a23e9e7b-default-certificate\") pod \"router-default-5444994796-ll9lb\" (UID: \"ebf750fb-f47c-420b-931d-3ac0a23e9e7b\") " pod="openshift-ingress/router-default-5444994796-ll9lb" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.495417 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/be8e4dfb-5b8a-4bf4-8d54-880d51f990de-srv-cert\") pod \"olm-operator-6b444d44fb-lv842\" (UID: \"be8e4dfb-5b8a-4bf4-8d54-880d51f990de\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-lv842" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.507275 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7pfxf\" (UniqueName: \"kubernetes.io/projected/dbe65326-0856-4b33-8894-4c9266bc1a90-kube-api-access-7pfxf\") pod \"etcd-operator-b45778765-q2jq4\" (UID: \"dbe65326-0856-4b33-8894-4c9266bc1a90\") " pod="openshift-etcd-operator/etcd-operator-b45778765-q2jq4" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.510303 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/be8e4dfb-5b8a-4bf4-8d54-880d51f990de-profile-collector-cert\") pod \"olm-operator-6b444d44fb-lv842\" (UID: \"be8e4dfb-5b8a-4bf4-8d54-880d51f990de\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-lv842" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.515068 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/091efc31-55eb-4026-bf38-626e09a13253-apiservice-cert\") pod \"packageserver-d55dfcdfc-nrd6b\" (UID: \"091efc31-55eb-4026-bf38-626e09a13253\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nrd6b" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.516538 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/b96341ba-da55-442f-851e-5666717a1d5a-metrics-tls\") pod \"ingress-operator-5b745b69d9-9vndv\" (UID: \"b96341ba-da55-442f-851e-5666717a1d5a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-9vndv" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.516962 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/230b53bd-7c59-42f5-9604-0c5414fcdf80-cert\") pod \"ingress-canary-7g427\" (UID: \"230b53bd-7c59-42f5-9604-0c5414fcdf80\") " pod="openshift-ingress-canary/ingress-canary-7g427" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.517112 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/c1c4fc78-c305-4dba-b4d5-9f05d58a8ef6-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-ll5jq\" (UID: \"c1c4fc78-c305-4dba-b4d5-9f05d58a8ef6\") " pod="openshift-marketplace/marketplace-operator-79b997595-ll5jq" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.517392 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/cce92b6a-69b0-49d5-bb66-36349ca96841-proxy-tls\") pod \"machine-config-controller-84d6567774-wz2dg\" (UID: \"cce92b6a-69b0-49d5-bb66-36349ca96841\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-wz2dg" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.517557 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/8797a357-3b55-4d63-9215-4c64bb676a3d-signing-key\") pod \"service-ca-9c57cc56f-chv7x\" (UID: \"8797a357-3b55-4d63-9215-4c64bb676a3d\") " pod="openshift-service-ca/service-ca-9c57cc56f-chv7x" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.517759 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/51547f65-61cc-4504-8399-5d540160fea1-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-682zv\" (UID: \"51547f65-61cc-4504-8399-5d540160fea1\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-682zv" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.520848 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dw4mx\" (UniqueName: \"kubernetes.io/projected/3934dd94-b9fa-4079-a268-60a9470dab41-kube-api-access-dw4mx\") pod \"downloads-7954f5f757-ksczt\" (UID: \"3934dd94-b9fa-4079-a268-60a9470dab41\") " pod="openshift-console/downloads-7954f5f757-ksczt" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.523236 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/d1b2b973-8b95-4f6a-83f1-468f84fb006d-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-5wx56\" (UID: \"d1b2b973-8b95-4f6a-83f1-468f84fb006d\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-5wx56" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.524934 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wtph5\" (UniqueName: \"kubernetes.io/projected/bc8bf6c1-363b-4e24-830c-898e4f7c6325-kube-api-access-wtph5\") pod \"dns-operator-744455d44c-2qmwc\" (UID: \"bc8bf6c1-363b-4e24-830c-898e4f7c6325\") " pod="openshift-dns-operator/dns-operator-744455d44c-2qmwc" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.529075 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/c1c75259-4cf0-46b9-aea0-2c2ad2a6928c-certs\") pod \"machine-config-server-bndqf\" (UID: \"c1c75259-4cf0-46b9-aea0-2c2ad2a6928c\") " pod="openshift-machine-config-operator/machine-config-server-bndqf" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.547938 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 20:56:59 crc kubenswrapper[4754]: E1005 20:56:59.548265 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 20:57:00.048212741 +0000 UTC m=+143.952331451 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.548624 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5sqj9\" (UID: \"c64d57c5-4bd4-48b5-85fc-9d613aea29c4\") " pod="openshift-image-registry/image-registry-697d97f7c8-5sqj9" Oct 05 20:56:59 crc kubenswrapper[4754]: E1005 20:56:59.551527 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 20:57:00.051504797 +0000 UTC m=+143.955623507 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5sqj9" (UID: "c64d57c5-4bd4-48b5-85fc-9d613aea29c4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.556855 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-5scsb"] Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.560329 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-gchnw" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.562476 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7qtr7\" (UniqueName: \"kubernetes.io/projected/c64d57c5-4bd4-48b5-85fc-9d613aea29c4-kube-api-access-7qtr7\") pod \"image-registry-697d97f7c8-5sqj9\" (UID: \"c64d57c5-4bd4-48b5-85fc-9d613aea29c4\") " pod="openshift-image-registry/image-registry-697d97f7c8-5sqj9" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.574192 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cjn98\" (UniqueName: \"kubernetes.io/projected/11048ba4-fef7-4675-b3c3-2671af4a4e27-kube-api-access-cjn98\") pod \"kube-storage-version-migrator-operator-b67b599dd-rl5n7\" (UID: \"11048ba4-fef7-4675-b3c3-2671af4a4e27\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rl5n7" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.578238 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-q2jq4" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.591726 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-2qmwc" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.597476 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wx9l2" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.603760 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jkkxq\" (UniqueName: \"kubernetes.io/projected/c1c75259-4cf0-46b9-aea0-2c2ad2a6928c-kube-api-access-jkkxq\") pod \"machine-config-server-bndqf\" (UID: \"c1c75259-4cf0-46b9-aea0-2c2ad2a6928c\") " pod="openshift-machine-config-operator/machine-config-server-bndqf" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.633787 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gb999\" (UniqueName: \"kubernetes.io/projected/be8e4dfb-5b8a-4bf4-8d54-880d51f990de-kube-api-access-gb999\") pod \"olm-operator-6b444d44fb-lv842\" (UID: \"be8e4dfb-5b8a-4bf4-8d54-880d51f990de\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-lv842" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.639136 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gxjxv\" (UniqueName: \"kubernetes.io/projected/0b4c539c-22cf-4337-9959-afcc73e64370-kube-api-access-gxjxv\") pod \"package-server-manager-789f6589d5-nrt9t\" (UID: \"0b4c539c-22cf-4337-9959-afcc73e64370\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-nrt9t" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.649895 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9nms8"] Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.652031 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.655122 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s6vzh\" (UniqueName: \"kubernetes.io/projected/4da215ae-75ec-4ab1-81b0-39966d449b5f-kube-api-access-s6vzh\") pod \"machine-config-operator-74547568cd-hjpzm\" (UID: \"4da215ae-75ec-4ab1-81b0-39966d449b5f\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-hjpzm" Oct 05 20:56:59 crc kubenswrapper[4754]: E1005 20:56:59.655237 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 20:57:00.154166426 +0000 UTC m=+144.058285136 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.655305 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5sqj9\" (UID: \"c64d57c5-4bd4-48b5-85fc-9d613aea29c4\") " pod="openshift-image-registry/image-registry-697d97f7c8-5sqj9" Oct 05 20:56:59 crc kubenswrapper[4754]: E1005 20:56:59.655847 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 20:57:00.155834831 +0000 UTC m=+144.059953541 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5sqj9" (UID: "c64d57c5-4bd4-48b5-85fc-9d613aea29c4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.661819 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rl5n7" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.675199 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bmkt9\" (UniqueName: \"kubernetes.io/projected/c2c0ebd0-5cd9-49f4-9cd6-df5995cfcd74-kube-api-access-bmkt9\") pod \"catalog-operator-68c6474976-jpbfr\" (UID: \"c2c0ebd0-5cd9-49f4-9cd6-df5995cfcd74\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jpbfr" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.696065 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p4pmn\" (UniqueName: \"kubernetes.io/projected/b96341ba-da55-442f-851e-5666717a1d5a-kube-api-access-p4pmn\") pod \"ingress-operator-5b745b69d9-9vndv\" (UID: \"b96341ba-da55-442f-851e-5666717a1d5a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-9vndv" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.708119 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lxp7v" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.720538 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jpbfr" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.723212 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-944m8\" (UniqueName: \"kubernetes.io/projected/51547f65-61cc-4504-8399-5d540160fea1-kube-api-access-944m8\") pod \"multus-admission-controller-857f4d67dd-682zv\" (UID: \"51547f65-61cc-4504-8399-5d540160fea1\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-682zv" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.738438 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-cwtlx" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.744293 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/80ede041-caa4-451e-bdd1-083e2e03ec29-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-8qm2l\" (UID: \"80ede041-caa4-451e-bdd1-083e2e03ec29\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-8qm2l" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.754535 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-nrt9t" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.755320 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xpdzs\" (UniqueName: \"kubernetes.io/projected/d1b2b973-8b95-4f6a-83f1-468f84fb006d-kube-api-access-xpdzs\") pod \"control-plane-machine-set-operator-78cbb6b69f-5wx56\" (UID: \"d1b2b973-8b95-4f6a-83f1-468f84fb006d\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-5wx56" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.756013 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 20:56:59 crc kubenswrapper[4754]: E1005 20:56:59.756355 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 20:57:00.256337753 +0000 UTC m=+144.160456463 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.778152 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-lv842" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.778373 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-ksczt" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.783131 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rfjtr\" (UniqueName: \"kubernetes.io/projected/8797a357-3b55-4d63-9215-4c64bb676a3d-kube-api-access-rfjtr\") pod \"service-ca-9c57cc56f-chv7x\" (UID: \"8797a357-3b55-4d63-9215-4c64bb676a3d\") " pod="openshift-service-ca/service-ca-9c57cc56f-chv7x" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.796033 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z9jr9\" (UniqueName: \"kubernetes.io/projected/c1c4fc78-c305-4dba-b4d5-9f05d58a8ef6-kube-api-access-z9jr9\") pod \"marketplace-operator-79b997595-ll5jq\" (UID: \"c1c4fc78-c305-4dba-b4d5-9f05d58a8ef6\") " pod="openshift-marketplace/marketplace-operator-79b997595-ll5jq" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.802865 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-bndqf" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.814053 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hm4ch" event={"ID":"3d4ca0c2-4e63-425f-bdb6-89da1a2e04c5","Type":"ContainerStarted","Data":"afcd40c9fdf147615947cb4c99f50d5d17f015b9b596e2d1ecbf7a427cae70f9"} Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.815155 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gv99q\" (UniqueName: \"kubernetes.io/projected/5a5b1574-7f76-4bcf-8c24-9ff7773e86f2-kube-api-access-gv99q\") pod \"csi-hostpathplugin-s4zph\" (UID: \"5a5b1574-7f76-4bcf-8c24-9ff7773e86f2\") " pod="hostpath-provisioner/csi-hostpathplugin-s4zph" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.833420 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9nms8" event={"ID":"9e7db3dd-7fae-469a-9ca5-eba1d965cbe8","Type":"ContainerStarted","Data":"27219ea09ea7d4d41548417ee91aed57e4acbf6838767912f74f670997bba4ee"} Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.834022 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9263c4f4-1540-4190-9e85-b1d8e2ca6c2f-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-6wnmn\" (UID: \"9263c4f4-1540-4190-9e85-b1d8e2ca6c2f\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6wnmn" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.834817 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-5scsb" event={"ID":"5a98537a-186b-4150-bd41-c2627ec772f9","Type":"ContainerStarted","Data":"4af61b422898067fd145863660204ba0d5833595f98b13da1921ea07f7d19601"} Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.843976 4754 generic.go:334] "Generic (PLEG): container finished" podID="20a5d042-ffe2-4244-b25d-700b26534a04" containerID="ee308f0ed1fc0a20f35f63bbe67f1cf169b4a6ad46d7fd9ae01a6997ee44d817" exitCode=0 Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.845019 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-22htd"] Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.845054 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-k8qff" event={"ID":"20a5d042-ffe2-4244-b25d-700b26534a04","Type":"ContainerDied","Data":"ee308f0ed1fc0a20f35f63bbe67f1cf169b4a6ad46d7fd9ae01a6997ee44d817"} Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.845072 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-k8qff" event={"ID":"20a5d042-ffe2-4244-b25d-700b26534a04","Type":"ContainerStarted","Data":"01d451612454e4be1a7c2c3c729bbaca4c7c441aead0c314183fdd48806793a0"} Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.845086 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-pnhf9"] Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.860915 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5sqj9\" (UID: \"c64d57c5-4bd4-48b5-85fc-9d613aea29c4\") " pod="openshift-image-registry/image-registry-697d97f7c8-5sqj9" Oct 05 20:56:59 crc kubenswrapper[4754]: E1005 20:56:59.862950 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 20:57:00.362933896 +0000 UTC m=+144.267052606 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5sqj9" (UID: "c64d57c5-4bd4-48b5-85fc-9d613aea29c4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.864059 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b96341ba-da55-442f-851e-5666717a1d5a-bound-sa-token\") pod \"ingress-operator-5b745b69d9-9vndv\" (UID: \"b96341ba-da55-442f-851e-5666717a1d5a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-9vndv" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.896753 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e542cd22-7a8b-4a4d-8205-d3dc60b2b8c6-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-w6krf\" (UID: \"e542cd22-7a8b-4a4d-8205-d3dc60b2b8c6\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-w6krf" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.905975 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-8qm2l" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.911966 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6wnmn" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.917001 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-rvsbj" event={"ID":"b1286ed6-4e15-46e0-b4c3-3b7e3cad6057","Type":"ContainerStarted","Data":"c237d2e1ac6335fad6f3391f43562c0780d719ecafeac3b0c81b33874c1d1da4"} Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.917042 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-rvsbj" event={"ID":"b1286ed6-4e15-46e0-b4c3-3b7e3cad6057","Type":"ContainerStarted","Data":"2fd8d2336ece00bcebac4481b957ad7bb2adf173e4f8f96087c3c4d72ea26934"} Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.917559 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-rvsbj" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.920659 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-w6krf" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.925640 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-5sffm" event={"ID":"c56f5eb8-3a12-428c-b347-78369dc19008","Type":"ContainerStarted","Data":"e89d93557492b507b25bbe982c756e8821d7f1efc3a6df2fb3a35171cd924f24"} Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.925716 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-5sffm" event={"ID":"c56f5eb8-3a12-428c-b347-78369dc19008","Type":"ContainerStarted","Data":"1b5f2b6c650425bb321b1679f0ac14848e2a2b5706e1443dbdeda08a6b35615f"} Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.929160 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-9vndv" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.933960 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-hjpzm" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.933958 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tcjmf" event={"ID":"afbf61d3-201e-4326-8463-4c660057a865","Type":"ContainerStarted","Data":"afc096de8c426d51034eacf4fabc640ea5c526bae4528f96b2ab2f7f5a575edd"} Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.935999 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-czgsx\" (UniqueName: \"kubernetes.io/projected/ebf750fb-f47c-420b-931d-3ac0a23e9e7b-kube-api-access-czgsx\") pod \"router-default-5444994796-ll9lb\" (UID: \"ebf750fb-f47c-420b-931d-3ac0a23e9e7b\") " pod="openshift-ingress/router-default-5444994796-ll9lb" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.940835 4754 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-rvsbj container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" start-of-body= Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.940922 4754 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-rvsbj" podUID="b1286ed6-4e15-46e0-b4c3-3b7e3cad6057" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.944638 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zqm88\" (UniqueName: \"kubernetes.io/projected/cce92b6a-69b0-49d5-bb66-36349ca96841-kube-api-access-zqm88\") pod \"machine-config-controller-84d6567774-wz2dg\" (UID: \"cce92b6a-69b0-49d5-bb66-36349ca96841\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-wz2dg" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.965995 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-ll5jq" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.966595 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 20:56:59 crc kubenswrapper[4754]: E1005 20:56:59.966719 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 20:57:00.466687854 +0000 UTC m=+144.370806564 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.966965 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5sqj9\" (UID: \"c64d57c5-4bd4-48b5-85fc-9d613aea29c4\") " pod="openshift-image-registry/image-registry-697d97f7c8-5sqj9" Oct 05 20:56:59 crc kubenswrapper[4754]: E1005 20:56:59.971741 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 20:57:00.471152032 +0000 UTC m=+144.375270742 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5sqj9" (UID: "c64d57c5-4bd4-48b5-85fc-9d613aea29c4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.978070 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-gchnw"] Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.978524 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-682zv" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.983323 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-ll9lb" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.988125 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8zqxw\" (UniqueName: \"kubernetes.io/projected/2e5d7ee3-ea83-4445-a74d-9efa3864b3bb-kube-api-access-8zqxw\") pod \"service-ca-operator-777779d784-nrgbt\" (UID: \"2e5d7ee3-ea83-4445-a74d-9efa3864b3bb\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-nrgbt" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.991446 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cjb58\" (UniqueName: \"kubernetes.io/projected/830a0a27-fab1-4efc-aeb3-e38a962ffcfa-kube-api-access-cjb58\") pod \"dns-default-6xm7j\" (UID: \"830a0a27-fab1-4efc-aeb3-e38a962ffcfa\") " pod="openshift-dns/dns-default-6xm7j" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.996179 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mv4pz\" (UniqueName: \"kubernetes.io/projected/60ff4c5c-9e45-47d3-b45d-a365e988342e-kube-api-access-mv4pz\") pod \"collect-profiles-29328285-mrswx\" (UID: \"60ff4c5c-9e45-47d3-b45d-a365e988342e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29328285-mrswx" Oct 05 20:56:59 crc kubenswrapper[4754]: I1005 20:56:59.997378 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7p9rv\" (UniqueName: \"kubernetes.io/projected/0593c91b-48f4-47aa-a57f-f9b50d7f385e-kube-api-access-7p9rv\") pod \"migrator-59844c95c7-67bjd\" (UID: \"0593c91b-48f4-47aa-a57f-f9b50d7f385e\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-67bjd" Oct 05 20:57:00 crc kubenswrapper[4754]: I1005 20:56:59.999615 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-5wx56" Oct 05 20:57:00 crc kubenswrapper[4754]: I1005 20:57:00.019178 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xjxn7\" (UniqueName: \"kubernetes.io/projected/091efc31-55eb-4026-bf38-626e09a13253-kube-api-access-xjxn7\") pod \"packageserver-d55dfcdfc-nrd6b\" (UID: \"091efc31-55eb-4026-bf38-626e09a13253\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nrd6b" Oct 05 20:57:00 crc kubenswrapper[4754]: I1005 20:57:00.032701 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-67bjd" Oct 05 20:57:00 crc kubenswrapper[4754]: I1005 20:57:00.036255 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29328285-mrswx" Oct 05 20:57:00 crc kubenswrapper[4754]: I1005 20:57:00.044841 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-nrgbt" Oct 05 20:57:00 crc kubenswrapper[4754]: I1005 20:57:00.061056 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x42wt\" (UniqueName: \"kubernetes.io/projected/230b53bd-7c59-42f5-9604-0c5414fcdf80-kube-api-access-x42wt\") pod \"ingress-canary-7g427\" (UID: \"230b53bd-7c59-42f5-9604-0c5414fcdf80\") " pod="openshift-ingress-canary/ingress-canary-7g427" Oct 05 20:57:00 crc kubenswrapper[4754]: I1005 20:57:00.062012 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-chv7x" Oct 05 20:57:00 crc kubenswrapper[4754]: I1005 20:57:00.070527 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 20:57:00 crc kubenswrapper[4754]: E1005 20:57:00.072330 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 20:57:00.572309702 +0000 UTC m=+144.476428412 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:57:00 crc kubenswrapper[4754]: I1005 20:57:00.092910 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-s4zph" Oct 05 20:57:00 crc kubenswrapper[4754]: I1005 20:57:00.111075 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-6xm7j" Oct 05 20:57:00 crc kubenswrapper[4754]: I1005 20:57:00.120813 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-7g427" Oct 05 20:57:00 crc kubenswrapper[4754]: I1005 20:57:00.181166 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5sqj9\" (UID: \"c64d57c5-4bd4-48b5-85fc-9d613aea29c4\") " pod="openshift-image-registry/image-registry-697d97f7c8-5sqj9" Oct 05 20:57:00 crc kubenswrapper[4754]: E1005 20:57:00.183262 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 20:57:00.683241349 +0000 UTC m=+144.587360059 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5sqj9" (UID: "c64d57c5-4bd4-48b5-85fc-9d613aea29c4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:57:00 crc kubenswrapper[4754]: I1005 20:57:00.203852 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-9chlx" Oct 05 20:57:00 crc kubenswrapper[4754]: I1005 20:57:00.241241 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-wz2dg" Oct 05 20:57:00 crc kubenswrapper[4754]: I1005 20:57:00.289633 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 20:57:00 crc kubenswrapper[4754]: E1005 20:57:00.290619 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 20:57:00.790594552 +0000 UTC m=+144.694713262 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:57:00 crc kubenswrapper[4754]: I1005 20:57:00.314380 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nrd6b" Oct 05 20:57:00 crc kubenswrapper[4754]: I1005 20:57:00.391899 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5sqj9\" (UID: \"c64d57c5-4bd4-48b5-85fc-9d613aea29c4\") " pod="openshift-image-registry/image-registry-697d97f7c8-5sqj9" Oct 05 20:57:00 crc kubenswrapper[4754]: E1005 20:57:00.392963 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 20:57:00.892852041 +0000 UTC m=+144.796970751 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5sqj9" (UID: "c64d57c5-4bd4-48b5-85fc-9d613aea29c4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:57:00 crc kubenswrapper[4754]: I1005 20:57:00.394701 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-q2jq4"] Oct 05 20:57:00 crc kubenswrapper[4754]: I1005 20:57:00.432561 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-2qmwc"] Oct 05 20:57:00 crc kubenswrapper[4754]: I1005 20:57:00.461366 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rl5n7"] Oct 05 20:57:00 crc kubenswrapper[4754]: I1005 20:57:00.461419 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-9chlx" podStartSLOduration=122.46140039 podStartE2EDuration="2m2.46140039s" podCreationTimestamp="2025-10-05 20:54:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 20:57:00.458093852 +0000 UTC m=+144.362212562" watchObservedRunningTime="2025-10-05 20:57:00.46140039 +0000 UTC m=+144.365519100" Oct 05 20:57:00 crc kubenswrapper[4754]: I1005 20:57:00.493167 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 20:57:00 crc kubenswrapper[4754]: E1005 20:57:00.493610 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 20:57:00.993581619 +0000 UTC m=+144.897700329 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:57:00 crc kubenswrapper[4754]: I1005 20:57:00.540552 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wx9l2"] Oct 05 20:57:00 crc kubenswrapper[4754]: I1005 20:57:00.561078 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-lxp7v"] Oct 05 20:57:00 crc kubenswrapper[4754]: I1005 20:57:00.595161 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5sqj9\" (UID: \"c64d57c5-4bd4-48b5-85fc-9d613aea29c4\") " pod="openshift-image-registry/image-registry-697d97f7c8-5sqj9" Oct 05 20:57:00 crc kubenswrapper[4754]: E1005 20:57:00.595693 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 20:57:01.095678543 +0000 UTC m=+144.999797253 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5sqj9" (UID: "c64d57c5-4bd4-48b5-85fc-9d613aea29c4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:57:00 crc kubenswrapper[4754]: I1005 20:57:00.662127 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tcjmf" podStartSLOduration=123.662102646 podStartE2EDuration="2m3.662102646s" podCreationTimestamp="2025-10-05 20:54:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 20:57:00.658972053 +0000 UTC m=+144.563090783" watchObservedRunningTime="2025-10-05 20:57:00.662102646 +0000 UTC m=+144.566221356" Oct 05 20:57:00 crc kubenswrapper[4754]: I1005 20:57:00.662711 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-cwtlx"] Oct 05 20:57:00 crc kubenswrapper[4754]: I1005 20:57:00.698985 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 20:57:00 crc kubenswrapper[4754]: E1005 20:57:00.699398 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 20:57:01.19937375 +0000 UTC m=+145.103492460 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:57:00 crc kubenswrapper[4754]: I1005 20:57:00.800396 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5sqj9\" (UID: \"c64d57c5-4bd4-48b5-85fc-9d613aea29c4\") " pod="openshift-image-registry/image-registry-697d97f7c8-5sqj9" Oct 05 20:57:00 crc kubenswrapper[4754]: E1005 20:57:00.800946 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 20:57:01.300919359 +0000 UTC m=+145.205038069 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5sqj9" (UID: "c64d57c5-4bd4-48b5-85fc-9d613aea29c4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:57:00 crc kubenswrapper[4754]: I1005 20:57:00.905807 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 20:57:00 crc kubenswrapper[4754]: E1005 20:57:00.906358 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 20:57:01.40634077 +0000 UTC m=+145.310459480 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:57:01 crc kubenswrapper[4754]: I1005 20:57:01.007324 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5sqj9\" (UID: \"c64d57c5-4bd4-48b5-85fc-9d613aea29c4\") " pod="openshift-image-registry/image-registry-697d97f7c8-5sqj9" Oct 05 20:57:01 crc kubenswrapper[4754]: E1005 20:57:01.008471 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 20:57:01.508453235 +0000 UTC m=+145.412571935 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5sqj9" (UID: "c64d57c5-4bd4-48b5-85fc-9d613aea29c4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:57:01 crc kubenswrapper[4754]: I1005 20:57:01.010003 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rl5n7" event={"ID":"11048ba4-fef7-4675-b3c3-2671af4a4e27","Type":"ContainerStarted","Data":"67645284694f58232976dc208a233cedf19eff78f00c4381ce4ced98b1909b46"} Oct 05 20:57:01 crc kubenswrapper[4754]: I1005 20:57:01.129176 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 20:57:01 crc kubenswrapper[4754]: E1005 20:57:01.129745 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 20:57:01.629727245 +0000 UTC m=+145.533845955 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:57:01 crc kubenswrapper[4754]: I1005 20:57:01.171979 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-pnhf9" event={"ID":"1e16d79e-eb1e-4b8b-8f0a-acb370163d17","Type":"ContainerStarted","Data":"8d09260517c2f421547a20e3a6a926190586ddec9333c715783763a4f76d958f"} Oct 05 20:57:01 crc kubenswrapper[4754]: I1005 20:57:01.192991 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lxp7v" event={"ID":"d3c66e2a-8404-498c-b8a9-2bd2e6fd318b","Type":"ContainerStarted","Data":"7fc72b15c4129a95cbc9e04b3531915d1694d66fbebca3770b283e89af939b0f"} Oct 05 20:57:01 crc kubenswrapper[4754]: I1005 20:57:01.200270 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-gchnw" event={"ID":"6c0dfaff-08a7-4152-8624-1973e4a4d148","Type":"ContainerStarted","Data":"7fbaca9acf75fd1b626ab41f3763fae48be5ed525f7498ed5c9fe6eb230a2537"} Oct 05 20:57:01 crc kubenswrapper[4754]: I1005 20:57:01.216012 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-5scsb" event={"ID":"5a98537a-186b-4150-bd41-c2627ec772f9","Type":"ContainerStarted","Data":"46594d9f598789a9d3ecd6ea6a4f6a85d52094a981ea3719f0f30658e2af3563"} Oct 05 20:57:01 crc kubenswrapper[4754]: I1005 20:57:01.230239 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5sqj9\" (UID: \"c64d57c5-4bd4-48b5-85fc-9d613aea29c4\") " pod="openshift-image-registry/image-registry-697d97f7c8-5sqj9" Oct 05 20:57:01 crc kubenswrapper[4754]: I1005 20:57:01.232369 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-ll9lb" event={"ID":"ebf750fb-f47c-420b-931d-3ac0a23e9e7b","Type":"ContainerStarted","Data":"c7fcf48bde3d7a370c8ce94e6849891b15303060333fd4651e33b6b5ae33d108"} Oct 05 20:57:01 crc kubenswrapper[4754]: E1005 20:57:01.255998 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 20:57:01.755970427 +0000 UTC m=+145.660089137 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5sqj9" (UID: "c64d57c5-4bd4-48b5-85fc-9d613aea29c4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:57:01 crc kubenswrapper[4754]: I1005 20:57:01.257043 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-bndqf" event={"ID":"c1c75259-4cf0-46b9-aea0-2c2ad2a6928c","Type":"ContainerStarted","Data":"6d9301ef13864b7e74af18ff5b2941af5cf9d558ff6a331aaa0983776c872fe6"} Oct 05 20:57:01 crc kubenswrapper[4754]: I1005 20:57:01.327445 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9nms8" event={"ID":"9e7db3dd-7fae-469a-9ca5-eba1d965cbe8","Type":"ContainerStarted","Data":"e6cd95b8371dbb33d65f724cac3b124fb828bde36d3713b97fd1a0bf3b97b60d"} Oct 05 20:57:01 crc kubenswrapper[4754]: I1005 20:57:01.341341 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 20:57:01 crc kubenswrapper[4754]: E1005 20:57:01.343048 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 20:57:01.843031024 +0000 UTC m=+145.747149734 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:57:01 crc kubenswrapper[4754]: I1005 20:57:01.365002 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-22htd" event={"ID":"811694fe-dcfd-4fe3-876c-964d09321d2f","Type":"ContainerStarted","Data":"1c0b18d114a66ff0735b56fa93d149fde1a9b3a0567fdf6fed980d9725ed0fe5"} Oct 05 20:57:01 crc kubenswrapper[4754]: I1005 20:57:01.365067 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-22htd" event={"ID":"811694fe-dcfd-4fe3-876c-964d09321d2f","Type":"ContainerStarted","Data":"700803b1f3e935882833db087f56a7f6230fd860c72450dc4db3fbaf2e21085b"} Oct 05 20:57:01 crc kubenswrapper[4754]: I1005 20:57:01.373700 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-2qmwc" event={"ID":"bc8bf6c1-363b-4e24-830c-898e4f7c6325","Type":"ContainerStarted","Data":"583e69f88edfa76a7c8539f6af865f380cd7926edcf17ef39379cac36c5549f7"} Oct 05 20:57:01 crc kubenswrapper[4754]: I1005 20:57:01.459539 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5sqj9\" (UID: \"c64d57c5-4bd4-48b5-85fc-9d613aea29c4\") " pod="openshift-image-registry/image-registry-697d97f7c8-5sqj9" Oct 05 20:57:01 crc kubenswrapper[4754]: E1005 20:57:01.460243 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 20:57:01.960228327 +0000 UTC m=+145.864347037 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5sqj9" (UID: "c64d57c5-4bd4-48b5-85fc-9d613aea29c4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:57:01 crc kubenswrapper[4754]: I1005 20:57:01.470609 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-q2jq4" event={"ID":"dbe65326-0856-4b33-8894-4c9266bc1a90","Type":"ContainerStarted","Data":"077805aa99637612c9f5df0f61c15a80e5950ca9c50227688ca81980b1c6e42c"} Oct 05 20:57:01 crc kubenswrapper[4754]: I1005 20:57:01.503699 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6wnmn"] Oct 05 20:57:01 crc kubenswrapper[4754]: I1005 20:57:01.507620 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-cwtlx" event={"ID":"b157fa21-ae67-4c98-806c-93854b072965","Type":"ContainerStarted","Data":"979abd67db27e9df41b1ecda51ee8ce786d498e6f69e5d4a2077c7134904b8cc"} Oct 05 20:57:01 crc kubenswrapper[4754]: I1005 20:57:01.531887 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wx9l2" event={"ID":"3408670c-55e3-4bf0-9589-fef6200c3490","Type":"ContainerStarted","Data":"060bb86e6c7f57f140b2145dc8e01c0d1b31848bcf8fa0d846b5a357c6545f67"} Oct 05 20:57:01 crc kubenswrapper[4754]: I1005 20:57:01.542651 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-rvsbj" Oct 05 20:57:01 crc kubenswrapper[4754]: I1005 20:57:01.564256 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 20:57:01 crc kubenswrapper[4754]: E1005 20:57:01.566303 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 20:57:02.066274556 +0000 UTC m=+145.970393266 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:57:01 crc kubenswrapper[4754]: I1005 20:57:01.596482 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-p5wqw" podStartSLOduration=123.596457622 podStartE2EDuration="2m3.596457622s" podCreationTimestamp="2025-10-05 20:54:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 20:57:01.595350623 +0000 UTC m=+145.499469333" watchObservedRunningTime="2025-10-05 20:57:01.596457622 +0000 UTC m=+145.500576332" Oct 05 20:57:01 crc kubenswrapper[4754]: I1005 20:57:01.597320 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-rvsbj" podStartSLOduration=123.597313675 podStartE2EDuration="2m3.597313675s" podCreationTimestamp="2025-10-05 20:54:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 20:57:01.557176586 +0000 UTC m=+145.461295296" watchObservedRunningTime="2025-10-05 20:57:01.597313675 +0000 UTC m=+145.501432385" Oct 05 20:57:01 crc kubenswrapper[4754]: I1005 20:57:01.684929 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5sqj9\" (UID: \"c64d57c5-4bd4-48b5-85fc-9d613aea29c4\") " pod="openshift-image-registry/image-registry-697d97f7c8-5sqj9" Oct 05 20:57:01 crc kubenswrapper[4754]: E1005 20:57:01.685433 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 20:57:02.18541857 +0000 UTC m=+146.089537280 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5sqj9" (UID: "c64d57c5-4bd4-48b5-85fc-9d613aea29c4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:57:01 crc kubenswrapper[4754]: I1005 20:57:01.691072 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hm4ch" podStartSLOduration=124.691038168 podStartE2EDuration="2m4.691038168s" podCreationTimestamp="2025-10-05 20:54:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 20:57:01.63731178 +0000 UTC m=+145.541430490" watchObservedRunningTime="2025-10-05 20:57:01.691038168 +0000 UTC m=+145.595156878" Oct 05 20:57:01 crc kubenswrapper[4754]: I1005 20:57:01.691198 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-5sffm" podStartSLOduration=123.691193892 podStartE2EDuration="2m3.691193892s" podCreationTimestamp="2025-10-05 20:54:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 20:57:01.687474264 +0000 UTC m=+145.591592974" watchObservedRunningTime="2025-10-05 20:57:01.691193892 +0000 UTC m=+145.595312602" Oct 05 20:57:01 crc kubenswrapper[4754]: I1005 20:57:01.738259 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-nrt9t"] Oct 05 20:57:01 crc kubenswrapper[4754]: I1005 20:57:01.748767 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-ksczt"] Oct 05 20:57:01 crc kubenswrapper[4754]: I1005 20:57:01.788944 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 20:57:01 crc kubenswrapper[4754]: E1005 20:57:01.789406 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 20:57:02.289383743 +0000 UTC m=+146.193502463 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:57:01 crc kubenswrapper[4754]: I1005 20:57:01.850191 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jpbfr"] Oct 05 20:57:01 crc kubenswrapper[4754]: I1005 20:57:01.852603 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-lv842"] Oct 05 20:57:01 crc kubenswrapper[4754]: I1005 20:57:01.891447 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5sqj9\" (UID: \"c64d57c5-4bd4-48b5-85fc-9d613aea29c4\") " pod="openshift-image-registry/image-registry-697d97f7c8-5sqj9" Oct 05 20:57:01 crc kubenswrapper[4754]: E1005 20:57:01.891810 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 20:57:02.391796856 +0000 UTC m=+146.295915566 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5sqj9" (UID: "c64d57c5-4bd4-48b5-85fc-9d613aea29c4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:57:01 crc kubenswrapper[4754]: I1005 20:57:01.949938 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-9vndv"] Oct 05 20:57:01 crc kubenswrapper[4754]: I1005 20:57:01.999088 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 20:57:02 crc kubenswrapper[4754]: E1005 20:57:02.000076 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 20:57:02.499885888 +0000 UTC m=+146.404004598 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:57:02 crc kubenswrapper[4754]: I1005 20:57:02.028628 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-hjpzm"] Oct 05 20:57:02 crc kubenswrapper[4754]: I1005 20:57:02.059632 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-8qm2l"] Oct 05 20:57:02 crc kubenswrapper[4754]: I1005 20:57:02.102978 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5sqj9\" (UID: \"c64d57c5-4bd4-48b5-85fc-9d613aea29c4\") " pod="openshift-image-registry/image-registry-697d97f7c8-5sqj9" Oct 05 20:57:02 crc kubenswrapper[4754]: E1005 20:57:02.104029 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 20:57:02.604012846 +0000 UTC m=+146.508131556 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5sqj9" (UID: "c64d57c5-4bd4-48b5-85fc-9d613aea29c4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:57:02 crc kubenswrapper[4754]: I1005 20:57:02.172282 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29328285-mrswx"] Oct 05 20:57:02 crc kubenswrapper[4754]: I1005 20:57:02.174452 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wx9l2" podStartSLOduration=124.174432074 podStartE2EDuration="2m4.174432074s" podCreationTimestamp="2025-10-05 20:54:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 20:57:02.00069888 +0000 UTC m=+145.904817600" watchObservedRunningTime="2025-10-05 20:57:02.174432074 +0000 UTC m=+146.078550774" Oct 05 20:57:02 crc kubenswrapper[4754]: I1005 20:57:02.205663 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 20:57:02 crc kubenswrapper[4754]: E1005 20:57:02.206239 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 20:57:02.706218363 +0000 UTC m=+146.610337063 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:57:02 crc kubenswrapper[4754]: I1005 20:57:02.219011 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-22htd" podStartSLOduration=125.21899329 podStartE2EDuration="2m5.21899329s" podCreationTimestamp="2025-10-05 20:54:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 20:57:02.064108773 +0000 UTC m=+145.968227473" watchObservedRunningTime="2025-10-05 20:57:02.21899329 +0000 UTC m=+146.123112000" Oct 05 20:57:02 crc kubenswrapper[4754]: I1005 20:57:02.226364 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nrd6b"] Oct 05 20:57:02 crc kubenswrapper[4754]: I1005 20:57:02.226465 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9nms8" podStartSLOduration=124.226449157 podStartE2EDuration="2m4.226449157s" podCreationTimestamp="2025-10-05 20:54:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 20:57:02.144434053 +0000 UTC m=+146.048552753" watchObservedRunningTime="2025-10-05 20:57:02.226449157 +0000 UTC m=+146.130567867" Oct 05 20:57:02 crc kubenswrapper[4754]: I1005 20:57:02.285839 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-67bjd"] Oct 05 20:57:02 crc kubenswrapper[4754]: I1005 20:57:02.307591 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5sqj9\" (UID: \"c64d57c5-4bd4-48b5-85fc-9d613aea29c4\") " pod="openshift-image-registry/image-registry-697d97f7c8-5sqj9" Oct 05 20:57:02 crc kubenswrapper[4754]: E1005 20:57:02.308029 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 20:57:02.80800994 +0000 UTC m=+146.712128650 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5sqj9" (UID: "c64d57c5-4bd4-48b5-85fc-9d613aea29c4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:57:02 crc kubenswrapper[4754]: W1005 20:57:02.348553 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4da215ae_75ec_4ab1_81b0_39966d449b5f.slice/crio-0b0746dd46765e6feb34965fed817481e05bd0fe7166d6068cbd4fe8b0761d52 WatchSource:0}: Error finding container 0b0746dd46765e6feb34965fed817481e05bd0fe7166d6068cbd4fe8b0761d52: Status 404 returned error can't find the container with id 0b0746dd46765e6feb34965fed817481e05bd0fe7166d6068cbd4fe8b0761d52 Oct 05 20:57:02 crc kubenswrapper[4754]: I1005 20:57:02.411115 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 20:57:02 crc kubenswrapper[4754]: E1005 20:57:02.411539 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 20:57:02.911519611 +0000 UTC m=+146.815638321 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:57:02 crc kubenswrapper[4754]: I1005 20:57:02.517031 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5sqj9\" (UID: \"c64d57c5-4bd4-48b5-85fc-9d613aea29c4\") " pod="openshift-image-registry/image-registry-697d97f7c8-5sqj9" Oct 05 20:57:02 crc kubenswrapper[4754]: E1005 20:57:02.517780 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 20:57:03.017765195 +0000 UTC m=+146.921883905 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5sqj9" (UID: "c64d57c5-4bd4-48b5-85fc-9d613aea29c4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:57:02 crc kubenswrapper[4754]: I1005 20:57:02.618146 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 20:57:02 crc kubenswrapper[4754]: E1005 20:57:02.618492 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 20:57:03.118475163 +0000 UTC m=+147.022593873 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:57:02 crc kubenswrapper[4754]: I1005 20:57:02.693488 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-w6krf"] Oct 05 20:57:02 crc kubenswrapper[4754]: I1005 20:57:02.694014 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-k8qff" event={"ID":"20a5d042-ffe2-4244-b25d-700b26534a04","Type":"ContainerStarted","Data":"ce581db761198d4ef5b9d00173e44d3c0e3009f3b0c46888a3103b7b394ed715"} Oct 05 20:57:02 crc kubenswrapper[4754]: I1005 20:57:02.713606 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-682zv"] Oct 05 20:57:02 crc kubenswrapper[4754]: I1005 20:57:02.720549 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5sqj9\" (UID: \"c64d57c5-4bd4-48b5-85fc-9d613aea29c4\") " pod="openshift-image-registry/image-registry-697d97f7c8-5sqj9" Oct 05 20:57:02 crc kubenswrapper[4754]: E1005 20:57:02.721183 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 20:57:03.221165592 +0000 UTC m=+147.125284302 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5sqj9" (UID: "c64d57c5-4bd4-48b5-85fc-9d613aea29c4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:57:02 crc kubenswrapper[4754]: I1005 20:57:02.727790 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-chv7x"] Oct 05 20:57:02 crc kubenswrapper[4754]: I1005 20:57:02.728187 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-6xm7j"] Oct 05 20:57:02 crc kubenswrapper[4754]: I1005 20:57:02.776097 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-bndqf" event={"ID":"c1c75259-4cf0-46b9-aea0-2c2ad2a6928c","Type":"ContainerStarted","Data":"d14cbf0d5a519ec9d0af4261c9a8c2ce474aa8aa253b634b814fc8d889c6c994"} Oct 05 20:57:02 crc kubenswrapper[4754]: I1005 20:57:02.802371 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-7g427"] Oct 05 20:57:02 crc kubenswrapper[4754]: I1005 20:57:02.835176 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-ll5jq"] Oct 05 20:57:02 crc kubenswrapper[4754]: I1005 20:57:02.838111 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 20:57:02 crc kubenswrapper[4754]: E1005 20:57:02.839721 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 20:57:03.33969078 +0000 UTC m=+147.243809490 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:57:02 crc kubenswrapper[4754]: W1005 20:57:02.876602 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode542cd22_7a8b_4a4d_8205_d3dc60b2b8c6.slice/crio-c878dfd75be1b179225fead5d922ee8e3abe2be012d4573d79c5dd6f3c8a1181 WatchSource:0}: Error finding container c878dfd75be1b179225fead5d922ee8e3abe2be012d4573d79c5dd6f3c8a1181: Status 404 returned error can't find the container with id c878dfd75be1b179225fead5d922ee8e3abe2be012d4573d79c5dd6f3c8a1181 Oct 05 20:57:02 crc kubenswrapper[4754]: I1005 20:57:02.909001 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-bndqf" podStartSLOduration=6.908979039 podStartE2EDuration="6.908979039s" podCreationTimestamp="2025-10-05 20:56:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 20:57:02.884039131 +0000 UTC m=+146.788157841" watchObservedRunningTime="2025-10-05 20:57:02.908979039 +0000 UTC m=+146.813097749" Oct 05 20:57:02 crc kubenswrapper[4754]: W1005 20:57:02.936799 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8797a357_3b55_4d63_9215_4c64bb676a3d.slice/crio-d981b2bdcbeb275a7ae60ca2eaaf8ff7f9a3fb8e88ee1ac383c1685b75e77158 WatchSource:0}: Error finding container d981b2bdcbeb275a7ae60ca2eaaf8ff7f9a3fb8e88ee1ac383c1685b75e77158: Status 404 returned error can't find the container with id d981b2bdcbeb275a7ae60ca2eaaf8ff7f9a3fb8e88ee1ac383c1685b75e77158 Oct 05 20:57:02 crc kubenswrapper[4754]: I1005 20:57:02.941231 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5sqj9\" (UID: \"c64d57c5-4bd4-48b5-85fc-9d613aea29c4\") " pod="openshift-image-registry/image-registry-697d97f7c8-5sqj9" Oct 05 20:57:02 crc kubenswrapper[4754]: W1005 20:57:02.941513 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod230b53bd_7c59_42f5_9604_0c5414fcdf80.slice/crio-cc46eb3d5a3478d549363b81ddae95df9caaae64465ea3982d633ecf5995ed04 WatchSource:0}: Error finding container cc46eb3d5a3478d549363b81ddae95df9caaae64465ea3982d633ecf5995ed04: Status 404 returned error can't find the container with id cc46eb3d5a3478d549363b81ddae95df9caaae64465ea3982d633ecf5995ed04 Oct 05 20:57:02 crc kubenswrapper[4754]: E1005 20:57:02.946704 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 20:57:03.446686674 +0000 UTC m=+147.350805384 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5sqj9" (UID: "c64d57c5-4bd4-48b5-85fc-9d613aea29c4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:57:03 crc kubenswrapper[4754]: I1005 20:57:03.012678 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-wz2dg"] Oct 05 20:57:03 crc kubenswrapper[4754]: I1005 20:57:03.013038 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-nrgbt"] Oct 05 20:57:03 crc kubenswrapper[4754]: I1005 20:57:03.013055 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wx9l2" event={"ID":"3408670c-55e3-4bf0-9589-fef6200c3490","Type":"ContainerStarted","Data":"034d05dd75fb8363ed4f45facdc38d1fb20e096a6385280c20cba323984208ab"} Oct 05 20:57:03 crc kubenswrapper[4754]: I1005 20:57:03.013075 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-s4zph"] Oct 05 20:57:03 crc kubenswrapper[4754]: I1005 20:57:03.015620 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-67bjd" event={"ID":"0593c91b-48f4-47aa-a57f-f9b50d7f385e","Type":"ContainerStarted","Data":"fc41e785d303160ccda66690762fb4ec5029125af382340f96583b906a2dae19"} Oct 05 20:57:03 crc kubenswrapper[4754]: W1005 20:57:03.037737 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2e5d7ee3_ea83_4445_a74d_9efa3864b3bb.slice/crio-283dcdb7107cef137855f14115356a50a71da3682a9eb0f7a6b562a616d5ddf0 WatchSource:0}: Error finding container 283dcdb7107cef137855f14115356a50a71da3682a9eb0f7a6b562a616d5ddf0: Status 404 returned error can't find the container with id 283dcdb7107cef137855f14115356a50a71da3682a9eb0f7a6b562a616d5ddf0 Oct 05 20:57:03 crc kubenswrapper[4754]: I1005 20:57:03.050801 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-pnhf9" event={"ID":"1e16d79e-eb1e-4b8b-8f0a-acb370163d17","Type":"ContainerStarted","Data":"2a2f7d826d242de3ef14c3ee7cf71e919b4aa69848335790746b587439befeb6"} Oct 05 20:57:03 crc kubenswrapper[4754]: I1005 20:57:03.052318 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-pnhf9" Oct 05 20:57:03 crc kubenswrapper[4754]: I1005 20:57:03.054183 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 20:57:03 crc kubenswrapper[4754]: I1005 20:57:03.063350 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-lv842" event={"ID":"be8e4dfb-5b8a-4bf4-8d54-880d51f990de","Type":"ContainerStarted","Data":"1e520347feb8e4e059d44021d93ff9f89c09748723f025638755ca86677d03d1"} Oct 05 20:57:03 crc kubenswrapper[4754]: E1005 20:57:03.071961 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 20:57:03.571926409 +0000 UTC m=+147.476045119 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:57:03 crc kubenswrapper[4754]: I1005 20:57:03.092149 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-5scsb" event={"ID":"5a98537a-186b-4150-bd41-c2627ec772f9","Type":"ContainerStarted","Data":"f04c0591594e925308a7505b9cd0f65bbb15451f0b69c8a9fabcd8343f4ba94a"} Oct 05 20:57:03 crc kubenswrapper[4754]: I1005 20:57:03.094702 4754 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-pnhf9 container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.21:8443/healthz\": dial tcp 10.217.0.21:8443: connect: connection refused" start-of-body= Oct 05 20:57:03 crc kubenswrapper[4754]: I1005 20:57:03.094873 4754 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-pnhf9" podUID="1e16d79e-eb1e-4b8b-8f0a-acb370163d17" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.21:8443/healthz\": dial tcp 10.217.0.21:8443: connect: connection refused" Oct 05 20:57:03 crc kubenswrapper[4754]: I1005 20:57:03.099704 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-hjpzm" event={"ID":"4da215ae-75ec-4ab1-81b0-39966d449b5f","Type":"ContainerStarted","Data":"0b0746dd46765e6feb34965fed817481e05bd0fe7166d6068cbd4fe8b0761d52"} Oct 05 20:57:03 crc kubenswrapper[4754]: I1005 20:57:03.117392 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-pnhf9" podStartSLOduration=125.117372068 podStartE2EDuration="2m5.117372068s" podCreationTimestamp="2025-10-05 20:54:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 20:57:03.112787167 +0000 UTC m=+147.016905877" watchObservedRunningTime="2025-10-05 20:57:03.117372068 +0000 UTC m=+147.021490778" Oct 05 20:57:03 crc kubenswrapper[4754]: I1005 20:57:03.121192 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29328285-mrswx" event={"ID":"60ff4c5c-9e45-47d3-b45d-a365e988342e","Type":"ContainerStarted","Data":"92251d5fc79e9f76bcbac6e3baa7658adb4cc5be49e94d1e9aba4a934ab16083"} Oct 05 20:57:03 crc kubenswrapper[4754]: W1005 20:57:03.126295 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5a5b1574_7f76_4bcf_8c24_9ff7773e86f2.slice/crio-b4e6bf7ac25d1b361be8785224e2c1b2d9da9a0b7d95656e1b8d2345745a13f4 WatchSource:0}: Error finding container b4e6bf7ac25d1b361be8785224e2c1b2d9da9a0b7d95656e1b8d2345745a13f4: Status 404 returned error can't find the container with id b4e6bf7ac25d1b361be8785224e2c1b2d9da9a0b7d95656e1b8d2345745a13f4 Oct 05 20:57:03 crc kubenswrapper[4754]: I1005 20:57:03.154370 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-cwtlx" event={"ID":"b157fa21-ae67-4c98-806c-93854b072965","Type":"ContainerStarted","Data":"5c707afa174976df75428f71cff936966298c71d75e8767f5b82690b0bb4b005"} Oct 05 20:57:03 crc kubenswrapper[4754]: I1005 20:57:03.158418 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5sqj9\" (UID: \"c64d57c5-4bd4-48b5-85fc-9d613aea29c4\") " pod="openshift-image-registry/image-registry-697d97f7c8-5sqj9" Oct 05 20:57:03 crc kubenswrapper[4754]: E1005 20:57:03.160065 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 20:57:03.660053354 +0000 UTC m=+147.564172064 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5sqj9" (UID: "c64d57c5-4bd4-48b5-85fc-9d613aea29c4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:57:03 crc kubenswrapper[4754]: I1005 20:57:03.167917 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-ll9lb" event={"ID":"ebf750fb-f47c-420b-931d-3ac0a23e9e7b","Type":"ContainerStarted","Data":"dbd7b33db94a3b5cb3fb459a809273d868b355fff75236f3eaad37e707b04a79"} Oct 05 20:57:03 crc kubenswrapper[4754]: I1005 20:57:03.191008 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nrd6b" event={"ID":"091efc31-55eb-4026-bf38-626e09a13253","Type":"ContainerStarted","Data":"5c94931e396d2f614ce53c986f9b1a0f35bb4845dbd59cf856b671451d0047df"} Oct 05 20:57:03 crc kubenswrapper[4754]: I1005 20:57:03.191693 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nrd6b" Oct 05 20:57:03 crc kubenswrapper[4754]: I1005 20:57:03.214177 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-ll9lb" podStartSLOduration=125.214158232 podStartE2EDuration="2m5.214158232s" podCreationTimestamp="2025-10-05 20:54:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 20:57:03.213206017 +0000 UTC m=+147.117324727" watchObservedRunningTime="2025-10-05 20:57:03.214158232 +0000 UTC m=+147.118276942" Oct 05 20:57:03 crc kubenswrapper[4754]: I1005 20:57:03.214988 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-5scsb" podStartSLOduration=126.214982324 podStartE2EDuration="2m6.214982324s" podCreationTimestamp="2025-10-05 20:54:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 20:57:03.171543888 +0000 UTC m=+147.075662598" watchObservedRunningTime="2025-10-05 20:57:03.214982324 +0000 UTC m=+147.119101034" Oct 05 20:57:03 crc kubenswrapper[4754]: I1005 20:57:03.227511 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rl5n7" event={"ID":"11048ba4-fef7-4675-b3c3-2671af4a4e27","Type":"ContainerStarted","Data":"dc7ae8ec6a9615c20806667ec7f279b5e4a268b25f1e00fe1da2c749c2ebc31f"} Oct 05 20:57:03 crc kubenswrapper[4754]: I1005 20:57:03.240166 4754 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-nrd6b container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.27:5443/healthz\": dial tcp 10.217.0.27:5443: connect: connection refused" start-of-body= Oct 05 20:57:03 crc kubenswrapper[4754]: I1005 20:57:03.240234 4754 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nrd6b" podUID="091efc31-55eb-4026-bf38-626e09a13253" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.27:5443/healthz\": dial tcp 10.217.0.27:5443: connect: connection refused" Oct 05 20:57:03 crc kubenswrapper[4754]: I1005 20:57:03.240582 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6wnmn" event={"ID":"9263c4f4-1540-4190-9e85-b1d8e2ca6c2f","Type":"ContainerStarted","Data":"b74bb753f21eb03762c9b7772022b77c6fd39f2b26e6d6d9d8218efd040b37fb"} Oct 05 20:57:03 crc kubenswrapper[4754]: I1005 20:57:03.246327 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-ksczt" event={"ID":"3934dd94-b9fa-4079-a268-60a9470dab41","Type":"ContainerStarted","Data":"a529de5dfecaa0467abf1a3739ea225d67526f76e63ac42f69d78637bbcec47d"} Oct 05 20:57:03 crc kubenswrapper[4754]: I1005 20:57:03.257834 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-9vndv" event={"ID":"b96341ba-da55-442f-851e-5666717a1d5a","Type":"ContainerStarted","Data":"d3ae03f0c96cadf4f156859d0ea5426ede05cee0f9797220c55db8aa4ba30cab"} Oct 05 20:57:03 crc kubenswrapper[4754]: I1005 20:57:03.261111 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 20:57:03 crc kubenswrapper[4754]: E1005 20:57:03.262722 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 20:57:03.762700673 +0000 UTC m=+147.666819373 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:57:03 crc kubenswrapper[4754]: I1005 20:57:03.271959 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-8qm2l" event={"ID":"80ede041-caa4-451e-bdd1-083e2e03ec29","Type":"ContainerStarted","Data":"b090addabff6b9bbe9325e2b45405168ff4e3241e31d5f84986aa28ab46751ed"} Oct 05 20:57:03 crc kubenswrapper[4754]: I1005 20:57:03.296728 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-5wx56"] Oct 05 20:57:03 crc kubenswrapper[4754]: I1005 20:57:03.303280 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-nrt9t" event={"ID":"0b4c539c-22cf-4337-9959-afcc73e64370","Type":"ContainerStarted","Data":"e130cb92f635ca5986769523e7102176fdf883ec7fcb6d6c967161dfa0134fac"} Oct 05 20:57:03 crc kubenswrapper[4754]: I1005 20:57:03.321506 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nrd6b" podStartSLOduration=125.321473524 podStartE2EDuration="2m5.321473524s" podCreationTimestamp="2025-10-05 20:54:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 20:57:03.319585774 +0000 UTC m=+147.223704484" watchObservedRunningTime="2025-10-05 20:57:03.321473524 +0000 UTC m=+147.225592234" Oct 05 20:57:03 crc kubenswrapper[4754]: I1005 20:57:03.349093 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jpbfr" event={"ID":"c2c0ebd0-5cd9-49f4-9cd6-df5995cfcd74","Type":"ContainerStarted","Data":"68b8ae952289823a4dbcbae475c9a850c4fb3dbbd575f195bf848dd8c0c21474"} Oct 05 20:57:03 crc kubenswrapper[4754]: I1005 20:57:03.362785 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5sqj9\" (UID: \"c64d57c5-4bd4-48b5-85fc-9d613aea29c4\") " pod="openshift-image-registry/image-registry-697d97f7c8-5sqj9" Oct 05 20:57:03 crc kubenswrapper[4754]: E1005 20:57:03.364934 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 20:57:03.864915871 +0000 UTC m=+147.769034581 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5sqj9" (UID: "c64d57c5-4bd4-48b5-85fc-9d613aea29c4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:57:03 crc kubenswrapper[4754]: I1005 20:57:03.377630 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rl5n7" podStartSLOduration=125.377608026 podStartE2EDuration="2m5.377608026s" podCreationTimestamp="2025-10-05 20:54:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 20:57:03.376887587 +0000 UTC m=+147.281006287" watchObservedRunningTime="2025-10-05 20:57:03.377608026 +0000 UTC m=+147.281726736" Oct 05 20:57:03 crc kubenswrapper[4754]: I1005 20:57:03.463610 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 20:57:03 crc kubenswrapper[4754]: E1005 20:57:03.464348 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 20:57:03.964318934 +0000 UTC m=+147.868437644 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:57:03 crc kubenswrapper[4754]: I1005 20:57:03.468870 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5sqj9\" (UID: \"c64d57c5-4bd4-48b5-85fc-9d613aea29c4\") " pod="openshift-image-registry/image-registry-697d97f7c8-5sqj9" Oct 05 20:57:03 crc kubenswrapper[4754]: E1005 20:57:03.470537 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 20:57:03.970523048 +0000 UTC m=+147.874641758 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5sqj9" (UID: "c64d57c5-4bd4-48b5-85fc-9d613aea29c4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:57:03 crc kubenswrapper[4754]: I1005 20:57:03.571209 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 20:57:03 crc kubenswrapper[4754]: E1005 20:57:03.571334 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 20:57:04.071311207 +0000 UTC m=+147.975429917 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:57:03 crc kubenswrapper[4754]: I1005 20:57:03.571604 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5sqj9\" (UID: \"c64d57c5-4bd4-48b5-85fc-9d613aea29c4\") " pod="openshift-image-registry/image-registry-697d97f7c8-5sqj9" Oct 05 20:57:03 crc kubenswrapper[4754]: E1005 20:57:03.572907 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 20:57:04.072882929 +0000 UTC m=+147.977001839 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5sqj9" (UID: "c64d57c5-4bd4-48b5-85fc-9d613aea29c4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:57:03 crc kubenswrapper[4754]: I1005 20:57:03.672264 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 20:57:03 crc kubenswrapper[4754]: E1005 20:57:03.672768 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 20:57:04.172749014 +0000 UTC m=+148.076867724 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:57:03 crc kubenswrapper[4754]: I1005 20:57:03.774363 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5sqj9\" (UID: \"c64d57c5-4bd4-48b5-85fc-9d613aea29c4\") " pod="openshift-image-registry/image-registry-697d97f7c8-5sqj9" Oct 05 20:57:03 crc kubenswrapper[4754]: E1005 20:57:03.774726 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 20:57:04.274712345 +0000 UTC m=+148.178831055 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5sqj9" (UID: "c64d57c5-4bd4-48b5-85fc-9d613aea29c4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:57:03 crc kubenswrapper[4754]: I1005 20:57:03.875277 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 20:57:03 crc kubenswrapper[4754]: E1005 20:57:03.875445 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 20:57:04.375417443 +0000 UTC m=+148.279536153 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:57:03 crc kubenswrapper[4754]: I1005 20:57:03.875679 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5sqj9\" (UID: \"c64d57c5-4bd4-48b5-85fc-9d613aea29c4\") " pod="openshift-image-registry/image-registry-697d97f7c8-5sqj9" Oct 05 20:57:03 crc kubenswrapper[4754]: E1005 20:57:03.876072 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 20:57:04.376053959 +0000 UTC m=+148.280172669 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5sqj9" (UID: "c64d57c5-4bd4-48b5-85fc-9d613aea29c4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:57:03 crc kubenswrapper[4754]: I1005 20:57:03.976393 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 20:57:03 crc kubenswrapper[4754]: E1005 20:57:03.976937 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 20:57:04.47689225 +0000 UTC m=+148.381010950 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:57:03 crc kubenswrapper[4754]: I1005 20:57:03.984300 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-ll9lb" Oct 05 20:57:04 crc kubenswrapper[4754]: I1005 20:57:03.995344 4754 patch_prober.go:28] interesting pod/router-default-5444994796-ll9lb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 05 20:57:04 crc kubenswrapper[4754]: [-]has-synced failed: reason withheld Oct 05 20:57:04 crc kubenswrapper[4754]: [+]process-running ok Oct 05 20:57:04 crc kubenswrapper[4754]: healthz check failed Oct 05 20:57:04 crc kubenswrapper[4754]: I1005 20:57:03.995447 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ll9lb" podUID="ebf750fb-f47c-420b-931d-3ac0a23e9e7b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 05 20:57:04 crc kubenswrapper[4754]: I1005 20:57:04.078327 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5sqj9\" (UID: \"c64d57c5-4bd4-48b5-85fc-9d613aea29c4\") " pod="openshift-image-registry/image-registry-697d97f7c8-5sqj9" Oct 05 20:57:04 crc kubenswrapper[4754]: E1005 20:57:04.078881 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 20:57:04.578865081 +0000 UTC m=+148.482983791 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5sqj9" (UID: "c64d57c5-4bd4-48b5-85fc-9d613aea29c4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:57:04 crc kubenswrapper[4754]: I1005 20:57:04.180224 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 20:57:04 crc kubenswrapper[4754]: E1005 20:57:04.180436 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 20:57:04.680404781 +0000 UTC m=+148.584523501 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:57:04 crc kubenswrapper[4754]: I1005 20:57:04.180635 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5sqj9\" (UID: \"c64d57c5-4bd4-48b5-85fc-9d613aea29c4\") " pod="openshift-image-registry/image-registry-697d97f7c8-5sqj9" Oct 05 20:57:04 crc kubenswrapper[4754]: E1005 20:57:04.181160 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 20:57:04.681151531 +0000 UTC m=+148.585270241 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5sqj9" (UID: "c64d57c5-4bd4-48b5-85fc-9d613aea29c4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:57:04 crc kubenswrapper[4754]: I1005 20:57:04.292260 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 20:57:04 crc kubenswrapper[4754]: E1005 20:57:04.293103 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 20:57:04.793060234 +0000 UTC m=+148.697179134 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:57:04 crc kubenswrapper[4754]: I1005 20:57:04.402797 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5sqj9\" (UID: \"c64d57c5-4bd4-48b5-85fc-9d613aea29c4\") " pod="openshift-image-registry/image-registry-697d97f7c8-5sqj9" Oct 05 20:57:04 crc kubenswrapper[4754]: E1005 20:57:04.404187 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 20:57:04.904169165 +0000 UTC m=+148.808287875 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5sqj9" (UID: "c64d57c5-4bd4-48b5-85fc-9d613aea29c4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:57:04 crc kubenswrapper[4754]: I1005 20:57:04.480700 4754 generic.go:334] "Generic (PLEG): container finished" podID="b157fa21-ae67-4c98-806c-93854b072965" containerID="5c707afa174976df75428f71cff936966298c71d75e8767f5b82690b0bb4b005" exitCode=0 Oct 05 20:57:04 crc kubenswrapper[4754]: I1005 20:57:04.481670 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-cwtlx" event={"ID":"b157fa21-ae67-4c98-806c-93854b072965","Type":"ContainerDied","Data":"5c707afa174976df75428f71cff936966298c71d75e8767f5b82690b0bb4b005"} Oct 05 20:57:04 crc kubenswrapper[4754]: I1005 20:57:04.505554 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 20:57:04 crc kubenswrapper[4754]: E1005 20:57:04.506091 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 20:57:05.006073174 +0000 UTC m=+148.910191874 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:57:04 crc kubenswrapper[4754]: I1005 20:57:04.509723 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lxp7v" event={"ID":"d3c66e2a-8404-498c-b8a9-2bd2e6fd318b","Type":"ContainerDied","Data":"274447830cb60a82d71e06897dee4840c2055fdf900c69234541a73c7195c72f"} Oct 05 20:57:04 crc kubenswrapper[4754]: I1005 20:57:04.524625 4754 generic.go:334] "Generic (PLEG): container finished" podID="d3c66e2a-8404-498c-b8a9-2bd2e6fd318b" containerID="274447830cb60a82d71e06897dee4840c2055fdf900c69234541a73c7195c72f" exitCode=0 Oct 05 20:57:04 crc kubenswrapper[4754]: I1005 20:57:04.550198 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jpbfr" event={"ID":"c2c0ebd0-5cd9-49f4-9cd6-df5995cfcd74","Type":"ContainerStarted","Data":"bf568e6521564aa099763a42aa44a134522fc7f741d635a91700479b74f007bc"} Oct 05 20:57:04 crc kubenswrapper[4754]: I1005 20:57:04.551305 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jpbfr" Oct 05 20:57:04 crc kubenswrapper[4754]: I1005 20:57:04.568737 4754 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-jpbfr container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.39:8443/healthz\": dial tcp 10.217.0.39:8443: connect: connection refused" start-of-body= Oct 05 20:57:04 crc kubenswrapper[4754]: I1005 20:57:04.568814 4754 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jpbfr" podUID="c2c0ebd0-5cd9-49f4-9cd6-df5995cfcd74" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.39:8443/healthz\": dial tcp 10.217.0.39:8443: connect: connection refused" Oct 05 20:57:04 crc kubenswrapper[4754]: I1005 20:57:04.598305 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29328285-mrswx" event={"ID":"60ff4c5c-9e45-47d3-b45d-a365e988342e","Type":"ContainerStarted","Data":"ff732b3e3d89105355ef4324b72c3db67106face7c8a9fd05f5ee00e8b2ceb24"} Oct 05 20:57:04 crc kubenswrapper[4754]: I1005 20:57:04.608952 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5sqj9\" (UID: \"c64d57c5-4bd4-48b5-85fc-9d613aea29c4\") " pod="openshift-image-registry/image-registry-697d97f7c8-5sqj9" Oct 05 20:57:04 crc kubenswrapper[4754]: E1005 20:57:04.619610 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 20:57:05.11959318 +0000 UTC m=+149.023711880 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5sqj9" (UID: "c64d57c5-4bd4-48b5-85fc-9d613aea29c4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:57:04 crc kubenswrapper[4754]: I1005 20:57:04.659196 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-nrgbt" event={"ID":"2e5d7ee3-ea83-4445-a74d-9efa3864b3bb","Type":"ContainerStarted","Data":"6fde5d9abe42a0ada022c86114cad72be3cbef7cf6ea3ee620945e4a647de714"} Oct 05 20:57:04 crc kubenswrapper[4754]: I1005 20:57:04.659257 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-nrgbt" event={"ID":"2e5d7ee3-ea83-4445-a74d-9efa3864b3bb","Type":"ContainerStarted","Data":"283dcdb7107cef137855f14115356a50a71da3682a9eb0f7a6b562a616d5ddf0"} Oct 05 20:57:04 crc kubenswrapper[4754]: I1005 20:57:04.671253 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-8qm2l" event={"ID":"80ede041-caa4-451e-bdd1-083e2e03ec29","Type":"ContainerStarted","Data":"eb2bd60f17fd03ce6a7a9d2abb4fc17c6312b4f4d95219ac883d1daec9ebd6b2"} Oct 05 20:57:04 crc kubenswrapper[4754]: I1005 20:57:04.705329 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jpbfr" podStartSLOduration=126.705303822 podStartE2EDuration="2m6.705303822s" podCreationTimestamp="2025-10-05 20:54:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 20:57:04.69691802 +0000 UTC m=+148.601036740" watchObservedRunningTime="2025-10-05 20:57:04.705303822 +0000 UTC m=+148.609422532" Oct 05 20:57:04 crc kubenswrapper[4754]: I1005 20:57:04.707250 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nrd6b" event={"ID":"091efc31-55eb-4026-bf38-626e09a13253","Type":"ContainerStarted","Data":"54e2c54070f67cddf9f35d3725e7ff1e6a33f9c6c228321b870904127b78bcb4"} Oct 05 20:57:04 crc kubenswrapper[4754]: I1005 20:57:04.713707 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 20:57:04 crc kubenswrapper[4754]: E1005 20:57:04.715810 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 20:57:05.215790918 +0000 UTC m=+149.119909628 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:57:04 crc kubenswrapper[4754]: I1005 20:57:04.736113 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-67bjd" event={"ID":"0593c91b-48f4-47aa-a57f-f9b50d7f385e","Type":"ContainerStarted","Data":"9ea835e379889351357772c0613bafcf32fa1204bd6dc04be2b1984fe7354aca"} Oct 05 20:57:04 crc kubenswrapper[4754]: I1005 20:57:04.767805 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-8qm2l" podStartSLOduration=126.76778401 podStartE2EDuration="2m6.76778401s" podCreationTimestamp="2025-10-05 20:54:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 20:57:04.767204575 +0000 UTC m=+148.671323285" watchObservedRunningTime="2025-10-05 20:57:04.76778401 +0000 UTC m=+148.671902720" Oct 05 20:57:04 crc kubenswrapper[4754]: I1005 20:57:04.774531 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-2qmwc" event={"ID":"bc8bf6c1-363b-4e24-830c-898e4f7c6325","Type":"ContainerStarted","Data":"a0b7694e70661a8e73d90a4cf4444412e67fce64d00d3542e76602b223e458f0"} Oct 05 20:57:04 crc kubenswrapper[4754]: I1005 20:57:04.802740 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-9vndv" event={"ID":"b96341ba-da55-442f-851e-5666717a1d5a","Type":"ContainerStarted","Data":"610d44e066b2e4cb793c517a3de0b43134be4f289ca47f07836f2f933291ca85"} Oct 05 20:57:04 crc kubenswrapper[4754]: I1005 20:57:04.815634 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5sqj9\" (UID: \"c64d57c5-4bd4-48b5-85fc-9d613aea29c4\") " pod="openshift-image-registry/image-registry-697d97f7c8-5sqj9" Oct 05 20:57:04 crc kubenswrapper[4754]: E1005 20:57:04.817876 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 20:57:05.317863342 +0000 UTC m=+149.221982042 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5sqj9" (UID: "c64d57c5-4bd4-48b5-85fc-9d613aea29c4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:57:04 crc kubenswrapper[4754]: I1005 20:57:04.856633 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-wz2dg" event={"ID":"cce92b6a-69b0-49d5-bb66-36349ca96841","Type":"ContainerStarted","Data":"7260493e409f96c48caef86b0c8bd5bcde859193f8a2a7a305612370fd7a9e7c"} Oct 05 20:57:04 crc kubenswrapper[4754]: I1005 20:57:04.879961 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-ksczt" event={"ID":"3934dd94-b9fa-4079-a268-60a9470dab41","Type":"ContainerStarted","Data":"94cbd304317e03cb1b11be58e998caa4de3a9bb0021d757486c074dde554dc2b"} Oct 05 20:57:04 crc kubenswrapper[4754]: I1005 20:57:04.882311 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-ksczt" Oct 05 20:57:04 crc kubenswrapper[4754]: I1005 20:57:04.882685 4754 patch_prober.go:28] interesting pod/downloads-7954f5f757-ksczt container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.20:8080/\": dial tcp 10.217.0.20:8080: connect: connection refused" start-of-body= Oct 05 20:57:04 crc kubenswrapper[4754]: I1005 20:57:04.882729 4754 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-ksczt" podUID="3934dd94-b9fa-4079-a268-60a9470dab41" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.20:8080/\": dial tcp 10.217.0.20:8080: connect: connection refused" Oct 05 20:57:04 crc kubenswrapper[4754]: I1005 20:57:04.916696 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-nrgbt" podStartSLOduration=126.91667428 podStartE2EDuration="2m6.91667428s" podCreationTimestamp="2025-10-05 20:54:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 20:57:04.913909867 +0000 UTC m=+148.818028577" watchObservedRunningTime="2025-10-05 20:57:04.91667428 +0000 UTC m=+148.820792990" Oct 05 20:57:04 crc kubenswrapper[4754]: I1005 20:57:04.917251 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 20:57:04 crc kubenswrapper[4754]: E1005 20:57:04.917695 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 20:57:05.417677046 +0000 UTC m=+149.321795756 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:57:04 crc kubenswrapper[4754]: I1005 20:57:04.973444 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29328285-mrswx" podStartSLOduration=126.973417927 podStartE2EDuration="2m6.973417927s" podCreationTimestamp="2025-10-05 20:54:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 20:57:04.960818964 +0000 UTC m=+148.864937674" watchObservedRunningTime="2025-10-05 20:57:04.973417927 +0000 UTC m=+148.877536637" Oct 05 20:57:04 crc kubenswrapper[4754]: I1005 20:57:04.980094 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-nrt9t" event={"ID":"0b4c539c-22cf-4337-9959-afcc73e64370","Type":"ContainerStarted","Data":"1627057c31bc921fae4eb22f34d169480c7c562668f566fac88e392b87c9039b"} Oct 05 20:57:04 crc kubenswrapper[4754]: I1005 20:57:04.981154 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-nrt9t" Oct 05 20:57:04 crc kubenswrapper[4754]: I1005 20:57:04.988814 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-lv842" event={"ID":"be8e4dfb-5b8a-4bf4-8d54-880d51f990de","Type":"ContainerStarted","Data":"c0319442c656a7f71ae1e2a0be9bc644cad62e3bf32f637684c5b6921d5608e7"} Oct 05 20:57:04 crc kubenswrapper[4754]: I1005 20:57:04.989735 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-lv842" Oct 05 20:57:04 crc kubenswrapper[4754]: I1005 20:57:04.990927 4754 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-lv842 container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.42:8443/healthz\": dial tcp 10.217.0.42:8443: connect: connection refused" start-of-body= Oct 05 20:57:04 crc kubenswrapper[4754]: I1005 20:57:04.990987 4754 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-lv842" podUID="be8e4dfb-5b8a-4bf4-8d54-880d51f990de" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.42:8443/healthz\": dial tcp 10.217.0.42:8443: connect: connection refused" Oct 05 20:57:04 crc kubenswrapper[4754]: I1005 20:57:04.995005 4754 patch_prober.go:28] interesting pod/router-default-5444994796-ll9lb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 05 20:57:04 crc kubenswrapper[4754]: [-]has-synced failed: reason withheld Oct 05 20:57:04 crc kubenswrapper[4754]: [+]process-running ok Oct 05 20:57:04 crc kubenswrapper[4754]: healthz check failed Oct 05 20:57:04 crc kubenswrapper[4754]: I1005 20:57:04.995059 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ll9lb" podUID="ebf750fb-f47c-420b-931d-3ac0a23e9e7b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 05 20:57:04 crc kubenswrapper[4754]: I1005 20:57:04.999374 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-hjpzm" event={"ID":"4da215ae-75ec-4ab1-81b0-39966d449b5f","Type":"ContainerStarted","Data":"588d94843e602f9bcae81cd4749bfacca82dc46ecd462943b43b19d7873662e0"} Oct 05 20:57:05 crc kubenswrapper[4754]: I1005 20:57:05.006369 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-ksczt" podStartSLOduration=127.006344876 podStartE2EDuration="2m7.006344876s" podCreationTimestamp="2025-10-05 20:54:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 20:57:05.003314766 +0000 UTC m=+148.907433476" watchObservedRunningTime="2025-10-05 20:57:05.006344876 +0000 UTC m=+148.910463606" Oct 05 20:57:05 crc kubenswrapper[4754]: I1005 20:57:05.019475 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5sqj9\" (UID: \"c64d57c5-4bd4-48b5-85fc-9d613aea29c4\") " pod="openshift-image-registry/image-registry-697d97f7c8-5sqj9" Oct 05 20:57:05 crc kubenswrapper[4754]: E1005 20:57:05.019943 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 20:57:05.519928814 +0000 UTC m=+149.424047524 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5sqj9" (UID: "c64d57c5-4bd4-48b5-85fc-9d613aea29c4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:57:05 crc kubenswrapper[4754]: I1005 20:57:05.050571 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-q2jq4" event={"ID":"dbe65326-0856-4b33-8894-4c9266bc1a90","Type":"ContainerStarted","Data":"e79bf7952caa4967ed4531a76d459d231f21530274f6fe74c31df88e3a8babba"} Oct 05 20:57:05 crc kubenswrapper[4754]: I1005 20:57:05.074084 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-nrt9t" podStartSLOduration=127.074065473 podStartE2EDuration="2m7.074065473s" podCreationTimestamp="2025-10-05 20:54:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 20:57:05.070084148 +0000 UTC m=+148.974202858" watchObservedRunningTime="2025-10-05 20:57:05.074065473 +0000 UTC m=+148.978184183" Oct 05 20:57:05 crc kubenswrapper[4754]: I1005 20:57:05.102238 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-k8qff" event={"ID":"20a5d042-ffe2-4244-b25d-700b26534a04","Type":"ContainerStarted","Data":"eb8315ba24f1f00a48be71f3b53a8130c7faad1288878a8dc79c4567463d0af7"} Oct 05 20:57:05 crc kubenswrapper[4754]: I1005 20:57:05.107015 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-chv7x" event={"ID":"8797a357-3b55-4d63-9215-4c64bb676a3d","Type":"ContainerStarted","Data":"d981b2bdcbeb275a7ae60ca2eaaf8ff7f9a3fb8e88ee1ac383c1685b75e77158"} Oct 05 20:57:05 crc kubenswrapper[4754]: I1005 20:57:05.126239 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 20:57:05 crc kubenswrapper[4754]: E1005 20:57:05.126471 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 20:57:05.626427045 +0000 UTC m=+149.530545755 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:57:05 crc kubenswrapper[4754]: I1005 20:57:05.135896 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5sqj9\" (UID: \"c64d57c5-4bd4-48b5-85fc-9d613aea29c4\") " pod="openshift-image-registry/image-registry-697d97f7c8-5sqj9" Oct 05 20:57:05 crc kubenswrapper[4754]: I1005 20:57:05.133341 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-hjpzm" podStartSLOduration=127.133319426 podStartE2EDuration="2m7.133319426s" podCreationTimestamp="2025-10-05 20:54:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 20:57:05.128505329 +0000 UTC m=+149.032624039" watchObservedRunningTime="2025-10-05 20:57:05.133319426 +0000 UTC m=+149.037438136" Oct 05 20:57:05 crc kubenswrapper[4754]: E1005 20:57:05.139849 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 20:57:05.639830958 +0000 UTC m=+149.543949668 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5sqj9" (UID: "c64d57c5-4bd4-48b5-85fc-9d613aea29c4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:57:05 crc kubenswrapper[4754]: I1005 20:57:05.155346 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-7g427" event={"ID":"230b53bd-7c59-42f5-9604-0c5414fcdf80","Type":"ContainerStarted","Data":"cc46eb3d5a3478d549363b81ddae95df9caaae64465ea3982d633ecf5995ed04"} Oct 05 20:57:05 crc kubenswrapper[4754]: I1005 20:57:05.163880 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-5wx56" event={"ID":"d1b2b973-8b95-4f6a-83f1-468f84fb006d","Type":"ContainerStarted","Data":"c03f08aa24d7dea2a45a925af90a2f5a4045550645f6abd891a9d6099ae3fc2f"} Oct 05 20:57:05 crc kubenswrapper[4754]: I1005 20:57:05.176914 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-682zv" event={"ID":"51547f65-61cc-4504-8399-5d540160fea1","Type":"ContainerStarted","Data":"e79e5005fcb746225ff8c9b4f7a843de34a842152f610c2adb04d638da322504"} Oct 05 20:57:05 crc kubenswrapper[4754]: I1005 20:57:05.197388 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-6xm7j" event={"ID":"830a0a27-fab1-4efc-aeb3-e38a962ffcfa","Type":"ContainerStarted","Data":"2aed80ad41ffe3d443bf1237da3c8720769d9f335799ce772a74041ecd7ae454"} Oct 05 20:57:05 crc kubenswrapper[4754]: I1005 20:57:05.207510 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-lv842" podStartSLOduration=127.207470543 podStartE2EDuration="2m7.207470543s" podCreationTimestamp="2025-10-05 20:54:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 20:57:05.205044009 +0000 UTC m=+149.109162719" watchObservedRunningTime="2025-10-05 20:57:05.207470543 +0000 UTC m=+149.111589253" Oct 05 20:57:05 crc kubenswrapper[4754]: I1005 20:57:05.245281 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 20:57:05 crc kubenswrapper[4754]: E1005 20:57:05.245913 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 20:57:05.745896347 +0000 UTC m=+149.650015057 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:57:05 crc kubenswrapper[4754]: I1005 20:57:05.246322 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5sqj9\" (UID: \"c64d57c5-4bd4-48b5-85fc-9d613aea29c4\") " pod="openshift-image-registry/image-registry-697d97f7c8-5sqj9" Oct 05 20:57:05 crc kubenswrapper[4754]: I1005 20:57:05.250934 4754 patch_prober.go:28] interesting pod/machine-config-daemon-b2h9k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 20:57:05 crc kubenswrapper[4754]: I1005 20:57:05.250999 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 20:57:05 crc kubenswrapper[4754]: I1005 20:57:05.251257 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-s4zph" event={"ID":"5a5b1574-7f76-4bcf-8c24-9ff7773e86f2","Type":"ContainerStarted","Data":"b4e6bf7ac25d1b361be8785224e2c1b2d9da9a0b7d95656e1b8d2345745a13f4"} Oct 05 20:57:05 crc kubenswrapper[4754]: E1005 20:57:05.284925 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 20:57:05.784896327 +0000 UTC m=+149.689015037 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5sqj9" (UID: "c64d57c5-4bd4-48b5-85fc-9d613aea29c4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:57:05 crc kubenswrapper[4754]: I1005 20:57:05.310590 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-w6krf" event={"ID":"e542cd22-7a8b-4a4d-8205-d3dc60b2b8c6","Type":"ContainerStarted","Data":"c878dfd75be1b179225fead5d922ee8e3abe2be012d4573d79c5dd6f3c8a1181"} Oct 05 20:57:05 crc kubenswrapper[4754]: I1005 20:57:05.326699 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-ll5jq" event={"ID":"c1c4fc78-c305-4dba-b4d5-9f05d58a8ef6","Type":"ContainerStarted","Data":"1775ab67489ca4d3c34d5d072cb2215fe929ee07d5f7e5d8cda18d28393efd0a"} Oct 05 20:57:05 crc kubenswrapper[4754]: I1005 20:57:05.330629 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-ll5jq" Oct 05 20:57:05 crc kubenswrapper[4754]: I1005 20:57:05.336929 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-gchnw" event={"ID":"6c0dfaff-08a7-4152-8624-1973e4a4d148","Type":"ContainerStarted","Data":"ef2a23bdc3d3f895d033e408cb22c61a2a1743aa1b90d1cb9e60b3bae60fa470"} Oct 05 20:57:05 crc kubenswrapper[4754]: I1005 20:57:05.338432 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-q2jq4" podStartSLOduration=127.338406439 podStartE2EDuration="2m7.338406439s" podCreationTimestamp="2025-10-05 20:54:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 20:57:05.291327586 +0000 UTC m=+149.195446296" watchObservedRunningTime="2025-10-05 20:57:05.338406439 +0000 UTC m=+149.242525149" Oct 05 20:57:05 crc kubenswrapper[4754]: I1005 20:57:05.348396 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-gchnw" Oct 05 20:57:05 crc kubenswrapper[4754]: I1005 20:57:05.351799 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-k8qff" podStartSLOduration=128.351776621 podStartE2EDuration="2m8.351776621s" podCreationTimestamp="2025-10-05 20:54:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 20:57:05.337536926 +0000 UTC m=+149.241655636" watchObservedRunningTime="2025-10-05 20:57:05.351776621 +0000 UTC m=+149.255895331" Oct 05 20:57:05 crc kubenswrapper[4754]: I1005 20:57:05.355301 4754 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-ll5jq container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.22:8080/healthz\": dial tcp 10.217.0.22:8080: connect: connection refused" start-of-body= Oct 05 20:57:05 crc kubenswrapper[4754]: I1005 20:57:05.355360 4754 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-ll5jq" podUID="c1c4fc78-c305-4dba-b4d5-9f05d58a8ef6" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.22:8080/healthz\": dial tcp 10.217.0.22:8080: connect: connection refused" Oct 05 20:57:05 crc kubenswrapper[4754]: I1005 20:57:05.359447 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 20:57:05 crc kubenswrapper[4754]: E1005 20:57:05.360119 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 20:57:05.860087311 +0000 UTC m=+149.764206021 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:57:05 crc kubenswrapper[4754]: I1005 20:57:05.369368 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5sqj9\" (UID: \"c64d57c5-4bd4-48b5-85fc-9d613aea29c4\") " pod="openshift-image-registry/image-registry-697d97f7c8-5sqj9" Oct 05 20:57:05 crc kubenswrapper[4754]: E1005 20:57:05.369798 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 20:57:05.869782437 +0000 UTC m=+149.773901147 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5sqj9" (UID: "c64d57c5-4bd4-48b5-85fc-9d613aea29c4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:57:05 crc kubenswrapper[4754]: I1005 20:57:05.382256 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-pnhf9" Oct 05 20:57:05 crc kubenswrapper[4754]: I1005 20:57:05.437952 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-chv7x" podStartSLOduration=127.437922275 podStartE2EDuration="2m7.437922275s" podCreationTimestamp="2025-10-05 20:54:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 20:57:05.427640264 +0000 UTC m=+149.331758974" watchObservedRunningTime="2025-10-05 20:57:05.437922275 +0000 UTC m=+149.342040985" Oct 05 20:57:05 crc kubenswrapper[4754]: I1005 20:57:05.482071 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 20:57:05 crc kubenswrapper[4754]: E1005 20:57:05.483463 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 20:57:05.983443036 +0000 UTC m=+149.887561746 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:57:05 crc kubenswrapper[4754]: I1005 20:57:05.537034 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-7g427" podStartSLOduration=8.53701143 podStartE2EDuration="8.53701143s" podCreationTimestamp="2025-10-05 20:56:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 20:57:05.480605591 +0000 UTC m=+149.384724301" watchObservedRunningTime="2025-10-05 20:57:05.53701143 +0000 UTC m=+149.441130140" Oct 05 20:57:05 crc kubenswrapper[4754]: I1005 20:57:05.537838 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-gchnw" podStartSLOduration=128.537833011 podStartE2EDuration="2m8.537833011s" podCreationTimestamp="2025-10-05 20:54:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 20:57:05.531714 +0000 UTC m=+149.435832730" watchObservedRunningTime="2025-10-05 20:57:05.537833011 +0000 UTC m=+149.441951721" Oct 05 20:57:05 crc kubenswrapper[4754]: I1005 20:57:05.558892 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-ll5jq" podStartSLOduration=127.558868296 podStartE2EDuration="2m7.558868296s" podCreationTimestamp="2025-10-05 20:54:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 20:57:05.556700199 +0000 UTC m=+149.460818909" watchObservedRunningTime="2025-10-05 20:57:05.558868296 +0000 UTC m=+149.462987006" Oct 05 20:57:05 crc kubenswrapper[4754]: I1005 20:57:05.586234 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5sqj9\" (UID: \"c64d57c5-4bd4-48b5-85fc-9d613aea29c4\") " pod="openshift-image-registry/image-registry-697d97f7c8-5sqj9" Oct 05 20:57:05 crc kubenswrapper[4754]: E1005 20:57:05.586652 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 20:57:06.086638829 +0000 UTC m=+149.990757529 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5sqj9" (UID: "c64d57c5-4bd4-48b5-85fc-9d613aea29c4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:57:05 crc kubenswrapper[4754]: I1005 20:57:05.679545 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-gchnw" Oct 05 20:57:05 crc kubenswrapper[4754]: I1005 20:57:05.696738 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 20:57:05 crc kubenswrapper[4754]: E1005 20:57:05.697262 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 20:57:06.197239638 +0000 UTC m=+150.101358348 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:57:05 crc kubenswrapper[4754]: I1005 20:57:05.708406 4754 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-nrd6b container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.27:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 05 20:57:05 crc kubenswrapper[4754]: I1005 20:57:05.708481 4754 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nrd6b" podUID="091efc31-55eb-4026-bf38-626e09a13253" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.27:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 05 20:57:05 crc kubenswrapper[4754]: I1005 20:57:05.798449 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5sqj9\" (UID: \"c64d57c5-4bd4-48b5-85fc-9d613aea29c4\") " pod="openshift-image-registry/image-registry-697d97f7c8-5sqj9" Oct 05 20:57:05 crc kubenswrapper[4754]: E1005 20:57:05.798826 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 20:57:06.298813078 +0000 UTC m=+150.202931788 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5sqj9" (UID: "c64d57c5-4bd4-48b5-85fc-9d613aea29c4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:57:05 crc kubenswrapper[4754]: I1005 20:57:05.899434 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 20:57:05 crc kubenswrapper[4754]: E1005 20:57:05.899681 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 20:57:06.399646559 +0000 UTC m=+150.303765269 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:57:05 crc kubenswrapper[4754]: I1005 20:57:05.899737 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5sqj9\" (UID: \"c64d57c5-4bd4-48b5-85fc-9d613aea29c4\") " pod="openshift-image-registry/image-registry-697d97f7c8-5sqj9" Oct 05 20:57:05 crc kubenswrapper[4754]: E1005 20:57:05.900325 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 20:57:06.400305227 +0000 UTC m=+150.304423927 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5sqj9" (UID: "c64d57c5-4bd4-48b5-85fc-9d613aea29c4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:57:05 crc kubenswrapper[4754]: I1005 20:57:05.988754 4754 patch_prober.go:28] interesting pod/router-default-5444994796-ll9lb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 05 20:57:05 crc kubenswrapper[4754]: [-]has-synced failed: reason withheld Oct 05 20:57:05 crc kubenswrapper[4754]: [+]process-running ok Oct 05 20:57:05 crc kubenswrapper[4754]: healthz check failed Oct 05 20:57:05 crc kubenswrapper[4754]: I1005 20:57:05.989101 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ll9lb" podUID="ebf750fb-f47c-420b-931d-3ac0a23e9e7b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 05 20:57:06 crc kubenswrapper[4754]: I1005 20:57:06.001445 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 20:57:06 crc kubenswrapper[4754]: E1005 20:57:06.001713 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 20:57:06.501674952 +0000 UTC m=+150.405793662 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:57:06 crc kubenswrapper[4754]: I1005 20:57:06.001867 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5sqj9\" (UID: \"c64d57c5-4bd4-48b5-85fc-9d613aea29c4\") " pod="openshift-image-registry/image-registry-697d97f7c8-5sqj9" Oct 05 20:57:06 crc kubenswrapper[4754]: E1005 20:57:06.002226 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 20:57:06.502208406 +0000 UTC m=+150.406327116 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5sqj9" (UID: "c64d57c5-4bd4-48b5-85fc-9d613aea29c4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:57:06 crc kubenswrapper[4754]: I1005 20:57:06.102796 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 20:57:06 crc kubenswrapper[4754]: E1005 20:57:06.103257 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 20:57:06.603225072 +0000 UTC m=+150.507343782 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:57:06 crc kubenswrapper[4754]: I1005 20:57:06.203966 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5sqj9\" (UID: \"c64d57c5-4bd4-48b5-85fc-9d613aea29c4\") " pod="openshift-image-registry/image-registry-697d97f7c8-5sqj9" Oct 05 20:57:06 crc kubenswrapper[4754]: E1005 20:57:06.204461 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 20:57:06.704438183 +0000 UTC m=+150.608556883 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5sqj9" (UID: "c64d57c5-4bd4-48b5-85fc-9d613aea29c4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:57:06 crc kubenswrapper[4754]: I1005 20:57:06.305127 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 20:57:06 crc kubenswrapper[4754]: E1005 20:57:06.305349 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 20:57:06.805315985 +0000 UTC m=+150.709434695 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:57:06 crc kubenswrapper[4754]: I1005 20:57:06.305677 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5sqj9\" (UID: \"c64d57c5-4bd4-48b5-85fc-9d613aea29c4\") " pod="openshift-image-registry/image-registry-697d97f7c8-5sqj9" Oct 05 20:57:06 crc kubenswrapper[4754]: E1005 20:57:06.306055 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 20:57:06.806045924 +0000 UTC m=+150.710164634 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5sqj9" (UID: "c64d57c5-4bd4-48b5-85fc-9d613aea29c4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:57:06 crc kubenswrapper[4754]: I1005 20:57:06.352680 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-hjpzm" event={"ID":"4da215ae-75ec-4ab1-81b0-39966d449b5f","Type":"ContainerStarted","Data":"454e02f8d9c7f6b24fcd759ad04ed608311127cf2dc258e7601fb53a11fa6472"} Oct 05 20:57:06 crc kubenswrapper[4754]: I1005 20:57:06.358215 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-wz2dg" event={"ID":"cce92b6a-69b0-49d5-bb66-36349ca96841","Type":"ContainerStarted","Data":"7aa5f0e52bcbe9a2c296300d13938ee28fab48e369a1d5de5e543eef8cbbdae4"} Oct 05 20:57:06 crc kubenswrapper[4754]: I1005 20:57:06.358269 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-wz2dg" event={"ID":"cce92b6a-69b0-49d5-bb66-36349ca96841","Type":"ContainerStarted","Data":"9856cd28aa3c2b1cec871a26acc0bfa5cc61cd5fb868b0e286747e514ed9bb07"} Oct 05 20:57:06 crc kubenswrapper[4754]: I1005 20:57:06.364065 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-5wx56" event={"ID":"d1b2b973-8b95-4f6a-83f1-468f84fb006d","Type":"ContainerStarted","Data":"0892e895648f77b7097bdb344500c3c3de436762f365028cf720337e54067b2c"} Oct 05 20:57:06 crc kubenswrapper[4754]: I1005 20:57:06.374191 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-cwtlx" event={"ID":"b157fa21-ae67-4c98-806c-93854b072965","Type":"ContainerStarted","Data":"10168e352f22daac00a7476c31cc1ce24673ae70ad2f3901429f4e4c8d8ff63c"} Oct 05 20:57:06 crc kubenswrapper[4754]: I1005 20:57:06.374310 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-cwtlx" Oct 05 20:57:06 crc kubenswrapper[4754]: I1005 20:57:06.375848 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-w6krf" event={"ID":"e542cd22-7a8b-4a4d-8205-d3dc60b2b8c6","Type":"ContainerStarted","Data":"00dbabb10931e4aeca1f262a58dd6550cdc94ecaa51155a003e411cc2486b179"} Oct 05 20:57:06 crc kubenswrapper[4754]: I1005 20:57:06.380240 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-wz2dg" podStartSLOduration=128.380228852 podStartE2EDuration="2m8.380228852s" podCreationTimestamp="2025-10-05 20:54:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 20:57:06.378304621 +0000 UTC m=+150.282423331" watchObservedRunningTime="2025-10-05 20:57:06.380228852 +0000 UTC m=+150.284347562" Oct 05 20:57:06 crc kubenswrapper[4754]: I1005 20:57:06.381723 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lxp7v" event={"ID":"d3c66e2a-8404-498c-b8a9-2bd2e6fd318b","Type":"ContainerStarted","Data":"4ea05210f3fa74df68d91ce86084eb60456b72b365097c1e645e7825691e2e44"} Oct 05 20:57:06 crc kubenswrapper[4754]: I1005 20:57:06.385367 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-682zv" event={"ID":"51547f65-61cc-4504-8399-5d540160fea1","Type":"ContainerStarted","Data":"e1f88eed66817bee63ac317b5482773755573da9cb2d6478d550cc732d98a9b2"} Oct 05 20:57:06 crc kubenswrapper[4754]: I1005 20:57:06.385409 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-682zv" event={"ID":"51547f65-61cc-4504-8399-5d540160fea1","Type":"ContainerStarted","Data":"7f02782f51a6187efa2ddd5fbffeb71d90315e60ef8f77fa0e091ececd35ebee"} Oct 05 20:57:06 crc kubenswrapper[4754]: I1005 20:57:06.395896 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-67bjd" event={"ID":"0593c91b-48f4-47aa-a57f-f9b50d7f385e","Type":"ContainerStarted","Data":"b5b78d1c33a715ab3f7053a0b4b028610e9107efa8cdb40c6ca7a91241007b71"} Oct 05 20:57:06 crc kubenswrapper[4754]: I1005 20:57:06.398116 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6wnmn" event={"ID":"9263c4f4-1540-4190-9e85-b1d8e2ca6c2f","Type":"ContainerStarted","Data":"40625f5fe234b6f1e30531df6e1d03aefffcf6cd198c42b8ece49bb8290a1ad9"} Oct 05 20:57:06 crc kubenswrapper[4754]: I1005 20:57:06.408015 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 20:57:06 crc kubenswrapper[4754]: E1005 20:57:06.408351 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 20:57:06.908330973 +0000 UTC m=+150.812449683 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:57:06 crc kubenswrapper[4754]: I1005 20:57:06.450657 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-2qmwc" event={"ID":"bc8bf6c1-363b-4e24-830c-898e4f7c6325","Type":"ContainerStarted","Data":"2abaa35ca89eae87b7eeedc4677c17b1382db3d5d93279fbe09950637d58c108"} Oct 05 20:57:06 crc kubenswrapper[4754]: I1005 20:57:06.452328 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-5wx56" podStartSLOduration=128.452315474 podStartE2EDuration="2m8.452315474s" podCreationTimestamp="2025-10-05 20:54:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 20:57:06.451578984 +0000 UTC m=+150.355697694" watchObservedRunningTime="2025-10-05 20:57:06.452315474 +0000 UTC m=+150.356434184" Oct 05 20:57:06 crc kubenswrapper[4754]: I1005 20:57:06.452459 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-cwtlx" podStartSLOduration=128.452455338 podStartE2EDuration="2m8.452455338s" podCreationTimestamp="2025-10-05 20:54:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 20:57:06.423274568 +0000 UTC m=+150.327393278" watchObservedRunningTime="2025-10-05 20:57:06.452455338 +0000 UTC m=+150.356574048" Oct 05 20:57:06 crc kubenswrapper[4754]: I1005 20:57:06.476847 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-chv7x" event={"ID":"8797a357-3b55-4d63-9215-4c64bb676a3d","Type":"ContainerStarted","Data":"c9623262d8c7a1b3d92ac55715d7782188710ba5dae6ad415f21544dbc43d8c6"} Oct 05 20:57:06 crc kubenswrapper[4754]: I1005 20:57:06.491770 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-nrt9t" event={"ID":"0b4c539c-22cf-4337-9959-afcc73e64370","Type":"ContainerStarted","Data":"2c3b92f98a7e947c68e3f386f443e051e97c06f8103dbb5e7ab7a500e666c1c4"} Oct 05 20:57:06 crc kubenswrapper[4754]: I1005 20:57:06.511427 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5sqj9\" (UID: \"c64d57c5-4bd4-48b5-85fc-9d613aea29c4\") " pod="openshift-image-registry/image-registry-697d97f7c8-5sqj9" Oct 05 20:57:06 crc kubenswrapper[4754]: E1005 20:57:06.515235 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 20:57:07.015214464 +0000 UTC m=+150.919333174 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5sqj9" (UID: "c64d57c5-4bd4-48b5-85fc-9d613aea29c4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:57:06 crc kubenswrapper[4754]: I1005 20:57:06.573440 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-w6krf" podStartSLOduration=128.57342145 podStartE2EDuration="2m8.57342145s" podCreationTimestamp="2025-10-05 20:54:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 20:57:06.531881354 +0000 UTC m=+150.436000064" watchObservedRunningTime="2025-10-05 20:57:06.57342145 +0000 UTC m=+150.477540160" Oct 05 20:57:06 crc kubenswrapper[4754]: I1005 20:57:06.574119 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-9vndv" event={"ID":"b96341ba-da55-442f-851e-5666717a1d5a","Type":"ContainerStarted","Data":"dc79623a22887ebad43771eee1907119658b785768fa9a666da308e66db1d372"} Oct 05 20:57:06 crc kubenswrapper[4754]: I1005 20:57:06.574900 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-682zv" podStartSLOduration=128.574894419 podStartE2EDuration="2m8.574894419s" podCreationTimestamp="2025-10-05 20:54:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 20:57:06.566845856 +0000 UTC m=+150.470964566" watchObservedRunningTime="2025-10-05 20:57:06.574894419 +0000 UTC m=+150.479013129" Oct 05 20:57:06 crc kubenswrapper[4754]: I1005 20:57:06.597100 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-ll5jq" event={"ID":"c1c4fc78-c305-4dba-b4d5-9f05d58a8ef6","Type":"ContainerStarted","Data":"2bc17b5eff1449d472fbd2fb6a61eaa7f22cf8a370a4c5c170b1d3dc3ee2ef1d"} Oct 05 20:57:06 crc kubenswrapper[4754]: I1005 20:57:06.598295 4754 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-ll5jq container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.22:8080/healthz\": dial tcp 10.217.0.22:8080: connect: connection refused" start-of-body= Oct 05 20:57:06 crc kubenswrapper[4754]: I1005 20:57:06.598338 4754 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-ll5jq" podUID="c1c4fc78-c305-4dba-b4d5-9f05d58a8ef6" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.22:8080/healthz\": dial tcp 10.217.0.22:8080: connect: connection refused" Oct 05 20:57:06 crc kubenswrapper[4754]: I1005 20:57:06.617082 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 20:57:06 crc kubenswrapper[4754]: I1005 20:57:06.617572 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 20:57:06 crc kubenswrapper[4754]: I1005 20:57:06.617600 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 20:57:06 crc kubenswrapper[4754]: I1005 20:57:06.617621 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 20:57:06 crc kubenswrapper[4754]: I1005 20:57:06.617654 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 20:57:06 crc kubenswrapper[4754]: I1005 20:57:06.625746 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 20:57:06 crc kubenswrapper[4754]: I1005 20:57:06.628340 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-7g427" event={"ID":"230b53bd-7c59-42f5-9604-0c5414fcdf80","Type":"ContainerStarted","Data":"0fdc3c8ffea443dd0b6e1e590bd6aceb1e16f5b413a41f7a90f5f977b46b2fea"} Oct 05 20:57:06 crc kubenswrapper[4754]: I1005 20:57:06.654975 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 20:57:06 crc kubenswrapper[4754]: I1005 20:57:06.655134 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 20:57:06 crc kubenswrapper[4754]: I1005 20:57:06.666346 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 20:57:06 crc kubenswrapper[4754]: E1005 20:57:06.683639 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 20:57:07.125821033 +0000 UTC m=+151.029939743 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:57:06 crc kubenswrapper[4754]: I1005 20:57:06.691070 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-6xm7j" event={"ID":"830a0a27-fab1-4efc-aeb3-e38a962ffcfa","Type":"ContainerStarted","Data":"b819c45e56f60a9db310383c00173a714cf780160477806917b24c9d143bd4ae"} Oct 05 20:57:06 crc kubenswrapper[4754]: I1005 20:57:06.691205 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-6xm7j" event={"ID":"830a0a27-fab1-4efc-aeb3-e38a962ffcfa","Type":"ContainerStarted","Data":"8cf5c98d0fde933545dc3b4d04953b499da370714eef1db4178a5d4f5bc06378"} Oct 05 20:57:06 crc kubenswrapper[4754]: I1005 20:57:06.691298 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-6xm7j" Oct 05 20:57:06 crc kubenswrapper[4754]: I1005 20:57:06.693030 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-67bjd" podStartSLOduration=128.693006116 podStartE2EDuration="2m8.693006116s" podCreationTimestamp="2025-10-05 20:54:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 20:57:06.616364173 +0000 UTC m=+150.520482883" watchObservedRunningTime="2025-10-05 20:57:06.693006116 +0000 UTC m=+150.597124826" Oct 05 20:57:06 crc kubenswrapper[4754]: I1005 20:57:06.697046 4754 patch_prober.go:28] interesting pod/downloads-7954f5f757-ksczt container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.20:8080/\": dial tcp 10.217.0.20:8080: connect: connection refused" start-of-body= Oct 05 20:57:06 crc kubenswrapper[4754]: I1005 20:57:06.697140 4754 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-ksczt" podUID="3934dd94-b9fa-4079-a268-60a9470dab41" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.20:8080/\": dial tcp 10.217.0.20:8080: connect: connection refused" Oct 05 20:57:06 crc kubenswrapper[4754]: I1005 20:57:06.702737 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lxp7v" podStartSLOduration=128.702726212 podStartE2EDuration="2m8.702726212s" podCreationTimestamp="2025-10-05 20:54:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 20:57:06.697960876 +0000 UTC m=+150.602079596" watchObservedRunningTime="2025-10-05 20:57:06.702726212 +0000 UTC m=+150.606844922" Oct 05 20:57:06 crc kubenswrapper[4754]: I1005 20:57:06.725439 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5sqj9\" (UID: \"c64d57c5-4bd4-48b5-85fc-9d613aea29c4\") " pod="openshift-image-registry/image-registry-697d97f7c8-5sqj9" Oct 05 20:57:06 crc kubenswrapper[4754]: I1005 20:57:06.743159 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jpbfr" Oct 05 20:57:06 crc kubenswrapper[4754]: I1005 20:57:06.743954 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nrd6b" Oct 05 20:57:06 crc kubenswrapper[4754]: E1005 20:57:06.744186 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 20:57:07.244164776 +0000 UTC m=+151.148283486 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5sqj9" (UID: "c64d57c5-4bd4-48b5-85fc-9d613aea29c4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:57:06 crc kubenswrapper[4754]: I1005 20:57:06.826123 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 20:57:06 crc kubenswrapper[4754]: E1005 20:57:06.828179 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 20:57:07.328146882 +0000 UTC m=+151.232265592 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:57:06 crc kubenswrapper[4754]: I1005 20:57:06.844997 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6wnmn" podStartSLOduration=128.844976306 podStartE2EDuration="2m8.844976306s" podCreationTimestamp="2025-10-05 20:54:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 20:57:06.844176455 +0000 UTC m=+150.748295165" watchObservedRunningTime="2025-10-05 20:57:06.844976306 +0000 UTC m=+150.749095016" Oct 05 20:57:06 crc kubenswrapper[4754]: I1005 20:57:06.859940 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 20:57:06 crc kubenswrapper[4754]: I1005 20:57:06.869323 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 20:57:06 crc kubenswrapper[4754]: I1005 20:57:06.874268 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 20:57:06 crc kubenswrapper[4754]: I1005 20:57:06.904957 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-lv842" Oct 05 20:57:06 crc kubenswrapper[4754]: I1005 20:57:06.928776 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5sqj9\" (UID: \"c64d57c5-4bd4-48b5-85fc-9d613aea29c4\") " pod="openshift-image-registry/image-registry-697d97f7c8-5sqj9" Oct 05 20:57:06 crc kubenswrapper[4754]: E1005 20:57:06.929261 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 20:57:07.42924656 +0000 UTC m=+151.333365270 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5sqj9" (UID: "c64d57c5-4bd4-48b5-85fc-9d613aea29c4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:57:07 crc kubenswrapper[4754]: I1005 20:57:07.008771 4754 patch_prober.go:28] interesting pod/router-default-5444994796-ll9lb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 05 20:57:07 crc kubenswrapper[4754]: [-]has-synced failed: reason withheld Oct 05 20:57:07 crc kubenswrapper[4754]: [+]process-running ok Oct 05 20:57:07 crc kubenswrapper[4754]: healthz check failed Oct 05 20:57:07 crc kubenswrapper[4754]: I1005 20:57:07.008834 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ll9lb" podUID="ebf750fb-f47c-420b-931d-3ac0a23e9e7b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 05 20:57:07 crc kubenswrapper[4754]: I1005 20:57:07.024141 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-2qmwc" podStartSLOduration=129.024115233 podStartE2EDuration="2m9.024115233s" podCreationTimestamp="2025-10-05 20:54:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 20:57:07.016952164 +0000 UTC m=+150.921070884" watchObservedRunningTime="2025-10-05 20:57:07.024115233 +0000 UTC m=+150.928233933" Oct 05 20:57:07 crc kubenswrapper[4754]: I1005 20:57:07.029774 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 20:57:07 crc kubenswrapper[4754]: E1005 20:57:07.030206 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 20:57:07.530187224 +0000 UTC m=+151.434305934 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:57:07 crc kubenswrapper[4754]: I1005 20:57:07.135577 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5sqj9\" (UID: \"c64d57c5-4bd4-48b5-85fc-9d613aea29c4\") " pod="openshift-image-registry/image-registry-697d97f7c8-5sqj9" Oct 05 20:57:07 crc kubenswrapper[4754]: E1005 20:57:07.135919 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 20:57:07.635904443 +0000 UTC m=+151.540023153 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5sqj9" (UID: "c64d57c5-4bd4-48b5-85fc-9d613aea29c4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:57:07 crc kubenswrapper[4754]: I1005 20:57:07.238274 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 20:57:07 crc kubenswrapper[4754]: E1005 20:57:07.238790 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 20:57:07.738772788 +0000 UTC m=+151.642891498 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:57:07 crc kubenswrapper[4754]: I1005 20:57:07.289319 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-9vndv" podStartSLOduration=129.289294881 podStartE2EDuration="2m9.289294881s" podCreationTimestamp="2025-10-05 20:54:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 20:57:07.252662084 +0000 UTC m=+151.156780794" watchObservedRunningTime="2025-10-05 20:57:07.289294881 +0000 UTC m=+151.193413591" Oct 05 20:57:07 crc kubenswrapper[4754]: I1005 20:57:07.292293 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-zwh6t"] Oct 05 20:57:07 crc kubenswrapper[4754]: I1005 20:57:07.293818 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zwh6t" Oct 05 20:57:07 crc kubenswrapper[4754]: I1005 20:57:07.303395 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-xknjb"] Oct 05 20:57:07 crc kubenswrapper[4754]: I1005 20:57:07.304858 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xknjb" Oct 05 20:57:07 crc kubenswrapper[4754]: I1005 20:57:07.316516 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Oct 05 20:57:07 crc kubenswrapper[4754]: I1005 20:57:07.322870 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Oct 05 20:57:07 crc kubenswrapper[4754]: I1005 20:57:07.341019 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xknjb"] Oct 05 20:57:07 crc kubenswrapper[4754]: I1005 20:57:07.352720 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5sqj9\" (UID: \"c64d57c5-4bd4-48b5-85fc-9d613aea29c4\") " pod="openshift-image-registry/image-registry-697d97f7c8-5sqj9" Oct 05 20:57:07 crc kubenswrapper[4754]: E1005 20:57:07.353205 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 20:57:07.853190757 +0000 UTC m=+151.757309467 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5sqj9" (UID: "c64d57c5-4bd4-48b5-85fc-9d613aea29c4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:57:07 crc kubenswrapper[4754]: I1005 20:57:07.388885 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zwh6t"] Oct 05 20:57:07 crc kubenswrapper[4754]: I1005 20:57:07.453508 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 20:57:07 crc kubenswrapper[4754]: I1005 20:57:07.453907 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p86kl\" (UniqueName: \"kubernetes.io/projected/b5b25be1-b393-4d98-aaf0-3f9693ba89d7-kube-api-access-p86kl\") pod \"certified-operators-xknjb\" (UID: \"b5b25be1-b393-4d98-aaf0-3f9693ba89d7\") " pod="openshift-marketplace/certified-operators-xknjb" Oct 05 20:57:07 crc kubenswrapper[4754]: I1005 20:57:07.453933 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2ngdc\" (UniqueName: \"kubernetes.io/projected/85d11cf6-caa4-4ee2-a8ee-48c44469d4b7-kube-api-access-2ngdc\") pod \"community-operators-zwh6t\" (UID: \"85d11cf6-caa4-4ee2-a8ee-48c44469d4b7\") " pod="openshift-marketplace/community-operators-zwh6t" Oct 05 20:57:07 crc kubenswrapper[4754]: I1005 20:57:07.453956 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b5b25be1-b393-4d98-aaf0-3f9693ba89d7-utilities\") pod \"certified-operators-xknjb\" (UID: \"b5b25be1-b393-4d98-aaf0-3f9693ba89d7\") " pod="openshift-marketplace/certified-operators-xknjb" Oct 05 20:57:07 crc kubenswrapper[4754]: I1005 20:57:07.453985 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b5b25be1-b393-4d98-aaf0-3f9693ba89d7-catalog-content\") pod \"certified-operators-xknjb\" (UID: \"b5b25be1-b393-4d98-aaf0-3f9693ba89d7\") " pod="openshift-marketplace/certified-operators-xknjb" Oct 05 20:57:07 crc kubenswrapper[4754]: I1005 20:57:07.454024 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/85d11cf6-caa4-4ee2-a8ee-48c44469d4b7-utilities\") pod \"community-operators-zwh6t\" (UID: \"85d11cf6-caa4-4ee2-a8ee-48c44469d4b7\") " pod="openshift-marketplace/community-operators-zwh6t" Oct 05 20:57:07 crc kubenswrapper[4754]: I1005 20:57:07.454044 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/85d11cf6-caa4-4ee2-a8ee-48c44469d4b7-catalog-content\") pod \"community-operators-zwh6t\" (UID: \"85d11cf6-caa4-4ee2-a8ee-48c44469d4b7\") " pod="openshift-marketplace/community-operators-zwh6t" Oct 05 20:57:07 crc kubenswrapper[4754]: E1005 20:57:07.454191 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 20:57:07.954175292 +0000 UTC m=+151.858294002 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:57:07 crc kubenswrapper[4754]: I1005 20:57:07.482416 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-5qjvg"] Oct 05 20:57:07 crc kubenswrapper[4754]: I1005 20:57:07.483766 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5qjvg" Oct 05 20:57:07 crc kubenswrapper[4754]: I1005 20:57:07.562136 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p86kl\" (UniqueName: \"kubernetes.io/projected/b5b25be1-b393-4d98-aaf0-3f9693ba89d7-kube-api-access-p86kl\") pod \"certified-operators-xknjb\" (UID: \"b5b25be1-b393-4d98-aaf0-3f9693ba89d7\") " pod="openshift-marketplace/certified-operators-xknjb" Oct 05 20:57:07 crc kubenswrapper[4754]: I1005 20:57:07.562173 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2ngdc\" (UniqueName: \"kubernetes.io/projected/85d11cf6-caa4-4ee2-a8ee-48c44469d4b7-kube-api-access-2ngdc\") pod \"community-operators-zwh6t\" (UID: \"85d11cf6-caa4-4ee2-a8ee-48c44469d4b7\") " pod="openshift-marketplace/community-operators-zwh6t" Oct 05 20:57:07 crc kubenswrapper[4754]: I1005 20:57:07.562194 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b5b25be1-b393-4d98-aaf0-3f9693ba89d7-utilities\") pod \"certified-operators-xknjb\" (UID: \"b5b25be1-b393-4d98-aaf0-3f9693ba89d7\") " pod="openshift-marketplace/certified-operators-xknjb" Oct 05 20:57:07 crc kubenswrapper[4754]: I1005 20:57:07.562216 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b5b25be1-b393-4d98-aaf0-3f9693ba89d7-catalog-content\") pod \"certified-operators-xknjb\" (UID: \"b5b25be1-b393-4d98-aaf0-3f9693ba89d7\") " pod="openshift-marketplace/certified-operators-xknjb" Oct 05 20:57:07 crc kubenswrapper[4754]: I1005 20:57:07.562255 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/85d11cf6-caa4-4ee2-a8ee-48c44469d4b7-utilities\") pod \"community-operators-zwh6t\" (UID: \"85d11cf6-caa4-4ee2-a8ee-48c44469d4b7\") " pod="openshift-marketplace/community-operators-zwh6t" Oct 05 20:57:07 crc kubenswrapper[4754]: I1005 20:57:07.562272 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/85d11cf6-caa4-4ee2-a8ee-48c44469d4b7-catalog-content\") pod \"community-operators-zwh6t\" (UID: \"85d11cf6-caa4-4ee2-a8ee-48c44469d4b7\") " pod="openshift-marketplace/community-operators-zwh6t" Oct 05 20:57:07 crc kubenswrapper[4754]: I1005 20:57:07.562307 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5sqj9\" (UID: \"c64d57c5-4bd4-48b5-85fc-9d613aea29c4\") " pod="openshift-image-registry/image-registry-697d97f7c8-5sqj9" Oct 05 20:57:07 crc kubenswrapper[4754]: E1005 20:57:07.562653 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 20:57:08.062640505 +0000 UTC m=+151.966759215 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5sqj9" (UID: "c64d57c5-4bd4-48b5-85fc-9d613aea29c4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:57:07 crc kubenswrapper[4754]: I1005 20:57:07.563193 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b5b25be1-b393-4d98-aaf0-3f9693ba89d7-catalog-content\") pod \"certified-operators-xknjb\" (UID: \"b5b25be1-b393-4d98-aaf0-3f9693ba89d7\") " pod="openshift-marketplace/certified-operators-xknjb" Oct 05 20:57:07 crc kubenswrapper[4754]: I1005 20:57:07.563538 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/85d11cf6-caa4-4ee2-a8ee-48c44469d4b7-utilities\") pod \"community-operators-zwh6t\" (UID: \"85d11cf6-caa4-4ee2-a8ee-48c44469d4b7\") " pod="openshift-marketplace/community-operators-zwh6t" Oct 05 20:57:07 crc kubenswrapper[4754]: I1005 20:57:07.563781 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/85d11cf6-caa4-4ee2-a8ee-48c44469d4b7-catalog-content\") pod \"community-operators-zwh6t\" (UID: \"85d11cf6-caa4-4ee2-a8ee-48c44469d4b7\") " pod="openshift-marketplace/community-operators-zwh6t" Oct 05 20:57:07 crc kubenswrapper[4754]: I1005 20:57:07.563798 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b5b25be1-b393-4d98-aaf0-3f9693ba89d7-utilities\") pod \"certified-operators-xknjb\" (UID: \"b5b25be1-b393-4d98-aaf0-3f9693ba89d7\") " pod="openshift-marketplace/certified-operators-xknjb" Oct 05 20:57:07 crc kubenswrapper[4754]: I1005 20:57:07.564558 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-5qjvg"] Oct 05 20:57:07 crc kubenswrapper[4754]: I1005 20:57:07.659127 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p86kl\" (UniqueName: \"kubernetes.io/projected/b5b25be1-b393-4d98-aaf0-3f9693ba89d7-kube-api-access-p86kl\") pod \"certified-operators-xknjb\" (UID: \"b5b25be1-b393-4d98-aaf0-3f9693ba89d7\") " pod="openshift-marketplace/certified-operators-xknjb" Oct 05 20:57:07 crc kubenswrapper[4754]: I1005 20:57:07.663137 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 20:57:07 crc kubenswrapper[4754]: I1005 20:57:07.663426 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d5e4795f-6f22-4cde-a939-970f1ec20119-catalog-content\") pod \"community-operators-5qjvg\" (UID: \"d5e4795f-6f22-4cde-a939-970f1ec20119\") " pod="openshift-marketplace/community-operators-5qjvg" Oct 05 20:57:07 crc kubenswrapper[4754]: I1005 20:57:07.663524 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h8xtz\" (UniqueName: \"kubernetes.io/projected/d5e4795f-6f22-4cde-a939-970f1ec20119-kube-api-access-h8xtz\") pod \"community-operators-5qjvg\" (UID: \"d5e4795f-6f22-4cde-a939-970f1ec20119\") " pod="openshift-marketplace/community-operators-5qjvg" Oct 05 20:57:07 crc kubenswrapper[4754]: I1005 20:57:07.663577 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d5e4795f-6f22-4cde-a939-970f1ec20119-utilities\") pod \"community-operators-5qjvg\" (UID: \"d5e4795f-6f22-4cde-a939-970f1ec20119\") " pod="openshift-marketplace/community-operators-5qjvg" Oct 05 20:57:07 crc kubenswrapper[4754]: E1005 20:57:07.663702 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 20:57:08.163678561 +0000 UTC m=+152.067797271 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:57:07 crc kubenswrapper[4754]: I1005 20:57:07.668576 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2ngdc\" (UniqueName: \"kubernetes.io/projected/85d11cf6-caa4-4ee2-a8ee-48c44469d4b7-kube-api-access-2ngdc\") pod \"community-operators-zwh6t\" (UID: \"85d11cf6-caa4-4ee2-a8ee-48c44469d4b7\") " pod="openshift-marketplace/community-operators-zwh6t" Oct 05 20:57:07 crc kubenswrapper[4754]: I1005 20:57:07.688848 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xknjb" Oct 05 20:57:07 crc kubenswrapper[4754]: I1005 20:57:07.748287 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-59nwx"] Oct 05 20:57:07 crc kubenswrapper[4754]: I1005 20:57:07.756094 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-s4zph" event={"ID":"5a5b1574-7f76-4bcf-8c24-9ff7773e86f2","Type":"ContainerStarted","Data":"8287a81de7d9e74a166188284ab3ae3643632aec2e17fe5f9c956058c4c7dda2"} Oct 05 20:57:07 crc kubenswrapper[4754]: I1005 20:57:07.759405 4754 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-ll5jq container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.22:8080/healthz\": dial tcp 10.217.0.22:8080: connect: connection refused" start-of-body= Oct 05 20:57:07 crc kubenswrapper[4754]: I1005 20:57:07.759527 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-59nwx" Oct 05 20:57:07 crc kubenswrapper[4754]: I1005 20:57:07.759523 4754 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-ll5jq" podUID="c1c4fc78-c305-4dba-b4d5-9f05d58a8ef6" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.22:8080/healthz\": dial tcp 10.217.0.22:8080: connect: connection refused" Oct 05 20:57:07 crc kubenswrapper[4754]: I1005 20:57:07.760041 4754 patch_prober.go:28] interesting pod/downloads-7954f5f757-ksczt container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.20:8080/\": dial tcp 10.217.0.20:8080: connect: connection refused" start-of-body= Oct 05 20:57:07 crc kubenswrapper[4754]: I1005 20:57:07.760095 4754 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-ksczt" podUID="3934dd94-b9fa-4079-a268-60a9470dab41" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.20:8080/\": dial tcp 10.217.0.20:8080: connect: connection refused" Oct 05 20:57:07 crc kubenswrapper[4754]: I1005 20:57:07.782877 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h8xtz\" (UniqueName: \"kubernetes.io/projected/d5e4795f-6f22-4cde-a939-970f1ec20119-kube-api-access-h8xtz\") pod \"community-operators-5qjvg\" (UID: \"d5e4795f-6f22-4cde-a939-970f1ec20119\") " pod="openshift-marketplace/community-operators-5qjvg" Oct 05 20:57:07 crc kubenswrapper[4754]: I1005 20:57:07.782966 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d5e4795f-6f22-4cde-a939-970f1ec20119-utilities\") pod \"community-operators-5qjvg\" (UID: \"d5e4795f-6f22-4cde-a939-970f1ec20119\") " pod="openshift-marketplace/community-operators-5qjvg" Oct 05 20:57:07 crc kubenswrapper[4754]: I1005 20:57:07.783020 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5sqj9\" (UID: \"c64d57c5-4bd4-48b5-85fc-9d613aea29c4\") " pod="openshift-image-registry/image-registry-697d97f7c8-5sqj9" Oct 05 20:57:07 crc kubenswrapper[4754]: I1005 20:57:07.783051 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d5e4795f-6f22-4cde-a939-970f1ec20119-catalog-content\") pod \"community-operators-5qjvg\" (UID: \"d5e4795f-6f22-4cde-a939-970f1ec20119\") " pod="openshift-marketplace/community-operators-5qjvg" Oct 05 20:57:07 crc kubenswrapper[4754]: I1005 20:57:07.784048 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d5e4795f-6f22-4cde-a939-970f1ec20119-catalog-content\") pod \"community-operators-5qjvg\" (UID: \"d5e4795f-6f22-4cde-a939-970f1ec20119\") " pod="openshift-marketplace/community-operators-5qjvg" Oct 05 20:57:07 crc kubenswrapper[4754]: E1005 20:57:07.786943 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 20:57:08.286922413 +0000 UTC m=+152.191041123 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5sqj9" (UID: "c64d57c5-4bd4-48b5-85fc-9d613aea29c4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:57:07 crc kubenswrapper[4754]: I1005 20:57:07.798804 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d5e4795f-6f22-4cde-a939-970f1ec20119-utilities\") pod \"community-operators-5qjvg\" (UID: \"d5e4795f-6f22-4cde-a939-970f1ec20119\") " pod="openshift-marketplace/community-operators-5qjvg" Oct 05 20:57:07 crc kubenswrapper[4754]: I1005 20:57:07.848623 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-6xm7j" podStartSLOduration=11.848591211 podStartE2EDuration="11.848591211s" podCreationTimestamp="2025-10-05 20:56:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 20:57:07.721709212 +0000 UTC m=+151.625827922" watchObservedRunningTime="2025-10-05 20:57:07.848591211 +0000 UTC m=+151.752709921" Oct 05 20:57:07 crc kubenswrapper[4754]: I1005 20:57:07.885127 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 20:57:07 crc kubenswrapper[4754]: I1005 20:57:07.885573 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t7p85\" (UniqueName: \"kubernetes.io/projected/232ff34c-dc36-4012-9038-e10014a179cd-kube-api-access-t7p85\") pod \"certified-operators-59nwx\" (UID: \"232ff34c-dc36-4012-9038-e10014a179cd\") " pod="openshift-marketplace/certified-operators-59nwx" Oct 05 20:57:07 crc kubenswrapper[4754]: I1005 20:57:07.886093 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/232ff34c-dc36-4012-9038-e10014a179cd-catalog-content\") pod \"certified-operators-59nwx\" (UID: \"232ff34c-dc36-4012-9038-e10014a179cd\") " pod="openshift-marketplace/certified-operators-59nwx" Oct 05 20:57:07 crc kubenswrapper[4754]: I1005 20:57:07.886286 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/232ff34c-dc36-4012-9038-e10014a179cd-utilities\") pod \"certified-operators-59nwx\" (UID: \"232ff34c-dc36-4012-9038-e10014a179cd\") " pod="openshift-marketplace/certified-operators-59nwx" Oct 05 20:57:07 crc kubenswrapper[4754]: E1005 20:57:07.892946 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 20:57:08.392925951 +0000 UTC m=+152.297044661 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:57:07 crc kubenswrapper[4754]: I1005 20:57:07.908350 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-59nwx"] Oct 05 20:57:07 crc kubenswrapper[4754]: I1005 20:57:07.943864 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zwh6t" Oct 05 20:57:07 crc kubenswrapper[4754]: I1005 20:57:07.946887 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h8xtz\" (UniqueName: \"kubernetes.io/projected/d5e4795f-6f22-4cde-a939-970f1ec20119-kube-api-access-h8xtz\") pod \"community-operators-5qjvg\" (UID: \"d5e4795f-6f22-4cde-a939-970f1ec20119\") " pod="openshift-marketplace/community-operators-5qjvg" Oct 05 20:57:08 crc kubenswrapper[4754]: I1005 20:57:07.998851 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5sqj9\" (UID: \"c64d57c5-4bd4-48b5-85fc-9d613aea29c4\") " pod="openshift-image-registry/image-registry-697d97f7c8-5sqj9" Oct 05 20:57:08 crc kubenswrapper[4754]: I1005 20:57:07.998920 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/232ff34c-dc36-4012-9038-e10014a179cd-catalog-content\") pod \"certified-operators-59nwx\" (UID: \"232ff34c-dc36-4012-9038-e10014a179cd\") " pod="openshift-marketplace/certified-operators-59nwx" Oct 05 20:57:08 crc kubenswrapper[4754]: I1005 20:57:07.999132 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/232ff34c-dc36-4012-9038-e10014a179cd-utilities\") pod \"certified-operators-59nwx\" (UID: \"232ff34c-dc36-4012-9038-e10014a179cd\") " pod="openshift-marketplace/certified-operators-59nwx" Oct 05 20:57:08 crc kubenswrapper[4754]: E1005 20:57:08.000234 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 20:57:08.500200291 +0000 UTC m=+152.404319011 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5sqj9" (UID: "c64d57c5-4bd4-48b5-85fc-9d613aea29c4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:57:08 crc kubenswrapper[4754]: I1005 20:57:08.001974 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t7p85\" (UniqueName: \"kubernetes.io/projected/232ff34c-dc36-4012-9038-e10014a179cd-kube-api-access-t7p85\") pod \"certified-operators-59nwx\" (UID: \"232ff34c-dc36-4012-9038-e10014a179cd\") " pod="openshift-marketplace/certified-operators-59nwx" Oct 05 20:57:08 crc kubenswrapper[4754]: I1005 20:57:08.005630 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/232ff34c-dc36-4012-9038-e10014a179cd-utilities\") pod \"certified-operators-59nwx\" (UID: \"232ff34c-dc36-4012-9038-e10014a179cd\") " pod="openshift-marketplace/certified-operators-59nwx" Oct 05 20:57:08 crc kubenswrapper[4754]: I1005 20:57:08.006523 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/232ff34c-dc36-4012-9038-e10014a179cd-catalog-content\") pod \"certified-operators-59nwx\" (UID: \"232ff34c-dc36-4012-9038-e10014a179cd\") " pod="openshift-marketplace/certified-operators-59nwx" Oct 05 20:57:08 crc kubenswrapper[4754]: I1005 20:57:08.030795 4754 patch_prober.go:28] interesting pod/router-default-5444994796-ll9lb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 05 20:57:08 crc kubenswrapper[4754]: [-]has-synced failed: reason withheld Oct 05 20:57:08 crc kubenswrapper[4754]: [+]process-running ok Oct 05 20:57:08 crc kubenswrapper[4754]: healthz check failed Oct 05 20:57:08 crc kubenswrapper[4754]: I1005 20:57:08.030873 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ll9lb" podUID="ebf750fb-f47c-420b-931d-3ac0a23e9e7b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 05 20:57:08 crc kubenswrapper[4754]: I1005 20:57:08.062960 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t7p85\" (UniqueName: \"kubernetes.io/projected/232ff34c-dc36-4012-9038-e10014a179cd-kube-api-access-t7p85\") pod \"certified-operators-59nwx\" (UID: \"232ff34c-dc36-4012-9038-e10014a179cd\") " pod="openshift-marketplace/certified-operators-59nwx" Oct 05 20:57:08 crc kubenswrapper[4754]: I1005 20:57:08.090854 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-59nwx" Oct 05 20:57:08 crc kubenswrapper[4754]: I1005 20:57:08.104649 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 20:57:08 crc kubenswrapper[4754]: E1005 20:57:08.105019 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 20:57:08.604997846 +0000 UTC m=+152.509116556 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:57:08 crc kubenswrapper[4754]: I1005 20:57:08.128993 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5qjvg" Oct 05 20:57:08 crc kubenswrapper[4754]: I1005 20:57:08.211910 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5sqj9\" (UID: \"c64d57c5-4bd4-48b5-85fc-9d613aea29c4\") " pod="openshift-image-registry/image-registry-697d97f7c8-5sqj9" Oct 05 20:57:08 crc kubenswrapper[4754]: E1005 20:57:08.212334 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 20:57:08.712320108 +0000 UTC m=+152.616438818 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5sqj9" (UID: "c64d57c5-4bd4-48b5-85fc-9d613aea29c4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:57:08 crc kubenswrapper[4754]: I1005 20:57:08.313177 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 20:57:08 crc kubenswrapper[4754]: E1005 20:57:08.313784 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 20:57:08.813761205 +0000 UTC m=+152.717879915 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:57:08 crc kubenswrapper[4754]: I1005 20:57:08.418293 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5sqj9\" (UID: \"c64d57c5-4bd4-48b5-85fc-9d613aea29c4\") " pod="openshift-image-registry/image-registry-697d97f7c8-5sqj9" Oct 05 20:57:08 crc kubenswrapper[4754]: E1005 20:57:08.418650 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 20:57:08.918637853 +0000 UTC m=+152.822756563 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5sqj9" (UID: "c64d57c5-4bd4-48b5-85fc-9d613aea29c4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:57:08 crc kubenswrapper[4754]: I1005 20:57:08.470108 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-5sffm" Oct 05 20:57:08 crc kubenswrapper[4754]: I1005 20:57:08.471434 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-5sffm" Oct 05 20:57:08 crc kubenswrapper[4754]: I1005 20:57:08.478642 4754 patch_prober.go:28] interesting pod/console-f9d7485db-5sffm container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.9:8443/health\": dial tcp 10.217.0.9:8443: connect: connection refused" start-of-body= Oct 05 20:57:08 crc kubenswrapper[4754]: I1005 20:57:08.478692 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-5sffm" podUID="c56f5eb8-3a12-428c-b347-78369dc19008" containerName="console" probeResult="failure" output="Get \"https://10.217.0.9:8443/health\": dial tcp 10.217.0.9:8443: connect: connection refused" Oct 05 20:57:08 crc kubenswrapper[4754]: I1005 20:57:08.519060 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 20:57:08 crc kubenswrapper[4754]: E1005 20:57:08.519627 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 20:57:09.019592667 +0000 UTC m=+152.923711377 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:57:08 crc kubenswrapper[4754]: I1005 20:57:08.519815 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5sqj9\" (UID: \"c64d57c5-4bd4-48b5-85fc-9d613aea29c4\") " pod="openshift-image-registry/image-registry-697d97f7c8-5sqj9" Oct 05 20:57:08 crc kubenswrapper[4754]: E1005 20:57:08.520162 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 20:57:09.020149712 +0000 UTC m=+152.924268422 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5sqj9" (UID: "c64d57c5-4bd4-48b5-85fc-9d613aea29c4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:57:08 crc kubenswrapper[4754]: I1005 20:57:08.625035 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 20:57:08 crc kubenswrapper[4754]: E1005 20:57:08.625408 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 20:57:09.125372349 +0000 UTC m=+153.029491059 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:57:08 crc kubenswrapper[4754]: I1005 20:57:08.625767 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5sqj9\" (UID: \"c64d57c5-4bd4-48b5-85fc-9d613aea29c4\") " pod="openshift-image-registry/image-registry-697d97f7c8-5sqj9" Oct 05 20:57:08 crc kubenswrapper[4754]: E1005 20:57:08.626786 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 20:57:09.126778356 +0000 UTC m=+153.030897056 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5sqj9" (UID: "c64d57c5-4bd4-48b5-85fc-9d613aea29c4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:57:08 crc kubenswrapper[4754]: I1005 20:57:08.642422 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-k8qff" Oct 05 20:57:08 crc kubenswrapper[4754]: I1005 20:57:08.642465 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-k8qff" Oct 05 20:57:08 crc kubenswrapper[4754]: I1005 20:57:08.674852 4754 patch_prober.go:28] interesting pod/apiserver-76f77b778f-k8qff container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Oct 05 20:57:08 crc kubenswrapper[4754]: [+]log ok Oct 05 20:57:08 crc kubenswrapper[4754]: [+]etcd ok Oct 05 20:57:08 crc kubenswrapper[4754]: [+]poststarthook/start-apiserver-admission-initializer ok Oct 05 20:57:08 crc kubenswrapper[4754]: [+]poststarthook/generic-apiserver-start-informers ok Oct 05 20:57:08 crc kubenswrapper[4754]: [+]poststarthook/max-in-flight-filter ok Oct 05 20:57:08 crc kubenswrapper[4754]: [+]poststarthook/storage-object-count-tracker-hook ok Oct 05 20:57:08 crc kubenswrapper[4754]: [+]poststarthook/image.openshift.io-apiserver-caches ok Oct 05 20:57:08 crc kubenswrapper[4754]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Oct 05 20:57:08 crc kubenswrapper[4754]: [+]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa ok Oct 05 20:57:08 crc kubenswrapper[4754]: [+]poststarthook/project.openshift.io-projectcache ok Oct 05 20:57:08 crc kubenswrapper[4754]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Oct 05 20:57:08 crc kubenswrapper[4754]: [+]poststarthook/openshift.io-startinformers ok Oct 05 20:57:08 crc kubenswrapper[4754]: [+]poststarthook/openshift.io-restmapperupdater ok Oct 05 20:57:08 crc kubenswrapper[4754]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Oct 05 20:57:08 crc kubenswrapper[4754]: livez check failed Oct 05 20:57:08 crc kubenswrapper[4754]: I1005 20:57:08.674928 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-k8qff" podUID="20a5d042-ffe2-4244-b25d-700b26534a04" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 05 20:57:08 crc kubenswrapper[4754]: I1005 20:57:08.729065 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 20:57:08 crc kubenswrapper[4754]: E1005 20:57:08.729430 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 20:57:09.229411524 +0000 UTC m=+153.133530224 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:57:08 crc kubenswrapper[4754]: I1005 20:57:08.755941 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-cwtlx" Oct 05 20:57:08 crc kubenswrapper[4754]: I1005 20:57:08.835567 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5sqj9\" (UID: \"c64d57c5-4bd4-48b5-85fc-9d613aea29c4\") " pod="openshift-image-registry/image-registry-697d97f7c8-5sqj9" Oct 05 20:57:08 crc kubenswrapper[4754]: E1005 20:57:08.838161 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 20:57:09.338142033 +0000 UTC m=+153.242260733 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5sqj9" (UID: "c64d57c5-4bd4-48b5-85fc-9d613aea29c4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:57:08 crc kubenswrapper[4754]: I1005 20:57:08.917247 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-s4zph" event={"ID":"5a5b1574-7f76-4bcf-8c24-9ff7773e86f2","Type":"ContainerStarted","Data":"19039b0e27f041d6cf634e40aade03ed453f067a56b4b4bb1ce213b1e58db2af"} Oct 05 20:57:08 crc kubenswrapper[4754]: I1005 20:57:08.917294 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"30425480a409ea9e03ed4da4b3932341e628482322539f468a64781b92d388d7"} Oct 05 20:57:08 crc kubenswrapper[4754]: I1005 20:57:08.923614 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"469f26ec15dbfdebfa4a2177e13671d52bb20b6a3385526748a893b65e466423"} Oct 05 20:57:08 crc kubenswrapper[4754]: I1005 20:57:08.936001 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"e80e94b7bb31f6fb5cce911e10cb5caf1bba2f850d093adfae7ae541eb9763e1"} Oct 05 20:57:08 crc kubenswrapper[4754]: I1005 20:57:08.938050 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 20:57:08 crc kubenswrapper[4754]: E1005 20:57:08.938126 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 20:57:09.438103781 +0000 UTC m=+153.342222491 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:57:08 crc kubenswrapper[4754]: I1005 20:57:08.939597 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5sqj9\" (UID: \"c64d57c5-4bd4-48b5-85fc-9d613aea29c4\") " pod="openshift-image-registry/image-registry-697d97f7c8-5sqj9" Oct 05 20:57:08 crc kubenswrapper[4754]: E1005 20:57:08.940142 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 20:57:09.440124975 +0000 UTC m=+153.344243685 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5sqj9" (UID: "c64d57c5-4bd4-48b5-85fc-9d613aea29c4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:57:09 crc kubenswrapper[4754]: I1005 20:57:08.996381 4754 patch_prober.go:28] interesting pod/router-default-5444994796-ll9lb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 05 20:57:09 crc kubenswrapper[4754]: [-]has-synced failed: reason withheld Oct 05 20:57:09 crc kubenswrapper[4754]: [+]process-running ok Oct 05 20:57:09 crc kubenswrapper[4754]: healthz check failed Oct 05 20:57:09 crc kubenswrapper[4754]: I1005 20:57:08.996455 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ll9lb" podUID="ebf750fb-f47c-420b-931d-3ac0a23e9e7b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 05 20:57:09 crc kubenswrapper[4754]: I1005 20:57:09.041855 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 20:57:09 crc kubenswrapper[4754]: E1005 20:57:09.042955 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 20:57:09.542935538 +0000 UTC m=+153.447054248 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:57:09 crc kubenswrapper[4754]: I1005 20:57:09.143327 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5sqj9\" (UID: \"c64d57c5-4bd4-48b5-85fc-9d613aea29c4\") " pod="openshift-image-registry/image-registry-697d97f7c8-5sqj9" Oct 05 20:57:09 crc kubenswrapper[4754]: E1005 20:57:09.143817 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 20:57:09.643797279 +0000 UTC m=+153.547915979 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5sqj9" (UID: "c64d57c5-4bd4-48b5-85fc-9d613aea29c4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:57:09 crc kubenswrapper[4754]: I1005 20:57:09.165386 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zwh6t"] Oct 05 20:57:09 crc kubenswrapper[4754]: I1005 20:57:09.244724 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 20:57:09 crc kubenswrapper[4754]: E1005 20:57:09.245573 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 20:57:09.745536434 +0000 UTC m=+153.649655144 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:57:09 crc kubenswrapper[4754]: I1005 20:57:09.278388 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xknjb"] Oct 05 20:57:09 crc kubenswrapper[4754]: I1005 20:57:09.346409 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5sqj9\" (UID: \"c64d57c5-4bd4-48b5-85fc-9d613aea29c4\") " pod="openshift-image-registry/image-registry-697d97f7c8-5sqj9" Oct 05 20:57:09 crc kubenswrapper[4754]: E1005 20:57:09.346823 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 20:57:09.846808867 +0000 UTC m=+153.750927577 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5sqj9" (UID: "c64d57c5-4bd4-48b5-85fc-9d613aea29c4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:57:09 crc kubenswrapper[4754]: I1005 20:57:09.374885 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-5889x"] Oct 05 20:57:09 crc kubenswrapper[4754]: I1005 20:57:09.375869 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-59nwx"] Oct 05 20:57:09 crc kubenswrapper[4754]: I1005 20:57:09.375980 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5889x" Oct 05 20:57:09 crc kubenswrapper[4754]: I1005 20:57:09.396656 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Oct 05 20:57:09 crc kubenswrapper[4754]: I1005 20:57:09.398306 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5889x"] Oct 05 20:57:09 crc kubenswrapper[4754]: I1005 20:57:09.448180 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 20:57:09 crc kubenswrapper[4754]: I1005 20:57:09.448348 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-68wh5\" (UniqueName: \"kubernetes.io/projected/61cebf9f-dd0d-40f9-9bd5-1ff0dd7038f8-kube-api-access-68wh5\") pod \"redhat-marketplace-5889x\" (UID: \"61cebf9f-dd0d-40f9-9bd5-1ff0dd7038f8\") " pod="openshift-marketplace/redhat-marketplace-5889x" Oct 05 20:57:09 crc kubenswrapper[4754]: I1005 20:57:09.448432 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/61cebf9f-dd0d-40f9-9bd5-1ff0dd7038f8-utilities\") pod \"redhat-marketplace-5889x\" (UID: \"61cebf9f-dd0d-40f9-9bd5-1ff0dd7038f8\") " pod="openshift-marketplace/redhat-marketplace-5889x" Oct 05 20:57:09 crc kubenswrapper[4754]: I1005 20:57:09.448456 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/61cebf9f-dd0d-40f9-9bd5-1ff0dd7038f8-catalog-content\") pod \"redhat-marketplace-5889x\" (UID: \"61cebf9f-dd0d-40f9-9bd5-1ff0dd7038f8\") " pod="openshift-marketplace/redhat-marketplace-5889x" Oct 05 20:57:09 crc kubenswrapper[4754]: E1005 20:57:09.448586 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 20:57:09.948565582 +0000 UTC m=+153.852684302 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:57:09 crc kubenswrapper[4754]: W1005 20:57:09.456219 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod232ff34c_dc36_4012_9038_e10014a179cd.slice/crio-f2262d22304f374a26601238d2879127c3c5921e6d1ca7f4ce13c6b08b7174b9 WatchSource:0}: Error finding container f2262d22304f374a26601238d2879127c3c5921e6d1ca7f4ce13c6b08b7174b9: Status 404 returned error can't find the container with id f2262d22304f374a26601238d2879127c3c5921e6d1ca7f4ce13c6b08b7174b9 Oct 05 20:57:09 crc kubenswrapper[4754]: I1005 20:57:09.530451 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-5qjvg"] Oct 05 20:57:09 crc kubenswrapper[4754]: I1005 20:57:09.550209 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5sqj9\" (UID: \"c64d57c5-4bd4-48b5-85fc-9d613aea29c4\") " pod="openshift-image-registry/image-registry-697d97f7c8-5sqj9" Oct 05 20:57:09 crc kubenswrapper[4754]: I1005 20:57:09.550257 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/61cebf9f-dd0d-40f9-9bd5-1ff0dd7038f8-utilities\") pod \"redhat-marketplace-5889x\" (UID: \"61cebf9f-dd0d-40f9-9bd5-1ff0dd7038f8\") " pod="openshift-marketplace/redhat-marketplace-5889x" Oct 05 20:57:09 crc kubenswrapper[4754]: I1005 20:57:09.550279 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/61cebf9f-dd0d-40f9-9bd5-1ff0dd7038f8-catalog-content\") pod \"redhat-marketplace-5889x\" (UID: \"61cebf9f-dd0d-40f9-9bd5-1ff0dd7038f8\") " pod="openshift-marketplace/redhat-marketplace-5889x" Oct 05 20:57:09 crc kubenswrapper[4754]: I1005 20:57:09.550310 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-68wh5\" (UniqueName: \"kubernetes.io/projected/61cebf9f-dd0d-40f9-9bd5-1ff0dd7038f8-kube-api-access-68wh5\") pod \"redhat-marketplace-5889x\" (UID: \"61cebf9f-dd0d-40f9-9bd5-1ff0dd7038f8\") " pod="openshift-marketplace/redhat-marketplace-5889x" Oct 05 20:57:09 crc kubenswrapper[4754]: E1005 20:57:09.551100 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 20:57:10.051087668 +0000 UTC m=+153.955206378 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5sqj9" (UID: "c64d57c5-4bd4-48b5-85fc-9d613aea29c4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:57:09 crc kubenswrapper[4754]: I1005 20:57:09.551608 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/61cebf9f-dd0d-40f9-9bd5-1ff0dd7038f8-utilities\") pod \"redhat-marketplace-5889x\" (UID: \"61cebf9f-dd0d-40f9-9bd5-1ff0dd7038f8\") " pod="openshift-marketplace/redhat-marketplace-5889x" Oct 05 20:57:09 crc kubenswrapper[4754]: I1005 20:57:09.551853 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/61cebf9f-dd0d-40f9-9bd5-1ff0dd7038f8-catalog-content\") pod \"redhat-marketplace-5889x\" (UID: \"61cebf9f-dd0d-40f9-9bd5-1ff0dd7038f8\") " pod="openshift-marketplace/redhat-marketplace-5889x" Oct 05 20:57:09 crc kubenswrapper[4754]: I1005 20:57:09.602902 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-68wh5\" (UniqueName: \"kubernetes.io/projected/61cebf9f-dd0d-40f9-9bd5-1ff0dd7038f8-kube-api-access-68wh5\") pod \"redhat-marketplace-5889x\" (UID: \"61cebf9f-dd0d-40f9-9bd5-1ff0dd7038f8\") " pod="openshift-marketplace/redhat-marketplace-5889x" Oct 05 20:57:09 crc kubenswrapper[4754]: I1005 20:57:09.654507 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 20:57:09 crc kubenswrapper[4754]: E1005 20:57:09.654866 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 20:57:10.154848516 +0000 UTC m=+154.058967226 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:57:09 crc kubenswrapper[4754]: I1005 20:57:09.708845 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lxp7v" Oct 05 20:57:09 crc kubenswrapper[4754]: I1005 20:57:09.708905 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lxp7v" Oct 05 20:57:09 crc kubenswrapper[4754]: I1005 20:57:09.721422 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lxp7v" Oct 05 20:57:09 crc kubenswrapper[4754]: I1005 20:57:09.738400 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5889x" Oct 05 20:57:09 crc kubenswrapper[4754]: I1005 20:57:09.757214 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5sqj9\" (UID: \"c64d57c5-4bd4-48b5-85fc-9d613aea29c4\") " pod="openshift-image-registry/image-registry-697d97f7c8-5sqj9" Oct 05 20:57:09 crc kubenswrapper[4754]: E1005 20:57:09.757702 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 20:57:10.2576855 +0000 UTC m=+154.161804210 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5sqj9" (UID: "c64d57c5-4bd4-48b5-85fc-9d613aea29c4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:57:09 crc kubenswrapper[4754]: I1005 20:57:09.767903 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-l2w5p"] Oct 05 20:57:09 crc kubenswrapper[4754]: I1005 20:57:09.769003 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-l2w5p" Oct 05 20:57:09 crc kubenswrapper[4754]: I1005 20:57:09.780129 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-l2w5p"] Oct 05 20:57:09 crc kubenswrapper[4754]: I1005 20:57:09.790554 4754 patch_prober.go:28] interesting pod/downloads-7954f5f757-ksczt container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.20:8080/\": dial tcp 10.217.0.20:8080: connect: connection refused" start-of-body= Oct 05 20:57:09 crc kubenswrapper[4754]: I1005 20:57:09.790595 4754 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-ksczt" podUID="3934dd94-b9fa-4079-a268-60a9470dab41" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.20:8080/\": dial tcp 10.217.0.20:8080: connect: connection refused" Oct 05 20:57:09 crc kubenswrapper[4754]: I1005 20:57:09.790645 4754 patch_prober.go:28] interesting pod/downloads-7954f5f757-ksczt container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.20:8080/\": dial tcp 10.217.0.20:8080: connect: connection refused" start-of-body= Oct 05 20:57:09 crc kubenswrapper[4754]: I1005 20:57:09.790658 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-ksczt" podUID="3934dd94-b9fa-4079-a268-60a9470dab41" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.20:8080/\": dial tcp 10.217.0.20:8080: connect: connection refused" Oct 05 20:57:09 crc kubenswrapper[4754]: I1005 20:57:09.858467 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 20:57:09 crc kubenswrapper[4754]: E1005 20:57:09.858706 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 20:57:10.358671365 +0000 UTC m=+154.262790075 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:57:09 crc kubenswrapper[4754]: I1005 20:57:09.858865 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/50c8cfa3-71a9-4c5e-895c-f15457183666-catalog-content\") pod \"redhat-marketplace-l2w5p\" (UID: \"50c8cfa3-71a9-4c5e-895c-f15457183666\") " pod="openshift-marketplace/redhat-marketplace-l2w5p" Oct 05 20:57:09 crc kubenswrapper[4754]: I1005 20:57:09.858925 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5sqj9\" (UID: \"c64d57c5-4bd4-48b5-85fc-9d613aea29c4\") " pod="openshift-image-registry/image-registry-697d97f7c8-5sqj9" Oct 05 20:57:09 crc kubenswrapper[4754]: I1005 20:57:09.858984 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6wq48\" (UniqueName: \"kubernetes.io/projected/50c8cfa3-71a9-4c5e-895c-f15457183666-kube-api-access-6wq48\") pod \"redhat-marketplace-l2w5p\" (UID: \"50c8cfa3-71a9-4c5e-895c-f15457183666\") " pod="openshift-marketplace/redhat-marketplace-l2w5p" Oct 05 20:57:09 crc kubenswrapper[4754]: I1005 20:57:09.859060 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/50c8cfa3-71a9-4c5e-895c-f15457183666-utilities\") pod \"redhat-marketplace-l2w5p\" (UID: \"50c8cfa3-71a9-4c5e-895c-f15457183666\") " pod="openshift-marketplace/redhat-marketplace-l2w5p" Oct 05 20:57:09 crc kubenswrapper[4754]: E1005 20:57:09.861118 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 20:57:10.361103579 +0000 UTC m=+154.265222289 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5sqj9" (UID: "c64d57c5-4bd4-48b5-85fc-9d613aea29c4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:57:09 crc kubenswrapper[4754]: I1005 20:57:09.964224 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 20:57:09 crc kubenswrapper[4754]: I1005 20:57:09.965185 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/50c8cfa3-71a9-4c5e-895c-f15457183666-catalog-content\") pod \"redhat-marketplace-l2w5p\" (UID: \"50c8cfa3-71a9-4c5e-895c-f15457183666\") " pod="openshift-marketplace/redhat-marketplace-l2w5p" Oct 05 20:57:09 crc kubenswrapper[4754]: I1005 20:57:09.965230 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6wq48\" (UniqueName: \"kubernetes.io/projected/50c8cfa3-71a9-4c5e-895c-f15457183666-kube-api-access-6wq48\") pod \"redhat-marketplace-l2w5p\" (UID: \"50c8cfa3-71a9-4c5e-895c-f15457183666\") " pod="openshift-marketplace/redhat-marketplace-l2w5p" Oct 05 20:57:09 crc kubenswrapper[4754]: I1005 20:57:09.965264 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/50c8cfa3-71a9-4c5e-895c-f15457183666-utilities\") pod \"redhat-marketplace-l2w5p\" (UID: \"50c8cfa3-71a9-4c5e-895c-f15457183666\") " pod="openshift-marketplace/redhat-marketplace-l2w5p" Oct 05 20:57:09 crc kubenswrapper[4754]: I1005 20:57:09.965720 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/50c8cfa3-71a9-4c5e-895c-f15457183666-utilities\") pod \"redhat-marketplace-l2w5p\" (UID: \"50c8cfa3-71a9-4c5e-895c-f15457183666\") " pod="openshift-marketplace/redhat-marketplace-l2w5p" Oct 05 20:57:09 crc kubenswrapper[4754]: E1005 20:57:09.965803 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 20:57:10.465785721 +0000 UTC m=+154.369904431 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:57:09 crc kubenswrapper[4754]: I1005 20:57:09.965995 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/50c8cfa3-71a9-4c5e-895c-f15457183666-catalog-content\") pod \"redhat-marketplace-l2w5p\" (UID: \"50c8cfa3-71a9-4c5e-895c-f15457183666\") " pod="openshift-marketplace/redhat-marketplace-l2w5p" Oct 05 20:57:09 crc kubenswrapper[4754]: I1005 20:57:09.985637 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-ll9lb" Oct 05 20:57:09 crc kubenswrapper[4754]: I1005 20:57:09.990872 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"b2b54d5159263b75b7182e5125cb57b23acddbd0179b42c030f1a455df804d21"} Oct 05 20:57:09 crc kubenswrapper[4754]: I1005 20:57:09.994545 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-ll5jq" Oct 05 20:57:09 crc kubenswrapper[4754]: I1005 20:57:09.994785 4754 patch_prober.go:28] interesting pod/router-default-5444994796-ll9lb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 05 20:57:09 crc kubenswrapper[4754]: [-]has-synced failed: reason withheld Oct 05 20:57:09 crc kubenswrapper[4754]: [+]process-running ok Oct 05 20:57:09 crc kubenswrapper[4754]: healthz check failed Oct 05 20:57:09 crc kubenswrapper[4754]: I1005 20:57:09.994851 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ll9lb" podUID="ebf750fb-f47c-420b-931d-3ac0a23e9e7b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 05 20:57:09 crc kubenswrapper[4754]: I1005 20:57:09.995197 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"eceef32d344784a3436a1ccdc5609e36e624e96e7114721faa975a6a73eae7d5"} Oct 05 20:57:09 crc kubenswrapper[4754]: I1005 20:57:09.995553 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 20:57:09 crc kubenswrapper[4754]: I1005 20:57:09.997479 4754 generic.go:334] "Generic (PLEG): container finished" podID="85d11cf6-caa4-4ee2-a8ee-48c44469d4b7" containerID="ed64ccc8ba22eb586a997140263b753574a123b89d48f5dbace62ef8f13c71a5" exitCode=0 Oct 05 20:57:09 crc kubenswrapper[4754]: I1005 20:57:09.998598 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zwh6t" event={"ID":"85d11cf6-caa4-4ee2-a8ee-48c44469d4b7","Type":"ContainerDied","Data":"ed64ccc8ba22eb586a997140263b753574a123b89d48f5dbace62ef8f13c71a5"} Oct 05 20:57:09 crc kubenswrapper[4754]: I1005 20:57:09.998649 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zwh6t" event={"ID":"85d11cf6-caa4-4ee2-a8ee-48c44469d4b7","Type":"ContainerStarted","Data":"7e2db30679f9bb3068ecaf9bb3b848ee881c5daac364bc6c5a62aeb4a367136f"} Oct 05 20:57:10 crc kubenswrapper[4754]: I1005 20:57:10.006256 4754 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 05 20:57:10 crc kubenswrapper[4754]: I1005 20:57:10.007255 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6wq48\" (UniqueName: \"kubernetes.io/projected/50c8cfa3-71a9-4c5e-895c-f15457183666-kube-api-access-6wq48\") pod \"redhat-marketplace-l2w5p\" (UID: \"50c8cfa3-71a9-4c5e-895c-f15457183666\") " pod="openshift-marketplace/redhat-marketplace-l2w5p" Oct 05 20:57:10 crc kubenswrapper[4754]: I1005 20:57:10.009887 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-59nwx" event={"ID":"232ff34c-dc36-4012-9038-e10014a179cd","Type":"ContainerStarted","Data":"f2262d22304f374a26601238d2879127c3c5921e6d1ca7f4ce13c6b08b7174b9"} Oct 05 20:57:10 crc kubenswrapper[4754]: I1005 20:57:10.044374 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"be1886f4a6abdd1d6d6b8dcfee479b4785ddd78a890db8c863c7c82b303fb73d"} Oct 05 20:57:10 crc kubenswrapper[4754]: I1005 20:57:10.064530 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5qjvg" event={"ID":"d5e4795f-6f22-4cde-a939-970f1ec20119","Type":"ContainerStarted","Data":"fffbb9b87fd079439336591f5d8cf694f7d51b2dfaf8d3b0f883da2b3b59c443"} Oct 05 20:57:10 crc kubenswrapper[4754]: I1005 20:57:10.083551 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5sqj9\" (UID: \"c64d57c5-4bd4-48b5-85fc-9d613aea29c4\") " pod="openshift-image-registry/image-registry-697d97f7c8-5sqj9" Oct 05 20:57:10 crc kubenswrapper[4754]: E1005 20:57:10.089950 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 20:57:10.589928897 +0000 UTC m=+154.494047607 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5sqj9" (UID: "c64d57c5-4bd4-48b5-85fc-9d613aea29c4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:57:10 crc kubenswrapper[4754]: I1005 20:57:10.124944 4754 generic.go:334] "Generic (PLEG): container finished" podID="b5b25be1-b393-4d98-aaf0-3f9693ba89d7" containerID="8457a90a27ace934a18c57e5a39bc1fa61bf28f1cf604f816967a9d1e074208c" exitCode=0 Oct 05 20:57:10 crc kubenswrapper[4754]: I1005 20:57:10.125136 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xknjb" event={"ID":"b5b25be1-b393-4d98-aaf0-3f9693ba89d7","Type":"ContainerDied","Data":"8457a90a27ace934a18c57e5a39bc1fa61bf28f1cf604f816967a9d1e074208c"} Oct 05 20:57:10 crc kubenswrapper[4754]: I1005 20:57:10.125190 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xknjb" event={"ID":"b5b25be1-b393-4d98-aaf0-3f9693ba89d7","Type":"ContainerStarted","Data":"2502a9d5b335e202508425ca7b4c43fee5b6e09a4bacb5ae9a9b9d3d07139080"} Oct 05 20:57:10 crc kubenswrapper[4754]: I1005 20:57:10.153083 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-s4zph" event={"ID":"5a5b1574-7f76-4bcf-8c24-9ff7773e86f2","Type":"ContainerStarted","Data":"a1274fd6ef39b1e71cc5f624557794be6da15072d2f5670ccf6db3de15212ed7"} Oct 05 20:57:10 crc kubenswrapper[4754]: I1005 20:57:10.162084 4754 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Oct 05 20:57:10 crc kubenswrapper[4754]: I1005 20:57:10.183692 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lxp7v" Oct 05 20:57:10 crc kubenswrapper[4754]: I1005 20:57:10.184905 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 20:57:10 crc kubenswrapper[4754]: E1005 20:57:10.186364 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 20:57:10.686348452 +0000 UTC m=+154.590467162 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:57:10 crc kubenswrapper[4754]: I1005 20:57:10.202854 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-xdqrj"] Oct 05 20:57:10 crc kubenswrapper[4754]: I1005 20:57:10.204020 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xdqrj" Oct 05 20:57:10 crc kubenswrapper[4754]: I1005 20:57:10.207905 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Oct 05 20:57:10 crc kubenswrapper[4754]: I1005 20:57:10.238954 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-l2w5p" Oct 05 20:57:10 crc kubenswrapper[4754]: I1005 20:57:10.243830 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xdqrj"] Oct 05 20:57:10 crc kubenswrapper[4754]: I1005 20:57:10.261164 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5889x"] Oct 05 20:57:10 crc kubenswrapper[4754]: I1005 20:57:10.292268 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/caf99098-6640-4077-8b97-69d1bfefa99c-catalog-content\") pod \"redhat-operators-xdqrj\" (UID: \"caf99098-6640-4077-8b97-69d1bfefa99c\") " pod="openshift-marketplace/redhat-operators-xdqrj" Oct 05 20:57:10 crc kubenswrapper[4754]: I1005 20:57:10.292301 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/caf99098-6640-4077-8b97-69d1bfefa99c-utilities\") pod \"redhat-operators-xdqrj\" (UID: \"caf99098-6640-4077-8b97-69d1bfefa99c\") " pod="openshift-marketplace/redhat-operators-xdqrj" Oct 05 20:57:10 crc kubenswrapper[4754]: I1005 20:57:10.292413 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9dgdm\" (UniqueName: \"kubernetes.io/projected/caf99098-6640-4077-8b97-69d1bfefa99c-kube-api-access-9dgdm\") pod \"redhat-operators-xdqrj\" (UID: \"caf99098-6640-4077-8b97-69d1bfefa99c\") " pod="openshift-marketplace/redhat-operators-xdqrj" Oct 05 20:57:10 crc kubenswrapper[4754]: I1005 20:57:10.292443 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5sqj9\" (UID: \"c64d57c5-4bd4-48b5-85fc-9d613aea29c4\") " pod="openshift-image-registry/image-registry-697d97f7c8-5sqj9" Oct 05 20:57:10 crc kubenswrapper[4754]: E1005 20:57:10.305868 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 20:57:10.805848605 +0000 UTC m=+154.709967315 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5sqj9" (UID: "c64d57c5-4bd4-48b5-85fc-9d613aea29c4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:57:10 crc kubenswrapper[4754]: I1005 20:57:10.389944 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-nz2vw"] Oct 05 20:57:10 crc kubenswrapper[4754]: I1005 20:57:10.391838 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nz2vw" Oct 05 20:57:10 crc kubenswrapper[4754]: I1005 20:57:10.405529 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 20:57:10 crc kubenswrapper[4754]: I1005 20:57:10.405856 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9dgdm\" (UniqueName: \"kubernetes.io/projected/caf99098-6640-4077-8b97-69d1bfefa99c-kube-api-access-9dgdm\") pod \"redhat-operators-xdqrj\" (UID: \"caf99098-6640-4077-8b97-69d1bfefa99c\") " pod="openshift-marketplace/redhat-operators-xdqrj" Oct 05 20:57:10 crc kubenswrapper[4754]: I1005 20:57:10.405938 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/caf99098-6640-4077-8b97-69d1bfefa99c-catalog-content\") pod \"redhat-operators-xdqrj\" (UID: \"caf99098-6640-4077-8b97-69d1bfefa99c\") " pod="openshift-marketplace/redhat-operators-xdqrj" Oct 05 20:57:10 crc kubenswrapper[4754]: I1005 20:57:10.405978 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/caf99098-6640-4077-8b97-69d1bfefa99c-utilities\") pod \"redhat-operators-xdqrj\" (UID: \"caf99098-6640-4077-8b97-69d1bfefa99c\") " pod="openshift-marketplace/redhat-operators-xdqrj" Oct 05 20:57:10 crc kubenswrapper[4754]: E1005 20:57:10.406325 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 20:57:10.906303576 +0000 UTC m=+154.810422286 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:57:10 crc kubenswrapper[4754]: I1005 20:57:10.407755 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/caf99098-6640-4077-8b97-69d1bfefa99c-catalog-content\") pod \"redhat-operators-xdqrj\" (UID: \"caf99098-6640-4077-8b97-69d1bfefa99c\") " pod="openshift-marketplace/redhat-operators-xdqrj" Oct 05 20:57:10 crc kubenswrapper[4754]: I1005 20:57:10.408007 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/caf99098-6640-4077-8b97-69d1bfefa99c-utilities\") pod \"redhat-operators-xdqrj\" (UID: \"caf99098-6640-4077-8b97-69d1bfefa99c\") " pod="openshift-marketplace/redhat-operators-xdqrj" Oct 05 20:57:10 crc kubenswrapper[4754]: I1005 20:57:10.410278 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-nz2vw"] Oct 05 20:57:10 crc kubenswrapper[4754]: I1005 20:57:10.456042 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" Oct 05 20:57:10 crc kubenswrapper[4754]: I1005 20:57:10.497948 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9dgdm\" (UniqueName: \"kubernetes.io/projected/caf99098-6640-4077-8b97-69d1bfefa99c-kube-api-access-9dgdm\") pod \"redhat-operators-xdqrj\" (UID: \"caf99098-6640-4077-8b97-69d1bfefa99c\") " pod="openshift-marketplace/redhat-operators-xdqrj" Oct 05 20:57:10 crc kubenswrapper[4754]: I1005 20:57:10.508568 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5dff6e7-5693-4f7c-af40-d7b10cbb73c7-utilities\") pod \"redhat-operators-nz2vw\" (UID: \"e5dff6e7-5693-4f7c-af40-d7b10cbb73c7\") " pod="openshift-marketplace/redhat-operators-nz2vw" Oct 05 20:57:10 crc kubenswrapper[4754]: I1005 20:57:10.508671 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5dff6e7-5693-4f7c-af40-d7b10cbb73c7-catalog-content\") pod \"redhat-operators-nz2vw\" (UID: \"e5dff6e7-5693-4f7c-af40-d7b10cbb73c7\") " pod="openshift-marketplace/redhat-operators-nz2vw" Oct 05 20:57:10 crc kubenswrapper[4754]: I1005 20:57:10.508706 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5sqj9\" (UID: \"c64d57c5-4bd4-48b5-85fc-9d613aea29c4\") " pod="openshift-image-registry/image-registry-697d97f7c8-5sqj9" Oct 05 20:57:10 crc kubenswrapper[4754]: I1005 20:57:10.508733 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b5k8b\" (UniqueName: \"kubernetes.io/projected/e5dff6e7-5693-4f7c-af40-d7b10cbb73c7-kube-api-access-b5k8b\") pod \"redhat-operators-nz2vw\" (UID: \"e5dff6e7-5693-4f7c-af40-d7b10cbb73c7\") " pod="openshift-marketplace/redhat-operators-nz2vw" Oct 05 20:57:10 crc kubenswrapper[4754]: E1005 20:57:10.509128 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 20:57:11.009112489 +0000 UTC m=+154.913231199 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5sqj9" (UID: "c64d57c5-4bd4-48b5-85fc-9d613aea29c4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:57:10 crc kubenswrapper[4754]: I1005 20:57:10.579975 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xdqrj" Oct 05 20:57:10 crc kubenswrapper[4754]: I1005 20:57:10.611669 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 20:57:10 crc kubenswrapper[4754]: E1005 20:57:10.612032 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 20:57:11.111988714 +0000 UTC m=+155.016107414 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:57:10 crc kubenswrapper[4754]: I1005 20:57:10.612238 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5dff6e7-5693-4f7c-af40-d7b10cbb73c7-catalog-content\") pod \"redhat-operators-nz2vw\" (UID: \"e5dff6e7-5693-4f7c-af40-d7b10cbb73c7\") " pod="openshift-marketplace/redhat-operators-nz2vw" Oct 05 20:57:10 crc kubenswrapper[4754]: I1005 20:57:10.612370 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5sqj9\" (UID: \"c64d57c5-4bd4-48b5-85fc-9d613aea29c4\") " pod="openshift-image-registry/image-registry-697d97f7c8-5sqj9" Oct 05 20:57:10 crc kubenswrapper[4754]: I1005 20:57:10.612415 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b5k8b\" (UniqueName: \"kubernetes.io/projected/e5dff6e7-5693-4f7c-af40-d7b10cbb73c7-kube-api-access-b5k8b\") pod \"redhat-operators-nz2vw\" (UID: \"e5dff6e7-5693-4f7c-af40-d7b10cbb73c7\") " pod="openshift-marketplace/redhat-operators-nz2vw" Oct 05 20:57:10 crc kubenswrapper[4754]: I1005 20:57:10.612579 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5dff6e7-5693-4f7c-af40-d7b10cbb73c7-utilities\") pod \"redhat-operators-nz2vw\" (UID: \"e5dff6e7-5693-4f7c-af40-d7b10cbb73c7\") " pod="openshift-marketplace/redhat-operators-nz2vw" Oct 05 20:57:10 crc kubenswrapper[4754]: I1005 20:57:10.612985 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5dff6e7-5693-4f7c-af40-d7b10cbb73c7-catalog-content\") pod \"redhat-operators-nz2vw\" (UID: \"e5dff6e7-5693-4f7c-af40-d7b10cbb73c7\") " pod="openshift-marketplace/redhat-operators-nz2vw" Oct 05 20:57:10 crc kubenswrapper[4754]: I1005 20:57:10.613099 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5dff6e7-5693-4f7c-af40-d7b10cbb73c7-utilities\") pod \"redhat-operators-nz2vw\" (UID: \"e5dff6e7-5693-4f7c-af40-d7b10cbb73c7\") " pod="openshift-marketplace/redhat-operators-nz2vw" Oct 05 20:57:10 crc kubenswrapper[4754]: E1005 20:57:10.613289 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 20:57:11.113273108 +0000 UTC m=+155.017392028 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5sqj9" (UID: "c64d57c5-4bd4-48b5-85fc-9d613aea29c4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:57:10 crc kubenswrapper[4754]: I1005 20:57:10.680795 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b5k8b\" (UniqueName: \"kubernetes.io/projected/e5dff6e7-5693-4f7c-af40-d7b10cbb73c7-kube-api-access-b5k8b\") pod \"redhat-operators-nz2vw\" (UID: \"e5dff6e7-5693-4f7c-af40-d7b10cbb73c7\") " pod="openshift-marketplace/redhat-operators-nz2vw" Oct 05 20:57:10 crc kubenswrapper[4754]: I1005 20:57:10.714254 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 20:57:10 crc kubenswrapper[4754]: E1005 20:57:10.714734 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 20:57:11.214715015 +0000 UTC m=+155.118833725 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:57:10 crc kubenswrapper[4754]: I1005 20:57:10.758788 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nz2vw" Oct 05 20:57:10 crc kubenswrapper[4754]: I1005 20:57:10.815884 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5sqj9\" (UID: \"c64d57c5-4bd4-48b5-85fc-9d613aea29c4\") " pod="openshift-image-registry/image-registry-697d97f7c8-5sqj9" Oct 05 20:57:10 crc kubenswrapper[4754]: E1005 20:57:10.816313 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 20:57:11.316298696 +0000 UTC m=+155.220417396 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5sqj9" (UID: "c64d57c5-4bd4-48b5-85fc-9d613aea29c4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:57:10 crc kubenswrapper[4754]: I1005 20:57:10.919425 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 20:57:10 crc kubenswrapper[4754]: E1005 20:57:10.919776 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 20:57:11.419729875 +0000 UTC m=+155.323848585 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:57:10 crc kubenswrapper[4754]: I1005 20:57:10.920560 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5sqj9\" (UID: \"c64d57c5-4bd4-48b5-85fc-9d613aea29c4\") " pod="openshift-image-registry/image-registry-697d97f7c8-5sqj9" Oct 05 20:57:10 crc kubenswrapper[4754]: E1005 20:57:10.921276 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 20:57:11.421260286 +0000 UTC m=+155.325378996 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5sqj9" (UID: "c64d57c5-4bd4-48b5-85fc-9d613aea29c4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:57:11 crc kubenswrapper[4754]: I1005 20:57:11.011373 4754 patch_prober.go:28] interesting pod/router-default-5444994796-ll9lb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 05 20:57:11 crc kubenswrapper[4754]: [-]has-synced failed: reason withheld Oct 05 20:57:11 crc kubenswrapper[4754]: [+]process-running ok Oct 05 20:57:11 crc kubenswrapper[4754]: healthz check failed Oct 05 20:57:11 crc kubenswrapper[4754]: I1005 20:57:11.011790 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ll9lb" podUID="ebf750fb-f47c-420b-931d-3ac0a23e9e7b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 05 20:57:11 crc kubenswrapper[4754]: I1005 20:57:11.022683 4754 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-10-05T20:57:10.162109462Z","Handler":null,"Name":""} Oct 05 20:57:11 crc kubenswrapper[4754]: I1005 20:57:11.023479 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 20:57:11 crc kubenswrapper[4754]: E1005 20:57:11.024209 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 20:57:11.524185982 +0000 UTC m=+155.428304692 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 20:57:11 crc kubenswrapper[4754]: I1005 20:57:11.025262 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-l2w5p"] Oct 05 20:57:11 crc kubenswrapper[4754]: I1005 20:57:11.035256 4754 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Oct 05 20:57:11 crc kubenswrapper[4754]: I1005 20:57:11.035295 4754 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Oct 05 20:57:11 crc kubenswrapper[4754]: I1005 20:57:11.124150 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xdqrj"] Oct 05 20:57:11 crc kubenswrapper[4754]: I1005 20:57:11.124952 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5sqj9\" (UID: \"c64d57c5-4bd4-48b5-85fc-9d613aea29c4\") " pod="openshift-image-registry/image-registry-697d97f7c8-5sqj9" Oct 05 20:57:11 crc kubenswrapper[4754]: I1005 20:57:11.131299 4754 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 05 20:57:11 crc kubenswrapper[4754]: I1005 20:57:11.131333 4754 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5sqj9\" (UID: \"c64d57c5-4bd4-48b5-85fc-9d613aea29c4\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-5sqj9" Oct 05 20:57:11 crc kubenswrapper[4754]: I1005 20:57:11.199191 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5sqj9\" (UID: \"c64d57c5-4bd4-48b5-85fc-9d613aea29c4\") " pod="openshift-image-registry/image-registry-697d97f7c8-5sqj9" Oct 05 20:57:11 crc kubenswrapper[4754]: I1005 20:57:11.202277 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-nz2vw"] Oct 05 20:57:11 crc kubenswrapper[4754]: I1005 20:57:11.216536 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-s4zph" event={"ID":"5a5b1574-7f76-4bcf-8c24-9ff7773e86f2","Type":"ContainerStarted","Data":"60eebd2df0629c5f1082c11e7f3c485dbeee219721d90dce84f9b4e2e3455473"} Oct 05 20:57:11 crc kubenswrapper[4754]: I1005 20:57:11.219081 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xdqrj" event={"ID":"caf99098-6640-4077-8b97-69d1bfefa99c","Type":"ContainerStarted","Data":"96cfcf8172f5b44ba90e77feec2d1e3ee1393f09d878f09c63656a9aa222713b"} Oct 05 20:57:11 crc kubenswrapper[4754]: I1005 20:57:11.222620 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l2w5p" event={"ID":"50c8cfa3-71a9-4c5e-895c-f15457183666","Type":"ContainerStarted","Data":"5ef1db83196df95c7d5303633ad6f3f36fc6aaf19fd40f10bbbf747b851b55b2"} Oct 05 20:57:11 crc kubenswrapper[4754]: I1005 20:57:11.226572 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 20:57:11 crc kubenswrapper[4754]: I1005 20:57:11.229110 4754 generic.go:334] "Generic (PLEG): container finished" podID="61cebf9f-dd0d-40f9-9bd5-1ff0dd7038f8" containerID="31cf7e5dfda9011d8644759cb52e266ca933e48a1579f6fa5323b051d9a5f29e" exitCode=0 Oct 05 20:57:11 crc kubenswrapper[4754]: I1005 20:57:11.229200 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5889x" event={"ID":"61cebf9f-dd0d-40f9-9bd5-1ff0dd7038f8","Type":"ContainerDied","Data":"31cf7e5dfda9011d8644759cb52e266ca933e48a1579f6fa5323b051d9a5f29e"} Oct 05 20:57:11 crc kubenswrapper[4754]: I1005 20:57:11.229233 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5889x" event={"ID":"61cebf9f-dd0d-40f9-9bd5-1ff0dd7038f8","Type":"ContainerStarted","Data":"831673c15f0c69e2fd6dc94ca3d8411e53559dcd1108802081f757966b315067"} Oct 05 20:57:11 crc kubenswrapper[4754]: I1005 20:57:11.236654 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-s4zph" podStartSLOduration=15.236409942 podStartE2EDuration="15.236409942s" podCreationTimestamp="2025-10-05 20:56:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 20:57:11.236135935 +0000 UTC m=+155.140254645" watchObservedRunningTime="2025-10-05 20:57:11.236409942 +0000 UTC m=+155.140528652" Oct 05 20:57:11 crc kubenswrapper[4754]: I1005 20:57:11.239657 4754 generic.go:334] "Generic (PLEG): container finished" podID="232ff34c-dc36-4012-9038-e10014a179cd" containerID="d6e446ac94cd1146dd3cc09eeb151e991447a00f7179a8d9f675e578ac774d23" exitCode=0 Oct 05 20:57:11 crc kubenswrapper[4754]: I1005 20:57:11.239734 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-59nwx" event={"ID":"232ff34c-dc36-4012-9038-e10014a179cd","Type":"ContainerDied","Data":"d6e446ac94cd1146dd3cc09eeb151e991447a00f7179a8d9f675e578ac774d23"} Oct 05 20:57:11 crc kubenswrapper[4754]: I1005 20:57:11.241870 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 05 20:57:11 crc kubenswrapper[4754]: I1005 20:57:11.247022 4754 generic.go:334] "Generic (PLEG): container finished" podID="d5e4795f-6f22-4cde-a939-970f1ec20119" containerID="d76d2da9cc0c7659052d793be6ef3070a7f1d693f236cce3327c8a5e99e14728" exitCode=0 Oct 05 20:57:11 crc kubenswrapper[4754]: I1005 20:57:11.248097 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5qjvg" event={"ID":"d5e4795f-6f22-4cde-a939-970f1ec20119","Type":"ContainerDied","Data":"d76d2da9cc0c7659052d793be6ef3070a7f1d693f236cce3327c8a5e99e14728"} Oct 05 20:57:11 crc kubenswrapper[4754]: I1005 20:57:11.283396 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-5sqj9" Oct 05 20:57:11 crc kubenswrapper[4754]: I1005 20:57:11.893919 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-5sqj9"] Oct 05 20:57:11 crc kubenswrapper[4754]: W1005 20:57:11.922016 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc64d57c5_4bd4_48b5_85fc_9d613aea29c4.slice/crio-aa4820d4875b0b4449cb2c7a512e82f56a182188f58561773f130af26b6b0ab5 WatchSource:0}: Error finding container aa4820d4875b0b4449cb2c7a512e82f56a182188f58561773f130af26b6b0ab5: Status 404 returned error can't find the container with id aa4820d4875b0b4449cb2c7a512e82f56a182188f58561773f130af26b6b0ab5 Oct 05 20:57:12 crc kubenswrapper[4754]: I1005 20:57:12.006894 4754 patch_prober.go:28] interesting pod/router-default-5444994796-ll9lb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 05 20:57:12 crc kubenswrapper[4754]: [-]has-synced failed: reason withheld Oct 05 20:57:12 crc kubenswrapper[4754]: [+]process-running ok Oct 05 20:57:12 crc kubenswrapper[4754]: healthz check failed Oct 05 20:57:12 crc kubenswrapper[4754]: I1005 20:57:12.006971 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ll9lb" podUID="ebf750fb-f47c-420b-931d-3ac0a23e9e7b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 05 20:57:12 crc kubenswrapper[4754]: I1005 20:57:12.268966 4754 generic.go:334] "Generic (PLEG): container finished" podID="e5dff6e7-5693-4f7c-af40-d7b10cbb73c7" containerID="73f7374ca314aac1bcf71496b9da4b7cc906e5d71e6a1ac6141bb10f41cb6ef0" exitCode=0 Oct 05 20:57:12 crc kubenswrapper[4754]: I1005 20:57:12.269645 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nz2vw" event={"ID":"e5dff6e7-5693-4f7c-af40-d7b10cbb73c7","Type":"ContainerDied","Data":"73f7374ca314aac1bcf71496b9da4b7cc906e5d71e6a1ac6141bb10f41cb6ef0"} Oct 05 20:57:12 crc kubenswrapper[4754]: I1005 20:57:12.269710 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nz2vw" event={"ID":"e5dff6e7-5693-4f7c-af40-d7b10cbb73c7","Type":"ContainerStarted","Data":"2f986afacdcb56a29c5711c4a2666164a622c988d71e1075471e7fd9ee0bbda1"} Oct 05 20:57:12 crc kubenswrapper[4754]: I1005 20:57:12.280450 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-5sqj9" event={"ID":"c64d57c5-4bd4-48b5-85fc-9d613aea29c4","Type":"ContainerStarted","Data":"591cdabaa80ef7bd9d0e08d68cf15ca0ca3b3a4940c23598ecfc3cecf551244d"} Oct 05 20:57:12 crc kubenswrapper[4754]: I1005 20:57:12.280510 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-5sqj9" event={"ID":"c64d57c5-4bd4-48b5-85fc-9d613aea29c4","Type":"ContainerStarted","Data":"aa4820d4875b0b4449cb2c7a512e82f56a182188f58561773f130af26b6b0ab5"} Oct 05 20:57:12 crc kubenswrapper[4754]: I1005 20:57:12.281926 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-5sqj9" Oct 05 20:57:12 crc kubenswrapper[4754]: I1005 20:57:12.284756 4754 generic.go:334] "Generic (PLEG): container finished" podID="caf99098-6640-4077-8b97-69d1bfefa99c" containerID="126c9a934dc7d9adf3eb6d7c0390dda2dfae5b78c59cb48672aba70080b5915f" exitCode=0 Oct 05 20:57:12 crc kubenswrapper[4754]: I1005 20:57:12.284830 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xdqrj" event={"ID":"caf99098-6640-4077-8b97-69d1bfefa99c","Type":"ContainerDied","Data":"126c9a934dc7d9adf3eb6d7c0390dda2dfae5b78c59cb48672aba70080b5915f"} Oct 05 20:57:12 crc kubenswrapper[4754]: I1005 20:57:12.297621 4754 generic.go:334] "Generic (PLEG): container finished" podID="50c8cfa3-71a9-4c5e-895c-f15457183666" containerID="2d164208c1e65f5aef85b54dc315b995b6c67703f87f75c98d92a28f0d7aa2ad" exitCode=0 Oct 05 20:57:12 crc kubenswrapper[4754]: I1005 20:57:12.297788 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l2w5p" event={"ID":"50c8cfa3-71a9-4c5e-895c-f15457183666","Type":"ContainerDied","Data":"2d164208c1e65f5aef85b54dc315b995b6c67703f87f75c98d92a28f0d7aa2ad"} Oct 05 20:57:12 crc kubenswrapper[4754]: I1005 20:57:12.353181 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-5sqj9" podStartSLOduration=134.35315795100001 podStartE2EDuration="2m14.353157951s" podCreationTimestamp="2025-10-05 20:54:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 20:57:12.35311384 +0000 UTC m=+156.257232540" watchObservedRunningTime="2025-10-05 20:57:12.353157951 +0000 UTC m=+156.257276661" Oct 05 20:57:12 crc kubenswrapper[4754]: I1005 20:57:12.854972 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Oct 05 20:57:12 crc kubenswrapper[4754]: I1005 20:57:12.989163 4754 patch_prober.go:28] interesting pod/router-default-5444994796-ll9lb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 05 20:57:12 crc kubenswrapper[4754]: [-]has-synced failed: reason withheld Oct 05 20:57:12 crc kubenswrapper[4754]: [+]process-running ok Oct 05 20:57:12 crc kubenswrapper[4754]: healthz check failed Oct 05 20:57:12 crc kubenswrapper[4754]: I1005 20:57:12.989242 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ll9lb" podUID="ebf750fb-f47c-420b-931d-3ac0a23e9e7b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 05 20:57:13 crc kubenswrapper[4754]: I1005 20:57:13.016438 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 05 20:57:13 crc kubenswrapper[4754]: I1005 20:57:13.017752 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 05 20:57:13 crc kubenswrapper[4754]: I1005 20:57:13.024293 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Oct 05 20:57:13 crc kubenswrapper[4754]: I1005 20:57:13.025829 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Oct 05 20:57:13 crc kubenswrapper[4754]: I1005 20:57:13.041097 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 05 20:57:13 crc kubenswrapper[4754]: I1005 20:57:13.178550 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/32533d89-e4d3-4c15-894e-cf9d09de5dfd-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"32533d89-e4d3-4c15-894e-cf9d09de5dfd\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 05 20:57:13 crc kubenswrapper[4754]: I1005 20:57:13.178713 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/32533d89-e4d3-4c15-894e-cf9d09de5dfd-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"32533d89-e4d3-4c15-894e-cf9d09de5dfd\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 05 20:57:13 crc kubenswrapper[4754]: I1005 20:57:13.280066 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/32533d89-e4d3-4c15-894e-cf9d09de5dfd-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"32533d89-e4d3-4c15-894e-cf9d09de5dfd\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 05 20:57:13 crc kubenswrapper[4754]: I1005 20:57:13.280179 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/32533d89-e4d3-4c15-894e-cf9d09de5dfd-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"32533d89-e4d3-4c15-894e-cf9d09de5dfd\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 05 20:57:13 crc kubenswrapper[4754]: I1005 20:57:13.280281 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/32533d89-e4d3-4c15-894e-cf9d09de5dfd-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"32533d89-e4d3-4c15-894e-cf9d09de5dfd\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 05 20:57:13 crc kubenswrapper[4754]: I1005 20:57:13.302053 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/32533d89-e4d3-4c15-894e-cf9d09de5dfd-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"32533d89-e4d3-4c15-894e-cf9d09de5dfd\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 05 20:57:13 crc kubenswrapper[4754]: I1005 20:57:13.349319 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 05 20:57:13 crc kubenswrapper[4754]: I1005 20:57:13.657144 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-k8qff" Oct 05 20:57:13 crc kubenswrapper[4754]: I1005 20:57:13.673400 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-k8qff" Oct 05 20:57:13 crc kubenswrapper[4754]: I1005 20:57:13.988941 4754 patch_prober.go:28] interesting pod/router-default-5444994796-ll9lb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 05 20:57:13 crc kubenswrapper[4754]: [-]has-synced failed: reason withheld Oct 05 20:57:13 crc kubenswrapper[4754]: [+]process-running ok Oct 05 20:57:13 crc kubenswrapper[4754]: healthz check failed Oct 05 20:57:13 crc kubenswrapper[4754]: I1005 20:57:13.989017 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ll9lb" podUID="ebf750fb-f47c-420b-931d-3ac0a23e9e7b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 05 20:57:14 crc kubenswrapper[4754]: I1005 20:57:14.080304 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 05 20:57:14 crc kubenswrapper[4754]: I1005 20:57:14.084312 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 05 20:57:14 crc kubenswrapper[4754]: I1005 20:57:14.089156 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Oct 05 20:57:14 crc kubenswrapper[4754]: I1005 20:57:14.089352 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Oct 05 20:57:14 crc kubenswrapper[4754]: I1005 20:57:14.105123 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 05 20:57:14 crc kubenswrapper[4754]: I1005 20:57:14.210379 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b7be5d67-453a-418d-8ad1-39eb78f538c1-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"b7be5d67-453a-418d-8ad1-39eb78f538c1\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 05 20:57:14 crc kubenswrapper[4754]: I1005 20:57:14.210443 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b7be5d67-453a-418d-8ad1-39eb78f538c1-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"b7be5d67-453a-418d-8ad1-39eb78f538c1\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 05 20:57:14 crc kubenswrapper[4754]: I1005 20:57:14.313270 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b7be5d67-453a-418d-8ad1-39eb78f538c1-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"b7be5d67-453a-418d-8ad1-39eb78f538c1\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 05 20:57:14 crc kubenswrapper[4754]: I1005 20:57:14.313328 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b7be5d67-453a-418d-8ad1-39eb78f538c1-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"b7be5d67-453a-418d-8ad1-39eb78f538c1\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 05 20:57:14 crc kubenswrapper[4754]: I1005 20:57:14.313472 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b7be5d67-453a-418d-8ad1-39eb78f538c1-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"b7be5d67-453a-418d-8ad1-39eb78f538c1\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 05 20:57:14 crc kubenswrapper[4754]: I1005 20:57:14.334484 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 05 20:57:14 crc kubenswrapper[4754]: I1005 20:57:14.341196 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b7be5d67-453a-418d-8ad1-39eb78f538c1-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"b7be5d67-453a-418d-8ad1-39eb78f538c1\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 05 20:57:14 crc kubenswrapper[4754]: I1005 20:57:14.427925 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 05 20:57:14 crc kubenswrapper[4754]: I1005 20:57:14.469139 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"32533d89-e4d3-4c15-894e-cf9d09de5dfd","Type":"ContainerStarted","Data":"236a902f807096517e886a0b02ffb8a5d0d85b150f23a3144f6743311d2a7bff"} Oct 05 20:57:15 crc kubenswrapper[4754]: I1005 20:57:15.005919 4754 patch_prober.go:28] interesting pod/router-default-5444994796-ll9lb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 05 20:57:15 crc kubenswrapper[4754]: [-]has-synced failed: reason withheld Oct 05 20:57:15 crc kubenswrapper[4754]: [+]process-running ok Oct 05 20:57:15 crc kubenswrapper[4754]: healthz check failed Oct 05 20:57:15 crc kubenswrapper[4754]: I1005 20:57:15.006303 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ll9lb" podUID="ebf750fb-f47c-420b-931d-3ac0a23e9e7b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 05 20:57:15 crc kubenswrapper[4754]: I1005 20:57:15.114941 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-6xm7j" Oct 05 20:57:15 crc kubenswrapper[4754]: I1005 20:57:15.279853 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 05 20:57:15 crc kubenswrapper[4754]: W1005 20:57:15.423696 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-podb7be5d67_453a_418d_8ad1_39eb78f538c1.slice/crio-db13ad573c3c967f137f78098bbc60cdfbf35306f63821812f1e4a269f0a3247 WatchSource:0}: Error finding container db13ad573c3c967f137f78098bbc60cdfbf35306f63821812f1e4a269f0a3247: Status 404 returned error can't find the container with id db13ad573c3c967f137f78098bbc60cdfbf35306f63821812f1e4a269f0a3247 Oct 05 20:57:15 crc kubenswrapper[4754]: I1005 20:57:15.550774 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"b7be5d67-453a-418d-8ad1-39eb78f538c1","Type":"ContainerStarted","Data":"db13ad573c3c967f137f78098bbc60cdfbf35306f63821812f1e4a269f0a3247"} Oct 05 20:57:15 crc kubenswrapper[4754]: I1005 20:57:15.989685 4754 patch_prober.go:28] interesting pod/router-default-5444994796-ll9lb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 05 20:57:15 crc kubenswrapper[4754]: [-]has-synced failed: reason withheld Oct 05 20:57:15 crc kubenswrapper[4754]: [+]process-running ok Oct 05 20:57:15 crc kubenswrapper[4754]: healthz check failed Oct 05 20:57:15 crc kubenswrapper[4754]: I1005 20:57:15.989785 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ll9lb" podUID="ebf750fb-f47c-420b-931d-3ac0a23e9e7b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 05 20:57:16 crc kubenswrapper[4754]: I1005 20:57:16.615261 4754 generic.go:334] "Generic (PLEG): container finished" podID="60ff4c5c-9e45-47d3-b45d-a365e988342e" containerID="ff732b3e3d89105355ef4324b72c3db67106face7c8a9fd05f5ee00e8b2ceb24" exitCode=0 Oct 05 20:57:16 crc kubenswrapper[4754]: I1005 20:57:16.615317 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29328285-mrswx" event={"ID":"60ff4c5c-9e45-47d3-b45d-a365e988342e","Type":"ContainerDied","Data":"ff732b3e3d89105355ef4324b72c3db67106face7c8a9fd05f5ee00e8b2ceb24"} Oct 05 20:57:16 crc kubenswrapper[4754]: I1005 20:57:16.646833 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"32533d89-e4d3-4c15-894e-cf9d09de5dfd","Type":"ContainerStarted","Data":"8663d654229f2eb1a75b59737cf7c57c75b43f2f158b02dca1eaa49911c3a5be"} Oct 05 20:57:16 crc kubenswrapper[4754]: I1005 20:57:16.650723 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"b7be5d67-453a-418d-8ad1-39eb78f538c1","Type":"ContainerStarted","Data":"a53441069e56d156f34ab65458b3b228b6d20d629b691e88f91382f7db84a375"} Oct 05 20:57:16 crc kubenswrapper[4754]: I1005 20:57:16.681555 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=4.681526943 podStartE2EDuration="4.681526943s" podCreationTimestamp="2025-10-05 20:57:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 20:57:16.668175901 +0000 UTC m=+160.572294641" watchObservedRunningTime="2025-10-05 20:57:16.681526943 +0000 UTC m=+160.585645653" Oct 05 20:57:16 crc kubenswrapper[4754]: I1005 20:57:16.987341 4754 patch_prober.go:28] interesting pod/router-default-5444994796-ll9lb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 05 20:57:16 crc kubenswrapper[4754]: [-]has-synced failed: reason withheld Oct 05 20:57:16 crc kubenswrapper[4754]: [+]process-running ok Oct 05 20:57:16 crc kubenswrapper[4754]: healthz check failed Oct 05 20:57:16 crc kubenswrapper[4754]: I1005 20:57:16.987419 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ll9lb" podUID="ebf750fb-f47c-420b-931d-3ac0a23e9e7b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 05 20:57:17 crc kubenswrapper[4754]: I1005 20:57:17.700451 4754 generic.go:334] "Generic (PLEG): container finished" podID="32533d89-e4d3-4c15-894e-cf9d09de5dfd" containerID="8663d654229f2eb1a75b59737cf7c57c75b43f2f158b02dca1eaa49911c3a5be" exitCode=0 Oct 05 20:57:17 crc kubenswrapper[4754]: I1005 20:57:17.700585 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"32533d89-e4d3-4c15-894e-cf9d09de5dfd","Type":"ContainerDied","Data":"8663d654229f2eb1a75b59737cf7c57c75b43f2f158b02dca1eaa49911c3a5be"} Oct 05 20:57:17 crc kubenswrapper[4754]: I1005 20:57:17.708407 4754 generic.go:334] "Generic (PLEG): container finished" podID="b7be5d67-453a-418d-8ad1-39eb78f538c1" containerID="a53441069e56d156f34ab65458b3b228b6d20d629b691e88f91382f7db84a375" exitCode=0 Oct 05 20:57:17 crc kubenswrapper[4754]: I1005 20:57:17.708523 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"b7be5d67-453a-418d-8ad1-39eb78f538c1","Type":"ContainerDied","Data":"a53441069e56d156f34ab65458b3b228b6d20d629b691e88f91382f7db84a375"} Oct 05 20:57:17 crc kubenswrapper[4754]: I1005 20:57:17.723438 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/revision-pruner-9-crc" podStartSLOduration=3.723416098 podStartE2EDuration="3.723416098s" podCreationTimestamp="2025-10-05 20:57:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 20:57:16.688864667 +0000 UTC m=+160.592983397" watchObservedRunningTime="2025-10-05 20:57:17.723416098 +0000 UTC m=+161.627534808" Oct 05 20:57:17 crc kubenswrapper[4754]: I1005 20:57:17.988023 4754 patch_prober.go:28] interesting pod/router-default-5444994796-ll9lb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 05 20:57:17 crc kubenswrapper[4754]: [-]has-synced failed: reason withheld Oct 05 20:57:17 crc kubenswrapper[4754]: [+]process-running ok Oct 05 20:57:17 crc kubenswrapper[4754]: healthz check failed Oct 05 20:57:17 crc kubenswrapper[4754]: I1005 20:57:17.988846 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ll9lb" podUID="ebf750fb-f47c-420b-931d-3ac0a23e9e7b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 05 20:57:18 crc kubenswrapper[4754]: I1005 20:57:18.170564 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29328285-mrswx" Oct 05 20:57:18 crc kubenswrapper[4754]: I1005 20:57:18.225035 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/60ff4c5c-9e45-47d3-b45d-a365e988342e-config-volume\") pod \"60ff4c5c-9e45-47d3-b45d-a365e988342e\" (UID: \"60ff4c5c-9e45-47d3-b45d-a365e988342e\") " Oct 05 20:57:18 crc kubenswrapper[4754]: I1005 20:57:18.225082 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/60ff4c5c-9e45-47d3-b45d-a365e988342e-secret-volume\") pod \"60ff4c5c-9e45-47d3-b45d-a365e988342e\" (UID: \"60ff4c5c-9e45-47d3-b45d-a365e988342e\") " Oct 05 20:57:18 crc kubenswrapper[4754]: I1005 20:57:18.225200 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mv4pz\" (UniqueName: \"kubernetes.io/projected/60ff4c5c-9e45-47d3-b45d-a365e988342e-kube-api-access-mv4pz\") pod \"60ff4c5c-9e45-47d3-b45d-a365e988342e\" (UID: \"60ff4c5c-9e45-47d3-b45d-a365e988342e\") " Oct 05 20:57:18 crc kubenswrapper[4754]: I1005 20:57:18.226586 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/60ff4c5c-9e45-47d3-b45d-a365e988342e-config-volume" (OuterVolumeSpecName: "config-volume") pod "60ff4c5c-9e45-47d3-b45d-a365e988342e" (UID: "60ff4c5c-9e45-47d3-b45d-a365e988342e"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 20:57:18 crc kubenswrapper[4754]: I1005 20:57:18.236341 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/60ff4c5c-9e45-47d3-b45d-a365e988342e-kube-api-access-mv4pz" (OuterVolumeSpecName: "kube-api-access-mv4pz") pod "60ff4c5c-9e45-47d3-b45d-a365e988342e" (UID: "60ff4c5c-9e45-47d3-b45d-a365e988342e"). InnerVolumeSpecName "kube-api-access-mv4pz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 20:57:18 crc kubenswrapper[4754]: I1005 20:57:18.242758 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/60ff4c5c-9e45-47d3-b45d-a365e988342e-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "60ff4c5c-9e45-47d3-b45d-a365e988342e" (UID: "60ff4c5c-9e45-47d3-b45d-a365e988342e"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 20:57:18 crc kubenswrapper[4754]: I1005 20:57:18.326993 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mv4pz\" (UniqueName: \"kubernetes.io/projected/60ff4c5c-9e45-47d3-b45d-a365e988342e-kube-api-access-mv4pz\") on node \"crc\" DevicePath \"\"" Oct 05 20:57:18 crc kubenswrapper[4754]: I1005 20:57:18.327029 4754 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/60ff4c5c-9e45-47d3-b45d-a365e988342e-config-volume\") on node \"crc\" DevicePath \"\"" Oct 05 20:57:18 crc kubenswrapper[4754]: I1005 20:57:18.327040 4754 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/60ff4c5c-9e45-47d3-b45d-a365e988342e-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 05 20:57:18 crc kubenswrapper[4754]: I1005 20:57:18.467504 4754 patch_prober.go:28] interesting pod/console-f9d7485db-5sffm container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.9:8443/health\": dial tcp 10.217.0.9:8443: connect: connection refused" start-of-body= Oct 05 20:57:18 crc kubenswrapper[4754]: I1005 20:57:18.468081 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-5sffm" podUID="c56f5eb8-3a12-428c-b347-78369dc19008" containerName="console" probeResult="failure" output="Get \"https://10.217.0.9:8443/health\": dial tcp 10.217.0.9:8443: connect: connection refused" Oct 05 20:57:18 crc kubenswrapper[4754]: I1005 20:57:18.722787 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29328285-mrswx" Oct 05 20:57:18 crc kubenswrapper[4754]: I1005 20:57:18.723851 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29328285-mrswx" event={"ID":"60ff4c5c-9e45-47d3-b45d-a365e988342e","Type":"ContainerDied","Data":"92251d5fc79e9f76bcbac6e3baa7658adb4cc5be49e94d1e9aba4a934ab16083"} Oct 05 20:57:18 crc kubenswrapper[4754]: I1005 20:57:18.723930 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="92251d5fc79e9f76bcbac6e3baa7658adb4cc5be49e94d1e9aba4a934ab16083" Oct 05 20:57:18 crc kubenswrapper[4754]: I1005 20:57:18.993667 4754 patch_prober.go:28] interesting pod/router-default-5444994796-ll9lb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 05 20:57:18 crc kubenswrapper[4754]: [-]has-synced failed: reason withheld Oct 05 20:57:18 crc kubenswrapper[4754]: [+]process-running ok Oct 05 20:57:18 crc kubenswrapper[4754]: healthz check failed Oct 05 20:57:18 crc kubenswrapper[4754]: I1005 20:57:18.995195 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ll9lb" podUID="ebf750fb-f47c-420b-931d-3ac0a23e9e7b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 05 20:57:19 crc kubenswrapper[4754]: I1005 20:57:19.051751 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 05 20:57:19 crc kubenswrapper[4754]: I1005 20:57:19.107084 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 05 20:57:19 crc kubenswrapper[4754]: I1005 20:57:19.139364 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/32533d89-e4d3-4c15-894e-cf9d09de5dfd-kube-api-access\") pod \"32533d89-e4d3-4c15-894e-cf9d09de5dfd\" (UID: \"32533d89-e4d3-4c15-894e-cf9d09de5dfd\") " Oct 05 20:57:19 crc kubenswrapper[4754]: I1005 20:57:19.139482 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b7be5d67-453a-418d-8ad1-39eb78f538c1-kube-api-access\") pod \"b7be5d67-453a-418d-8ad1-39eb78f538c1\" (UID: \"b7be5d67-453a-418d-8ad1-39eb78f538c1\") " Oct 05 20:57:19 crc kubenswrapper[4754]: I1005 20:57:19.139546 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b7be5d67-453a-418d-8ad1-39eb78f538c1-kubelet-dir\") pod \"b7be5d67-453a-418d-8ad1-39eb78f538c1\" (UID: \"b7be5d67-453a-418d-8ad1-39eb78f538c1\") " Oct 05 20:57:19 crc kubenswrapper[4754]: I1005 20:57:19.139716 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/32533d89-e4d3-4c15-894e-cf9d09de5dfd-kubelet-dir\") pod \"32533d89-e4d3-4c15-894e-cf9d09de5dfd\" (UID: \"32533d89-e4d3-4c15-894e-cf9d09de5dfd\") " Oct 05 20:57:19 crc kubenswrapper[4754]: I1005 20:57:19.140180 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/32533d89-e4d3-4c15-894e-cf9d09de5dfd-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "32533d89-e4d3-4c15-894e-cf9d09de5dfd" (UID: "32533d89-e4d3-4c15-894e-cf9d09de5dfd"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 05 20:57:19 crc kubenswrapper[4754]: I1005 20:57:19.141358 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b7be5d67-453a-418d-8ad1-39eb78f538c1-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "b7be5d67-453a-418d-8ad1-39eb78f538c1" (UID: "b7be5d67-453a-418d-8ad1-39eb78f538c1"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 05 20:57:19 crc kubenswrapper[4754]: I1005 20:57:19.151339 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/32533d89-e4d3-4c15-894e-cf9d09de5dfd-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "32533d89-e4d3-4c15-894e-cf9d09de5dfd" (UID: "32533d89-e4d3-4c15-894e-cf9d09de5dfd"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 20:57:19 crc kubenswrapper[4754]: I1005 20:57:19.161704 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b7be5d67-453a-418d-8ad1-39eb78f538c1-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "b7be5d67-453a-418d-8ad1-39eb78f538c1" (UID: "b7be5d67-453a-418d-8ad1-39eb78f538c1"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 20:57:19 crc kubenswrapper[4754]: I1005 20:57:19.243874 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b7be5d67-453a-418d-8ad1-39eb78f538c1-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 05 20:57:19 crc kubenswrapper[4754]: I1005 20:57:19.243931 4754 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b7be5d67-453a-418d-8ad1-39eb78f538c1-kubelet-dir\") on node \"crc\" DevicePath \"\"" Oct 05 20:57:19 crc kubenswrapper[4754]: I1005 20:57:19.243950 4754 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/32533d89-e4d3-4c15-894e-cf9d09de5dfd-kubelet-dir\") on node \"crc\" DevicePath \"\"" Oct 05 20:57:19 crc kubenswrapper[4754]: I1005 20:57:19.243968 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/32533d89-e4d3-4c15-894e-cf9d09de5dfd-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 05 20:57:19 crc kubenswrapper[4754]: I1005 20:57:19.775615 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 05 20:57:19 crc kubenswrapper[4754]: I1005 20:57:19.775676 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"32533d89-e4d3-4c15-894e-cf9d09de5dfd","Type":"ContainerDied","Data":"236a902f807096517e886a0b02ffb8a5d0d85b150f23a3144f6743311d2a7bff"} Oct 05 20:57:19 crc kubenswrapper[4754]: I1005 20:57:19.775751 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="236a902f807096517e886a0b02ffb8a5d0d85b150f23a3144f6743311d2a7bff" Oct 05 20:57:19 crc kubenswrapper[4754]: I1005 20:57:19.779671 4754 patch_prober.go:28] interesting pod/downloads-7954f5f757-ksczt container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.20:8080/\": dial tcp 10.217.0.20:8080: connect: connection refused" start-of-body= Oct 05 20:57:19 crc kubenswrapper[4754]: I1005 20:57:19.779715 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-ksczt" podUID="3934dd94-b9fa-4079-a268-60a9470dab41" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.20:8080/\": dial tcp 10.217.0.20:8080: connect: connection refused" Oct 05 20:57:19 crc kubenswrapper[4754]: I1005 20:57:19.779936 4754 patch_prober.go:28] interesting pod/downloads-7954f5f757-ksczt container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.20:8080/\": dial tcp 10.217.0.20:8080: connect: connection refused" start-of-body= Oct 05 20:57:19 crc kubenswrapper[4754]: I1005 20:57:19.780048 4754 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-ksczt" podUID="3934dd94-b9fa-4079-a268-60a9470dab41" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.20:8080/\": dial tcp 10.217.0.20:8080: connect: connection refused" Oct 05 20:57:19 crc kubenswrapper[4754]: I1005 20:57:19.811305 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"b7be5d67-453a-418d-8ad1-39eb78f538c1","Type":"ContainerDied","Data":"db13ad573c3c967f137f78098bbc60cdfbf35306f63821812f1e4a269f0a3247"} Oct 05 20:57:19 crc kubenswrapper[4754]: I1005 20:57:19.811351 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="db13ad573c3c967f137f78098bbc60cdfbf35306f63821812f1e4a269f0a3247" Oct 05 20:57:19 crc kubenswrapper[4754]: I1005 20:57:19.811416 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 05 20:57:19 crc kubenswrapper[4754]: I1005 20:57:19.985572 4754 patch_prober.go:28] interesting pod/router-default-5444994796-ll9lb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 05 20:57:19 crc kubenswrapper[4754]: [-]has-synced failed: reason withheld Oct 05 20:57:19 crc kubenswrapper[4754]: [+]process-running ok Oct 05 20:57:19 crc kubenswrapper[4754]: healthz check failed Oct 05 20:57:19 crc kubenswrapper[4754]: I1005 20:57:19.985670 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ll9lb" podUID="ebf750fb-f47c-420b-931d-3ac0a23e9e7b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 05 20:57:20 crc kubenswrapper[4754]: I1005 20:57:20.989855 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-ll9lb" Oct 05 20:57:20 crc kubenswrapper[4754]: I1005 20:57:20.994389 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-ll9lb" Oct 05 20:57:21 crc kubenswrapper[4754]: I1005 20:57:21.706236 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7dc728ea-2601-44f0-bcab-2913f034007d-metrics-certs\") pod \"network-metrics-daemon-nwrnt\" (UID: \"7dc728ea-2601-44f0-bcab-2913f034007d\") " pod="openshift-multus/network-metrics-daemon-nwrnt" Oct 05 20:57:21 crc kubenswrapper[4754]: I1005 20:57:21.717379 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7dc728ea-2601-44f0-bcab-2913f034007d-metrics-certs\") pod \"network-metrics-daemon-nwrnt\" (UID: \"7dc728ea-2601-44f0-bcab-2913f034007d\") " pod="openshift-multus/network-metrics-daemon-nwrnt" Oct 05 20:57:21 crc kubenswrapper[4754]: I1005 20:57:21.758938 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nwrnt" Oct 05 20:57:22 crc kubenswrapper[4754]: I1005 20:57:22.502302 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-nwrnt"] Oct 05 20:57:22 crc kubenswrapper[4754]: W1005 20:57:22.514449 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7dc728ea_2601_44f0_bcab_2913f034007d.slice/crio-31ee1283ab2af436e7a2804c1231cdbe8d6e4555319eb7443fe2b8dc0242a8ac WatchSource:0}: Error finding container 31ee1283ab2af436e7a2804c1231cdbe8d6e4555319eb7443fe2b8dc0242a8ac: Status 404 returned error can't find the container with id 31ee1283ab2af436e7a2804c1231cdbe8d6e4555319eb7443fe2b8dc0242a8ac Oct 05 20:57:22 crc kubenswrapper[4754]: I1005 20:57:22.866984 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-nwrnt" event={"ID":"7dc728ea-2601-44f0-bcab-2913f034007d","Type":"ContainerStarted","Data":"31ee1283ab2af436e7a2804c1231cdbe8d6e4555319eb7443fe2b8dc0242a8ac"} Oct 05 20:57:28 crc kubenswrapper[4754]: I1005 20:57:28.471445 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-5sffm" Oct 05 20:57:28 crc kubenswrapper[4754]: I1005 20:57:28.475145 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-5sffm" Oct 05 20:57:29 crc kubenswrapper[4754]: I1005 20:57:29.785569 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-ksczt" Oct 05 20:57:31 crc kubenswrapper[4754]: I1005 20:57:31.291177 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-5sqj9" Oct 05 20:57:32 crc kubenswrapper[4754]: I1005 20:57:32.952694 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-nwrnt" event={"ID":"7dc728ea-2601-44f0-bcab-2913f034007d","Type":"ContainerStarted","Data":"0f47d0d7db37b5a2e703edb57d070e206ea0dac5bf53c8ce97c2aeabe5ac409c"} Oct 05 20:57:35 crc kubenswrapper[4754]: I1005 20:57:35.245015 4754 patch_prober.go:28] interesting pod/machine-config-daemon-b2h9k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 20:57:35 crc kubenswrapper[4754]: I1005 20:57:35.245103 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 20:57:39 crc kubenswrapper[4754]: I1005 20:57:39.765371 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-nrt9t" Oct 05 20:57:46 crc kubenswrapper[4754]: I1005 20:57:46.865320 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 20:57:47 crc kubenswrapper[4754]: E1005 20:57:47.541531 4754 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Oct 05 20:57:47 crc kubenswrapper[4754]: E1005 20:57:47.541825 4754 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-68wh5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-5889x_openshift-marketplace(61cebf9f-dd0d-40f9-9bd5-1ff0dd7038f8): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 05 20:57:47 crc kubenswrapper[4754]: E1005 20:57:47.543057 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-5889x" podUID="61cebf9f-dd0d-40f9-9bd5-1ff0dd7038f8" Oct 05 20:57:49 crc kubenswrapper[4754]: E1005 20:57:49.130719 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-5889x" podUID="61cebf9f-dd0d-40f9-9bd5-1ff0dd7038f8" Oct 05 20:57:49 crc kubenswrapper[4754]: E1005 20:57:49.213317 4754 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Oct 05 20:57:49 crc kubenswrapper[4754]: E1005 20:57:49.213582 4754 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-2ngdc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-zwh6t_openshift-marketplace(85d11cf6-caa4-4ee2-a8ee-48c44469d4b7): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 05 20:57:49 crc kubenswrapper[4754]: E1005 20:57:49.214911 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-zwh6t" podUID="85d11cf6-caa4-4ee2-a8ee-48c44469d4b7" Oct 05 20:57:51 crc kubenswrapper[4754]: E1005 20:57:51.981943 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-zwh6t" podUID="85d11cf6-caa4-4ee2-a8ee-48c44469d4b7" Oct 05 20:57:53 crc kubenswrapper[4754]: E1005 20:57:53.011102 4754 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Oct 05 20:57:53 crc kubenswrapper[4754]: E1005 20:57:53.011277 4754 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-p86kl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-xknjb_openshift-marketplace(b5b25be1-b393-4d98-aaf0-3f9693ba89d7): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 05 20:57:53 crc kubenswrapper[4754]: E1005 20:57:53.014304 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-xknjb" podUID="b5b25be1-b393-4d98-aaf0-3f9693ba89d7" Oct 05 20:57:53 crc kubenswrapper[4754]: E1005 20:57:53.035720 4754 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Oct 05 20:57:53 crc kubenswrapper[4754]: E1005 20:57:53.035876 4754 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6wq48,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-l2w5p_openshift-marketplace(50c8cfa3-71a9-4c5e-895c-f15457183666): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 05 20:57:53 crc kubenswrapper[4754]: E1005 20:57:53.038975 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-l2w5p" podUID="50c8cfa3-71a9-4c5e-895c-f15457183666" Oct 05 20:57:53 crc kubenswrapper[4754]: E1005 20:57:53.081744 4754 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Oct 05 20:57:53 crc kubenswrapper[4754]: E1005 20:57:53.081927 4754 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-h8xtz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-5qjvg_openshift-marketplace(d5e4795f-6f22-4cde-a939-970f1ec20119): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 05 20:57:53 crc kubenswrapper[4754]: E1005 20:57:53.085721 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-5qjvg" podUID="d5e4795f-6f22-4cde-a939-970f1ec20119" Oct 05 20:57:53 crc kubenswrapper[4754]: E1005 20:57:53.186304 4754 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Oct 05 20:57:53 crc kubenswrapper[4754]: E1005 20:57:53.186474 4754 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-t7p85,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-59nwx_openshift-marketplace(232ff34c-dc36-4012-9038-e10014a179cd): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 05 20:57:53 crc kubenswrapper[4754]: E1005 20:57:53.187789 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-59nwx" podUID="232ff34c-dc36-4012-9038-e10014a179cd" Oct 05 20:57:56 crc kubenswrapper[4754]: E1005 20:57:56.841123 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-5qjvg" podUID="d5e4795f-6f22-4cde-a939-970f1ec20119" Oct 05 20:57:56 crc kubenswrapper[4754]: E1005 20:57:56.842202 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-xknjb" podUID="b5b25be1-b393-4d98-aaf0-3f9693ba89d7" Oct 05 20:57:56 crc kubenswrapper[4754]: E1005 20:57:56.842326 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-l2w5p" podUID="50c8cfa3-71a9-4c5e-895c-f15457183666" Oct 05 20:57:56 crc kubenswrapper[4754]: E1005 20:57:56.842402 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-59nwx" podUID="232ff34c-dc36-4012-9038-e10014a179cd" Oct 05 20:57:57 crc kubenswrapper[4754]: E1005 20:57:57.345826 4754 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Oct 05 20:57:57 crc kubenswrapper[4754]: E1005 20:57:57.346329 4754 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9dgdm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-xdqrj_openshift-marketplace(caf99098-6640-4077-8b97-69d1bfefa99c): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 05 20:57:57 crc kubenswrapper[4754]: E1005 20:57:57.348478 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-xdqrj" podUID="caf99098-6640-4077-8b97-69d1bfefa99c" Oct 05 20:57:57 crc kubenswrapper[4754]: E1005 20:57:57.423029 4754 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Oct 05 20:57:57 crc kubenswrapper[4754]: E1005 20:57:57.423172 4754 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-b5k8b,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-nz2vw_openshift-marketplace(e5dff6e7-5693-4f7c-af40-d7b10cbb73c7): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 05 20:57:57 crc kubenswrapper[4754]: E1005 20:57:57.424465 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-nz2vw" podUID="e5dff6e7-5693-4f7c-af40-d7b10cbb73c7" Oct 05 20:57:58 crc kubenswrapper[4754]: I1005 20:57:58.120643 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-nwrnt" event={"ID":"7dc728ea-2601-44f0-bcab-2913f034007d","Type":"ContainerStarted","Data":"cf4699a89df0f0312550df4dba9ceb05fdd7b847df28852f0c96d525ef49e2d4"} Oct 05 20:57:58 crc kubenswrapper[4754]: E1005 20:57:58.123133 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-nz2vw" podUID="e5dff6e7-5693-4f7c-af40-d7b10cbb73c7" Oct 05 20:57:58 crc kubenswrapper[4754]: E1005 20:57:58.125383 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-xdqrj" podUID="caf99098-6640-4077-8b97-69d1bfefa99c" Oct 05 20:58:03 crc kubenswrapper[4754]: I1005 20:58:03.868053 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-nwrnt" podStartSLOduration=185.868022526 podStartE2EDuration="3m5.868022526s" podCreationTimestamp="2025-10-05 20:54:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 20:57:58.201513563 +0000 UTC m=+202.105632273" watchObservedRunningTime="2025-10-05 20:58:03.868022526 +0000 UTC m=+207.772141226" Oct 05 20:58:05 crc kubenswrapper[4754]: I1005 20:58:05.143380 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-gchnw"] Oct 05 20:58:05 crc kubenswrapper[4754]: I1005 20:58:05.244797 4754 patch_prober.go:28] interesting pod/machine-config-daemon-b2h9k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 20:58:05 crc kubenswrapper[4754]: I1005 20:58:05.245544 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 20:58:05 crc kubenswrapper[4754]: I1005 20:58:05.245672 4754 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" Oct 05 20:58:05 crc kubenswrapper[4754]: I1005 20:58:05.246415 4754 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"4bb8c5903e8b70dd19eebb111097d96f725e17de44e06c3e74c067376f6c7360"} pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 05 20:58:05 crc kubenswrapper[4754]: I1005 20:58:05.246630 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" containerName="machine-config-daemon" containerID="cri-o://4bb8c5903e8b70dd19eebb111097d96f725e17de44e06c3e74c067376f6c7360" gracePeriod=600 Oct 05 20:58:06 crc kubenswrapper[4754]: I1005 20:58:06.201536 4754 generic.go:334] "Generic (PLEG): container finished" podID="be95a413-6a59-45b4-84b7-b43cae694a26" containerID="4bb8c5903e8b70dd19eebb111097d96f725e17de44e06c3e74c067376f6c7360" exitCode=0 Oct 05 20:58:06 crc kubenswrapper[4754]: I1005 20:58:06.201633 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" event={"ID":"be95a413-6a59-45b4-84b7-b43cae694a26","Type":"ContainerDied","Data":"4bb8c5903e8b70dd19eebb111097d96f725e17de44e06c3e74c067376f6c7360"} Oct 05 20:58:08 crc kubenswrapper[4754]: I1005 20:58:08.214735 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" event={"ID":"be95a413-6a59-45b4-84b7-b43cae694a26","Type":"ContainerStarted","Data":"1748c85d7a12794d4f4515c3318cf93b10e679b7d52cfbe7c2fbc83011ee324c"} Oct 05 20:58:08 crc kubenswrapper[4754]: I1005 20:58:08.221749 4754 generic.go:334] "Generic (PLEG): container finished" podID="61cebf9f-dd0d-40f9-9bd5-1ff0dd7038f8" containerID="85b2ac566e0435f9eb3619b63bd81c71c3f614dc0b553c34cf9826126a37aec2" exitCode=0 Oct 05 20:58:08 crc kubenswrapper[4754]: I1005 20:58:08.221865 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5889x" event={"ID":"61cebf9f-dd0d-40f9-9bd5-1ff0dd7038f8","Type":"ContainerDied","Data":"85b2ac566e0435f9eb3619b63bd81c71c3f614dc0b553c34cf9826126a37aec2"} Oct 05 20:58:08 crc kubenswrapper[4754]: I1005 20:58:08.224652 4754 generic.go:334] "Generic (PLEG): container finished" podID="85d11cf6-caa4-4ee2-a8ee-48c44469d4b7" containerID="25bc1b4acb780b87220b6a8388386b75b9f855100762c4b16dee83e294476085" exitCode=0 Oct 05 20:58:08 crc kubenswrapper[4754]: I1005 20:58:08.224710 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zwh6t" event={"ID":"85d11cf6-caa4-4ee2-a8ee-48c44469d4b7","Type":"ContainerDied","Data":"25bc1b4acb780b87220b6a8388386b75b9f855100762c4b16dee83e294476085"} Oct 05 20:58:09 crc kubenswrapper[4754]: I1005 20:58:09.250574 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5889x" event={"ID":"61cebf9f-dd0d-40f9-9bd5-1ff0dd7038f8","Type":"ContainerStarted","Data":"36c567a32b080330ac66b125a7e5a956a5cbe8f6902f9406a5ac19b451119129"} Oct 05 20:58:09 crc kubenswrapper[4754]: I1005 20:58:09.277396 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-5889x" podStartSLOduration=2.83232717 podStartE2EDuration="1m0.277373173s" podCreationTimestamp="2025-10-05 20:57:09 +0000 UTC" firstStartedPulling="2025-10-05 20:57:11.244556567 +0000 UTC m=+155.148675277" lastFinishedPulling="2025-10-05 20:58:08.68960257 +0000 UTC m=+212.593721280" observedRunningTime="2025-10-05 20:58:09.274717788 +0000 UTC m=+213.178836498" watchObservedRunningTime="2025-10-05 20:58:09.277373173 +0000 UTC m=+213.181491883" Oct 05 20:58:09 crc kubenswrapper[4754]: I1005 20:58:09.739547 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-5889x" Oct 05 20:58:09 crc kubenswrapper[4754]: I1005 20:58:09.739903 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-5889x" Oct 05 20:58:10 crc kubenswrapper[4754]: I1005 20:58:10.255832 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xdqrj" event={"ID":"caf99098-6640-4077-8b97-69d1bfefa99c","Type":"ContainerStarted","Data":"bbedaaf423efe4f7b92787043d222dab156fbeb2eeaa05812b1b89d2d2cfc01a"} Oct 05 20:58:10 crc kubenswrapper[4754]: I1005 20:58:10.257937 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zwh6t" event={"ID":"85d11cf6-caa4-4ee2-a8ee-48c44469d4b7","Type":"ContainerStarted","Data":"df11ecc12c5fb6fe0d12cdd353c9d061a0f848224746d9ac5db943911be9d0ff"} Oct 05 20:58:10 crc kubenswrapper[4754]: I1005 20:58:10.260170 4754 generic.go:334] "Generic (PLEG): container finished" podID="b5b25be1-b393-4d98-aaf0-3f9693ba89d7" containerID="04d86388c6f62ae88d96a2cded7ecd37f98d5c9ff7697d137554bbef2e3aa5c2" exitCode=0 Oct 05 20:58:10 crc kubenswrapper[4754]: I1005 20:58:10.260324 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xknjb" event={"ID":"b5b25be1-b393-4d98-aaf0-3f9693ba89d7","Type":"ContainerDied","Data":"04d86388c6f62ae88d96a2cded7ecd37f98d5c9ff7697d137554bbef2e3aa5c2"} Oct 05 20:58:10 crc kubenswrapper[4754]: I1005 20:58:10.306988 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-zwh6t" podStartSLOduration=4.389993192 podStartE2EDuration="1m3.30696765s" podCreationTimestamp="2025-10-05 20:57:07 +0000 UTC" firstStartedPulling="2025-10-05 20:57:10.005935041 +0000 UTC m=+153.910053741" lastFinishedPulling="2025-10-05 20:58:08.922909469 +0000 UTC m=+212.827028199" observedRunningTime="2025-10-05 20:58:10.302823769 +0000 UTC m=+214.206942479" watchObservedRunningTime="2025-10-05 20:58:10.30696765 +0000 UTC m=+214.211086360" Oct 05 20:58:10 crc kubenswrapper[4754]: I1005 20:58:10.935053 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-marketplace-5889x" podUID="61cebf9f-dd0d-40f9-9bd5-1ff0dd7038f8" containerName="registry-server" probeResult="failure" output=< Oct 05 20:58:10 crc kubenswrapper[4754]: timeout: failed to connect service ":50051" within 1s Oct 05 20:58:10 crc kubenswrapper[4754]: > Oct 05 20:58:11 crc kubenswrapper[4754]: I1005 20:58:11.266289 4754 generic.go:334] "Generic (PLEG): container finished" podID="d5e4795f-6f22-4cde-a939-970f1ec20119" containerID="533f474b59c58333faac7190e98ef2bb9bf1ea468e6acef27965f264795053e0" exitCode=0 Oct 05 20:58:11 crc kubenswrapper[4754]: I1005 20:58:11.266380 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5qjvg" event={"ID":"d5e4795f-6f22-4cde-a939-970f1ec20119","Type":"ContainerDied","Data":"533f474b59c58333faac7190e98ef2bb9bf1ea468e6acef27965f264795053e0"} Oct 05 20:58:11 crc kubenswrapper[4754]: I1005 20:58:11.271165 4754 generic.go:334] "Generic (PLEG): container finished" podID="caf99098-6640-4077-8b97-69d1bfefa99c" containerID="bbedaaf423efe4f7b92787043d222dab156fbeb2eeaa05812b1b89d2d2cfc01a" exitCode=0 Oct 05 20:58:11 crc kubenswrapper[4754]: I1005 20:58:11.271217 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xdqrj" event={"ID":"caf99098-6640-4077-8b97-69d1bfefa99c","Type":"ContainerDied","Data":"bbedaaf423efe4f7b92787043d222dab156fbeb2eeaa05812b1b89d2d2cfc01a"} Oct 05 20:58:11 crc kubenswrapper[4754]: I1005 20:58:11.276754 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nz2vw" event={"ID":"e5dff6e7-5693-4f7c-af40-d7b10cbb73c7","Type":"ContainerStarted","Data":"8b226498309d87d21927e83782cf58d174034d45662688c5c47399a05d078694"} Oct 05 20:58:11 crc kubenswrapper[4754]: I1005 20:58:11.279249 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xknjb" event={"ID":"b5b25be1-b393-4d98-aaf0-3f9693ba89d7","Type":"ContainerStarted","Data":"8f06273caa8a4b6742bd4cedff3d16610bf1ecc0d97633d087c4942fbbb30f70"} Oct 05 20:58:11 crc kubenswrapper[4754]: I1005 20:58:11.317274 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-xknjb" podStartSLOduration=3.751201994 podStartE2EDuration="1m4.317251534s" podCreationTimestamp="2025-10-05 20:57:07 +0000 UTC" firstStartedPulling="2025-10-05 20:57:10.133332903 +0000 UTC m=+154.037451613" lastFinishedPulling="2025-10-05 20:58:10.699382443 +0000 UTC m=+214.603501153" observedRunningTime="2025-10-05 20:58:11.315771208 +0000 UTC m=+215.219889918" watchObservedRunningTime="2025-10-05 20:58:11.317251534 +0000 UTC m=+215.221370244" Oct 05 20:58:12 crc kubenswrapper[4754]: I1005 20:58:12.295702 4754 generic.go:334] "Generic (PLEG): container finished" podID="e5dff6e7-5693-4f7c-af40-d7b10cbb73c7" containerID="8b226498309d87d21927e83782cf58d174034d45662688c5c47399a05d078694" exitCode=0 Oct 05 20:58:12 crc kubenswrapper[4754]: I1005 20:58:12.296149 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nz2vw" event={"ID":"e5dff6e7-5693-4f7c-af40-d7b10cbb73c7","Type":"ContainerDied","Data":"8b226498309d87d21927e83782cf58d174034d45662688c5c47399a05d078694"} Oct 05 20:58:13 crc kubenswrapper[4754]: I1005 20:58:13.310178 4754 generic.go:334] "Generic (PLEG): container finished" podID="50c8cfa3-71a9-4c5e-895c-f15457183666" containerID="7cbaef3132e2bdb3a993930804668b94e03510605a6cd26ce96e56c8e4277b5a" exitCode=0 Oct 05 20:58:13 crc kubenswrapper[4754]: I1005 20:58:13.310455 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l2w5p" event={"ID":"50c8cfa3-71a9-4c5e-895c-f15457183666","Type":"ContainerDied","Data":"7cbaef3132e2bdb3a993930804668b94e03510605a6cd26ce96e56c8e4277b5a"} Oct 05 20:58:13 crc kubenswrapper[4754]: I1005 20:58:13.322895 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nz2vw" event={"ID":"e5dff6e7-5693-4f7c-af40-d7b10cbb73c7","Type":"ContainerStarted","Data":"d33974403ef16ebac0e6d526b32622f3e89f6e3576632663f184ad4b670cb39e"} Oct 05 20:58:13 crc kubenswrapper[4754]: I1005 20:58:13.328619 4754 generic.go:334] "Generic (PLEG): container finished" podID="232ff34c-dc36-4012-9038-e10014a179cd" containerID="8f4390632ec901428a65e37624ed36ac020cb98eccefecc6caa6156dcd884c2e" exitCode=0 Oct 05 20:58:13 crc kubenswrapper[4754]: I1005 20:58:13.328684 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-59nwx" event={"ID":"232ff34c-dc36-4012-9038-e10014a179cd","Type":"ContainerDied","Data":"8f4390632ec901428a65e37624ed36ac020cb98eccefecc6caa6156dcd884c2e"} Oct 05 20:58:13 crc kubenswrapper[4754]: I1005 20:58:13.354298 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-nz2vw" podStartSLOduration=2.8636093259999997 podStartE2EDuration="1m3.354273234s" podCreationTimestamp="2025-10-05 20:57:10 +0000 UTC" firstStartedPulling="2025-10-05 20:57:12.276309143 +0000 UTC m=+156.180427853" lastFinishedPulling="2025-10-05 20:58:12.766973051 +0000 UTC m=+216.671091761" observedRunningTime="2025-10-05 20:58:13.346735259 +0000 UTC m=+217.250853969" watchObservedRunningTime="2025-10-05 20:58:13.354273234 +0000 UTC m=+217.258391944" Oct 05 20:58:14 crc kubenswrapper[4754]: I1005 20:58:14.336991 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xdqrj" event={"ID":"caf99098-6640-4077-8b97-69d1bfefa99c","Type":"ContainerStarted","Data":"4839e09a747a0b0ba463d4adc2e850e0ab819456da4ec507586bf54f71323ba8"} Oct 05 20:58:14 crc kubenswrapper[4754]: I1005 20:58:14.360143 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-xdqrj" podStartSLOduration=3.336768349 podStartE2EDuration="1m4.360116099s" podCreationTimestamp="2025-10-05 20:57:10 +0000 UTC" firstStartedPulling="2025-10-05 20:57:12.288328031 +0000 UTC m=+156.192446741" lastFinishedPulling="2025-10-05 20:58:13.311675781 +0000 UTC m=+217.215794491" observedRunningTime="2025-10-05 20:58:14.357876095 +0000 UTC m=+218.261994815" watchObservedRunningTime="2025-10-05 20:58:14.360116099 +0000 UTC m=+218.264234809" Oct 05 20:58:16 crc kubenswrapper[4754]: I1005 20:58:16.350330 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5qjvg" event={"ID":"d5e4795f-6f22-4cde-a939-970f1ec20119","Type":"ContainerStarted","Data":"c011f2109a4ff30788314d105b743e703d46bc1ffcc11537bc4d9abde8d93702"} Oct 05 20:58:16 crc kubenswrapper[4754]: I1005 20:58:16.370066 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-5qjvg" podStartSLOduration=5.554426966 podStartE2EDuration="1m9.370043566s" podCreationTimestamp="2025-10-05 20:57:07 +0000 UTC" firstStartedPulling="2025-10-05 20:57:11.252484326 +0000 UTC m=+155.156603036" lastFinishedPulling="2025-10-05 20:58:15.068100916 +0000 UTC m=+218.972219636" observedRunningTime="2025-10-05 20:58:16.365930026 +0000 UTC m=+220.270048736" watchObservedRunningTime="2025-10-05 20:58:16.370043566 +0000 UTC m=+220.274162276" Oct 05 20:58:17 crc kubenswrapper[4754]: I1005 20:58:17.689976 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-xknjb" Oct 05 20:58:17 crc kubenswrapper[4754]: I1005 20:58:17.690440 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-xknjb" Oct 05 20:58:17 crc kubenswrapper[4754]: I1005 20:58:17.770406 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-xknjb" Oct 05 20:58:17 crc kubenswrapper[4754]: I1005 20:58:17.945161 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-zwh6t" Oct 05 20:58:17 crc kubenswrapper[4754]: I1005 20:58:17.945215 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-zwh6t" Oct 05 20:58:17 crc kubenswrapper[4754]: I1005 20:58:17.997810 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-zwh6t" Oct 05 20:58:18 crc kubenswrapper[4754]: I1005 20:58:18.129573 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-5qjvg" Oct 05 20:58:18 crc kubenswrapper[4754]: I1005 20:58:18.129643 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-5qjvg" Oct 05 20:58:18 crc kubenswrapper[4754]: I1005 20:58:18.363606 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-59nwx" event={"ID":"232ff34c-dc36-4012-9038-e10014a179cd","Type":"ContainerStarted","Data":"beed44416a349711a5a4660be2a8a715e97d73dfe18584e74b7b560c36f2b3a3"} Oct 05 20:58:18 crc kubenswrapper[4754]: I1005 20:58:18.414702 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-xknjb" Oct 05 20:58:18 crc kubenswrapper[4754]: I1005 20:58:18.414850 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-zwh6t" Oct 05 20:58:19 crc kubenswrapper[4754]: I1005 20:58:19.174478 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-5qjvg" podUID="d5e4795f-6f22-4cde-a939-970f1ec20119" containerName="registry-server" probeResult="failure" output=< Oct 05 20:58:19 crc kubenswrapper[4754]: timeout: failed to connect service ":50051" within 1s Oct 05 20:58:19 crc kubenswrapper[4754]: > Oct 05 20:58:19 crc kubenswrapper[4754]: I1005 20:58:19.407942 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l2w5p" event={"ID":"50c8cfa3-71a9-4c5e-895c-f15457183666","Type":"ContainerStarted","Data":"f0f016d97f54585f98642d501e6940bba4c046c1d03d148eae9257499864ef47"} Oct 05 20:58:19 crc kubenswrapper[4754]: I1005 20:58:19.428813 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-l2w5p" podStartSLOduration=4.096332136 podStartE2EDuration="1m10.428791221s" podCreationTimestamp="2025-10-05 20:57:09 +0000 UTC" firstStartedPulling="2025-10-05 20:57:12.304947319 +0000 UTC m=+156.209066029" lastFinishedPulling="2025-10-05 20:58:18.637406404 +0000 UTC m=+222.541525114" observedRunningTime="2025-10-05 20:58:19.428483404 +0000 UTC m=+223.332602114" watchObservedRunningTime="2025-10-05 20:58:19.428791221 +0000 UTC m=+223.332909931" Oct 05 20:58:19 crc kubenswrapper[4754]: I1005 20:58:19.456746 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-59nwx" podStartSLOduration=6.331187082 podStartE2EDuration="1m12.456726275s" podCreationTimestamp="2025-10-05 20:57:07 +0000 UTC" firstStartedPulling="2025-10-05 20:57:11.244516096 +0000 UTC m=+155.148634806" lastFinishedPulling="2025-10-05 20:58:17.370055289 +0000 UTC m=+221.274173999" observedRunningTime="2025-10-05 20:58:19.452640325 +0000 UTC m=+223.356759035" watchObservedRunningTime="2025-10-05 20:58:19.456726275 +0000 UTC m=+223.360844985" Oct 05 20:58:19 crc kubenswrapper[4754]: I1005 20:58:19.793691 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-5889x" Oct 05 20:58:19 crc kubenswrapper[4754]: I1005 20:58:19.844615 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-5889x" Oct 05 20:58:20 crc kubenswrapper[4754]: I1005 20:58:20.239714 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-l2w5p" Oct 05 20:58:20 crc kubenswrapper[4754]: I1005 20:58:20.239785 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-l2w5p" Oct 05 20:58:20 crc kubenswrapper[4754]: I1005 20:58:20.301400 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-l2w5p" Oct 05 20:58:20 crc kubenswrapper[4754]: I1005 20:58:20.581947 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-xdqrj" Oct 05 20:58:20 crc kubenswrapper[4754]: I1005 20:58:20.582014 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-xdqrj" Oct 05 20:58:20 crc kubenswrapper[4754]: I1005 20:58:20.625150 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-xdqrj" Oct 05 20:58:20 crc kubenswrapper[4754]: I1005 20:58:20.759438 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-nz2vw" Oct 05 20:58:20 crc kubenswrapper[4754]: I1005 20:58:20.759545 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-nz2vw" Oct 05 20:58:20 crc kubenswrapper[4754]: I1005 20:58:20.807327 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-nz2vw" Oct 05 20:58:21 crc kubenswrapper[4754]: I1005 20:58:21.470783 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-xdqrj" Oct 05 20:58:21 crc kubenswrapper[4754]: I1005 20:58:21.470874 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-nz2vw" Oct 05 20:58:24 crc kubenswrapper[4754]: I1005 20:58:24.512191 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-nz2vw"] Oct 05 20:58:24 crc kubenswrapper[4754]: I1005 20:58:24.512471 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-nz2vw" podUID="e5dff6e7-5693-4f7c-af40-d7b10cbb73c7" containerName="registry-server" containerID="cri-o://d33974403ef16ebac0e6d526b32622f3e89f6e3576632663f184ad4b670cb39e" gracePeriod=2 Oct 05 20:58:26 crc kubenswrapper[4754]: I1005 20:58:26.462749 4754 generic.go:334] "Generic (PLEG): container finished" podID="e5dff6e7-5693-4f7c-af40-d7b10cbb73c7" containerID="d33974403ef16ebac0e6d526b32622f3e89f6e3576632663f184ad4b670cb39e" exitCode=0 Oct 05 20:58:26 crc kubenswrapper[4754]: I1005 20:58:26.463466 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nz2vw" event={"ID":"e5dff6e7-5693-4f7c-af40-d7b10cbb73c7","Type":"ContainerDied","Data":"d33974403ef16ebac0e6d526b32622f3e89f6e3576632663f184ad4b670cb39e"} Oct 05 20:58:26 crc kubenswrapper[4754]: I1005 20:58:26.616695 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nz2vw" Oct 05 20:58:26 crc kubenswrapper[4754]: I1005 20:58:26.745718 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b5k8b\" (UniqueName: \"kubernetes.io/projected/e5dff6e7-5693-4f7c-af40-d7b10cbb73c7-kube-api-access-b5k8b\") pod \"e5dff6e7-5693-4f7c-af40-d7b10cbb73c7\" (UID: \"e5dff6e7-5693-4f7c-af40-d7b10cbb73c7\") " Oct 05 20:58:26 crc kubenswrapper[4754]: I1005 20:58:26.745885 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5dff6e7-5693-4f7c-af40-d7b10cbb73c7-catalog-content\") pod \"e5dff6e7-5693-4f7c-af40-d7b10cbb73c7\" (UID: \"e5dff6e7-5693-4f7c-af40-d7b10cbb73c7\") " Oct 05 20:58:26 crc kubenswrapper[4754]: I1005 20:58:26.746109 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5dff6e7-5693-4f7c-af40-d7b10cbb73c7-utilities\") pod \"e5dff6e7-5693-4f7c-af40-d7b10cbb73c7\" (UID: \"e5dff6e7-5693-4f7c-af40-d7b10cbb73c7\") " Oct 05 20:58:26 crc kubenswrapper[4754]: I1005 20:58:26.749211 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e5dff6e7-5693-4f7c-af40-d7b10cbb73c7-utilities" (OuterVolumeSpecName: "utilities") pod "e5dff6e7-5693-4f7c-af40-d7b10cbb73c7" (UID: "e5dff6e7-5693-4f7c-af40-d7b10cbb73c7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 20:58:26 crc kubenswrapper[4754]: I1005 20:58:26.756550 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e5dff6e7-5693-4f7c-af40-d7b10cbb73c7-kube-api-access-b5k8b" (OuterVolumeSpecName: "kube-api-access-b5k8b") pod "e5dff6e7-5693-4f7c-af40-d7b10cbb73c7" (UID: "e5dff6e7-5693-4f7c-af40-d7b10cbb73c7"). InnerVolumeSpecName "kube-api-access-b5k8b". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 20:58:26 crc kubenswrapper[4754]: I1005 20:58:26.849056 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b5k8b\" (UniqueName: \"kubernetes.io/projected/e5dff6e7-5693-4f7c-af40-d7b10cbb73c7-kube-api-access-b5k8b\") on node \"crc\" DevicePath \"\"" Oct 05 20:58:26 crc kubenswrapper[4754]: I1005 20:58:26.849612 4754 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5dff6e7-5693-4f7c-af40-d7b10cbb73c7-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 20:58:26 crc kubenswrapper[4754]: I1005 20:58:26.863417 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e5dff6e7-5693-4f7c-af40-d7b10cbb73c7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e5dff6e7-5693-4f7c-af40-d7b10cbb73c7" (UID: "e5dff6e7-5693-4f7c-af40-d7b10cbb73c7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 20:58:26 crc kubenswrapper[4754]: I1005 20:58:26.950874 4754 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5dff6e7-5693-4f7c-af40-d7b10cbb73c7-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 20:58:27 crc kubenswrapper[4754]: I1005 20:58:27.475704 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nz2vw" Oct 05 20:58:27 crc kubenswrapper[4754]: I1005 20:58:27.475700 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nz2vw" event={"ID":"e5dff6e7-5693-4f7c-af40-d7b10cbb73c7","Type":"ContainerDied","Data":"2f986afacdcb56a29c5711c4a2666164a622c988d71e1075471e7fd9ee0bbda1"} Oct 05 20:58:27 crc kubenswrapper[4754]: I1005 20:58:27.475995 4754 scope.go:117] "RemoveContainer" containerID="d33974403ef16ebac0e6d526b32622f3e89f6e3576632663f184ad4b670cb39e" Oct 05 20:58:27 crc kubenswrapper[4754]: I1005 20:58:27.508095 4754 scope.go:117] "RemoveContainer" containerID="8b226498309d87d21927e83782cf58d174034d45662688c5c47399a05d078694" Oct 05 20:58:27 crc kubenswrapper[4754]: I1005 20:58:27.526839 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-nz2vw"] Oct 05 20:58:27 crc kubenswrapper[4754]: I1005 20:58:27.531646 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-nz2vw"] Oct 05 20:58:27 crc kubenswrapper[4754]: I1005 20:58:27.552764 4754 scope.go:117] "RemoveContainer" containerID="73f7374ca314aac1bcf71496b9da4b7cc906e5d71e6a1ac6141bb10f41cb6ef0" Oct 05 20:58:28 crc kubenswrapper[4754]: I1005 20:58:28.092176 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-59nwx" Oct 05 20:58:28 crc kubenswrapper[4754]: I1005 20:58:28.092283 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-59nwx" Oct 05 20:58:28 crc kubenswrapper[4754]: I1005 20:58:28.162986 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-59nwx" Oct 05 20:58:28 crc kubenswrapper[4754]: I1005 20:58:28.223675 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-5qjvg" Oct 05 20:58:28 crc kubenswrapper[4754]: I1005 20:58:28.296913 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-5qjvg" Oct 05 20:58:28 crc kubenswrapper[4754]: I1005 20:58:28.554078 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-59nwx" Oct 05 20:58:28 crc kubenswrapper[4754]: I1005 20:58:28.847352 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e5dff6e7-5693-4f7c-af40-d7b10cbb73c7" path="/var/lib/kubelet/pods/e5dff6e7-5693-4f7c-af40-d7b10cbb73c7/volumes" Oct 05 20:58:29 crc kubenswrapper[4754]: I1005 20:58:29.110005 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-59nwx"] Oct 05 20:58:30 crc kubenswrapper[4754]: I1005 20:58:30.185112 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-gchnw" podUID="6c0dfaff-08a7-4152-8624-1973e4a4d148" containerName="oauth-openshift" containerID="cri-o://ef2a23bdc3d3f895d033e408cb22c61a2a1743aa1b90d1cb9e60b3bae60fa470" gracePeriod=15 Oct 05 20:58:30 crc kubenswrapper[4754]: I1005 20:58:30.323233 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-l2w5p" Oct 05 20:58:30 crc kubenswrapper[4754]: I1005 20:58:30.516062 4754 generic.go:334] "Generic (PLEG): container finished" podID="6c0dfaff-08a7-4152-8624-1973e4a4d148" containerID="ef2a23bdc3d3f895d033e408cb22c61a2a1743aa1b90d1cb9e60b3bae60fa470" exitCode=0 Oct 05 20:58:30 crc kubenswrapper[4754]: I1005 20:58:30.516793 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-gchnw" event={"ID":"6c0dfaff-08a7-4152-8624-1973e4a4d148","Type":"ContainerDied","Data":"ef2a23bdc3d3f895d033e408cb22c61a2a1743aa1b90d1cb9e60b3bae60fa470"} Oct 05 20:58:30 crc kubenswrapper[4754]: I1005 20:58:30.517019 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-59nwx" podUID="232ff34c-dc36-4012-9038-e10014a179cd" containerName="registry-server" containerID="cri-o://beed44416a349711a5a4660be2a8a715e97d73dfe18584e74b7b560c36f2b3a3" gracePeriod=2 Oct 05 20:58:30 crc kubenswrapper[4754]: I1005 20:58:30.670530 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-gchnw" Oct 05 20:58:30 crc kubenswrapper[4754]: I1005 20:58:30.696785 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-76cf47c974-872xf"] Oct 05 20:58:30 crc kubenswrapper[4754]: E1005 20:58:30.697045 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5dff6e7-5693-4f7c-af40-d7b10cbb73c7" containerName="extract-content" Oct 05 20:58:30 crc kubenswrapper[4754]: I1005 20:58:30.697062 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5dff6e7-5693-4f7c-af40-d7b10cbb73c7" containerName="extract-content" Oct 05 20:58:30 crc kubenswrapper[4754]: E1005 20:58:30.697073 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5dff6e7-5693-4f7c-af40-d7b10cbb73c7" containerName="registry-server" Oct 05 20:58:30 crc kubenswrapper[4754]: I1005 20:58:30.697081 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5dff6e7-5693-4f7c-af40-d7b10cbb73c7" containerName="registry-server" Oct 05 20:58:30 crc kubenswrapper[4754]: E1005 20:58:30.697574 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32533d89-e4d3-4c15-894e-cf9d09de5dfd" containerName="pruner" Oct 05 20:58:30 crc kubenswrapper[4754]: I1005 20:58:30.697590 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="32533d89-e4d3-4c15-894e-cf9d09de5dfd" containerName="pruner" Oct 05 20:58:30 crc kubenswrapper[4754]: E1005 20:58:30.697614 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c0dfaff-08a7-4152-8624-1973e4a4d148" containerName="oauth-openshift" Oct 05 20:58:30 crc kubenswrapper[4754]: I1005 20:58:30.697625 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c0dfaff-08a7-4152-8624-1973e4a4d148" containerName="oauth-openshift" Oct 05 20:58:30 crc kubenswrapper[4754]: E1005 20:58:30.697638 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5dff6e7-5693-4f7c-af40-d7b10cbb73c7" containerName="extract-utilities" Oct 05 20:58:30 crc kubenswrapper[4754]: I1005 20:58:30.697647 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5dff6e7-5693-4f7c-af40-d7b10cbb73c7" containerName="extract-utilities" Oct 05 20:58:30 crc kubenswrapper[4754]: E1005 20:58:30.697665 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7be5d67-453a-418d-8ad1-39eb78f538c1" containerName="pruner" Oct 05 20:58:30 crc kubenswrapper[4754]: I1005 20:58:30.697673 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7be5d67-453a-418d-8ad1-39eb78f538c1" containerName="pruner" Oct 05 20:58:30 crc kubenswrapper[4754]: E1005 20:58:30.697682 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60ff4c5c-9e45-47d3-b45d-a365e988342e" containerName="collect-profiles" Oct 05 20:58:30 crc kubenswrapper[4754]: I1005 20:58:30.697691 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="60ff4c5c-9e45-47d3-b45d-a365e988342e" containerName="collect-profiles" Oct 05 20:58:30 crc kubenswrapper[4754]: I1005 20:58:30.697818 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="6c0dfaff-08a7-4152-8624-1973e4a4d148" containerName="oauth-openshift" Oct 05 20:58:30 crc kubenswrapper[4754]: I1005 20:58:30.697832 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="b7be5d67-453a-418d-8ad1-39eb78f538c1" containerName="pruner" Oct 05 20:58:30 crc kubenswrapper[4754]: I1005 20:58:30.697843 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="32533d89-e4d3-4c15-894e-cf9d09de5dfd" containerName="pruner" Oct 05 20:58:30 crc kubenswrapper[4754]: I1005 20:58:30.697858 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="60ff4c5c-9e45-47d3-b45d-a365e988342e" containerName="collect-profiles" Oct 05 20:58:30 crc kubenswrapper[4754]: I1005 20:58:30.697869 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="e5dff6e7-5693-4f7c-af40-d7b10cbb73c7" containerName="registry-server" Oct 05 20:58:30 crc kubenswrapper[4754]: I1005 20:58:30.698357 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-76cf47c974-872xf" Oct 05 20:58:30 crc kubenswrapper[4754]: I1005 20:58:30.718931 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-76cf47c974-872xf"] Oct 05 20:58:30 crc kubenswrapper[4754]: I1005 20:58:30.824428 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/6c0dfaff-08a7-4152-8624-1973e4a4d148-v4-0-config-system-serving-cert\") pod \"6c0dfaff-08a7-4152-8624-1973e4a4d148\" (UID: \"6c0dfaff-08a7-4152-8624-1973e4a4d148\") " Oct 05 20:58:30 crc kubenswrapper[4754]: I1005 20:58:30.824512 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6c0dfaff-08a7-4152-8624-1973e4a4d148-v4-0-config-system-trusted-ca-bundle\") pod \"6c0dfaff-08a7-4152-8624-1973e4a4d148\" (UID: \"6c0dfaff-08a7-4152-8624-1973e4a4d148\") " Oct 05 20:58:30 crc kubenswrapper[4754]: I1005 20:58:30.824556 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/6c0dfaff-08a7-4152-8624-1973e4a4d148-v4-0-config-system-service-ca\") pod \"6c0dfaff-08a7-4152-8624-1973e4a4d148\" (UID: \"6c0dfaff-08a7-4152-8624-1973e4a4d148\") " Oct 05 20:58:30 crc kubenswrapper[4754]: I1005 20:58:30.824649 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/6c0dfaff-08a7-4152-8624-1973e4a4d148-v4-0-config-user-template-provider-selection\") pod \"6c0dfaff-08a7-4152-8624-1973e4a4d148\" (UID: \"6c0dfaff-08a7-4152-8624-1973e4a4d148\") " Oct 05 20:58:30 crc kubenswrapper[4754]: I1005 20:58:30.824705 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/6c0dfaff-08a7-4152-8624-1973e4a4d148-v4-0-config-user-template-error\") pod \"6c0dfaff-08a7-4152-8624-1973e4a4d148\" (UID: \"6c0dfaff-08a7-4152-8624-1973e4a4d148\") " Oct 05 20:58:30 crc kubenswrapper[4754]: I1005 20:58:30.824735 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/6c0dfaff-08a7-4152-8624-1973e4a4d148-audit-policies\") pod \"6c0dfaff-08a7-4152-8624-1973e4a4d148\" (UID: \"6c0dfaff-08a7-4152-8624-1973e4a4d148\") " Oct 05 20:58:30 crc kubenswrapper[4754]: I1005 20:58:30.824783 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/6c0dfaff-08a7-4152-8624-1973e4a4d148-v4-0-config-system-ocp-branding-template\") pod \"6c0dfaff-08a7-4152-8624-1973e4a4d148\" (UID: \"6c0dfaff-08a7-4152-8624-1973e4a4d148\") " Oct 05 20:58:30 crc kubenswrapper[4754]: I1005 20:58:30.824813 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/6c0dfaff-08a7-4152-8624-1973e4a4d148-v4-0-config-system-session\") pod \"6c0dfaff-08a7-4152-8624-1973e4a4d148\" (UID: \"6c0dfaff-08a7-4152-8624-1973e4a4d148\") " Oct 05 20:58:30 crc kubenswrapper[4754]: I1005 20:58:30.824840 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/6c0dfaff-08a7-4152-8624-1973e4a4d148-v4-0-config-user-idp-0-file-data\") pod \"6c0dfaff-08a7-4152-8624-1973e4a4d148\" (UID: \"6c0dfaff-08a7-4152-8624-1973e4a4d148\") " Oct 05 20:58:30 crc kubenswrapper[4754]: I1005 20:58:30.824879 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/6c0dfaff-08a7-4152-8624-1973e4a4d148-v4-0-config-user-template-login\") pod \"6c0dfaff-08a7-4152-8624-1973e4a4d148\" (UID: \"6c0dfaff-08a7-4152-8624-1973e4a4d148\") " Oct 05 20:58:30 crc kubenswrapper[4754]: I1005 20:58:30.824917 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/6c0dfaff-08a7-4152-8624-1973e4a4d148-v4-0-config-system-router-certs\") pod \"6c0dfaff-08a7-4152-8624-1973e4a4d148\" (UID: \"6c0dfaff-08a7-4152-8624-1973e4a4d148\") " Oct 05 20:58:30 crc kubenswrapper[4754]: I1005 20:58:30.824950 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mqjlq\" (UniqueName: \"kubernetes.io/projected/6c0dfaff-08a7-4152-8624-1973e4a4d148-kube-api-access-mqjlq\") pod \"6c0dfaff-08a7-4152-8624-1973e4a4d148\" (UID: \"6c0dfaff-08a7-4152-8624-1973e4a4d148\") " Oct 05 20:58:30 crc kubenswrapper[4754]: I1005 20:58:30.824984 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/6c0dfaff-08a7-4152-8624-1973e4a4d148-audit-dir\") pod \"6c0dfaff-08a7-4152-8624-1973e4a4d148\" (UID: \"6c0dfaff-08a7-4152-8624-1973e4a4d148\") " Oct 05 20:58:30 crc kubenswrapper[4754]: I1005 20:58:30.825022 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/6c0dfaff-08a7-4152-8624-1973e4a4d148-v4-0-config-system-cliconfig\") pod \"6c0dfaff-08a7-4152-8624-1973e4a4d148\" (UID: \"6c0dfaff-08a7-4152-8624-1973e4a4d148\") " Oct 05 20:58:30 crc kubenswrapper[4754]: I1005 20:58:30.825267 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ade3e9f-98c9-4fb6-ae04-abff0806435c-v4-0-config-system-serving-cert\") pod \"oauth-openshift-76cf47c974-872xf\" (UID: \"3ade3e9f-98c9-4fb6-ae04-abff0806435c\") " pod="openshift-authentication/oauth-openshift-76cf47c974-872xf" Oct 05 20:58:30 crc kubenswrapper[4754]: I1005 20:58:30.825304 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/3ade3e9f-98c9-4fb6-ae04-abff0806435c-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-76cf47c974-872xf\" (UID: \"3ade3e9f-98c9-4fb6-ae04-abff0806435c\") " pod="openshift-authentication/oauth-openshift-76cf47c974-872xf" Oct 05 20:58:30 crc kubenswrapper[4754]: I1005 20:58:30.825337 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/3ade3e9f-98c9-4fb6-ae04-abff0806435c-audit-policies\") pod \"oauth-openshift-76cf47c974-872xf\" (UID: \"3ade3e9f-98c9-4fb6-ae04-abff0806435c\") " pod="openshift-authentication/oauth-openshift-76cf47c974-872xf" Oct 05 20:58:30 crc kubenswrapper[4754]: I1005 20:58:30.825375 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/3ade3e9f-98c9-4fb6-ae04-abff0806435c-v4-0-config-system-router-certs\") pod \"oauth-openshift-76cf47c974-872xf\" (UID: \"3ade3e9f-98c9-4fb6-ae04-abff0806435c\") " pod="openshift-authentication/oauth-openshift-76cf47c974-872xf" Oct 05 20:58:30 crc kubenswrapper[4754]: I1005 20:58:30.825402 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/3ade3e9f-98c9-4fb6-ae04-abff0806435c-v4-0-config-system-cliconfig\") pod \"oauth-openshift-76cf47c974-872xf\" (UID: \"3ade3e9f-98c9-4fb6-ae04-abff0806435c\") " pod="openshift-authentication/oauth-openshift-76cf47c974-872xf" Oct 05 20:58:30 crc kubenswrapper[4754]: I1005 20:58:30.825437 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/3ade3e9f-98c9-4fb6-ae04-abff0806435c-v4-0-config-system-service-ca\") pod \"oauth-openshift-76cf47c974-872xf\" (UID: \"3ade3e9f-98c9-4fb6-ae04-abff0806435c\") " pod="openshift-authentication/oauth-openshift-76cf47c974-872xf" Oct 05 20:58:30 crc kubenswrapper[4754]: I1005 20:58:30.825464 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/3ade3e9f-98c9-4fb6-ae04-abff0806435c-v4-0-config-user-template-login\") pod \"oauth-openshift-76cf47c974-872xf\" (UID: \"3ade3e9f-98c9-4fb6-ae04-abff0806435c\") " pod="openshift-authentication/oauth-openshift-76cf47c974-872xf" Oct 05 20:58:30 crc kubenswrapper[4754]: I1005 20:58:30.825520 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/3ade3e9f-98c9-4fb6-ae04-abff0806435c-v4-0-config-system-session\") pod \"oauth-openshift-76cf47c974-872xf\" (UID: \"3ade3e9f-98c9-4fb6-ae04-abff0806435c\") " pod="openshift-authentication/oauth-openshift-76cf47c974-872xf" Oct 05 20:58:30 crc kubenswrapper[4754]: I1005 20:58:30.825556 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/3ade3e9f-98c9-4fb6-ae04-abff0806435c-v4-0-config-user-template-error\") pod \"oauth-openshift-76cf47c974-872xf\" (UID: \"3ade3e9f-98c9-4fb6-ae04-abff0806435c\") " pod="openshift-authentication/oauth-openshift-76cf47c974-872xf" Oct 05 20:58:30 crc kubenswrapper[4754]: I1005 20:58:30.825580 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/3ade3e9f-98c9-4fb6-ae04-abff0806435c-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-76cf47c974-872xf\" (UID: \"3ade3e9f-98c9-4fb6-ae04-abff0806435c\") " pod="openshift-authentication/oauth-openshift-76cf47c974-872xf" Oct 05 20:58:30 crc kubenswrapper[4754]: I1005 20:58:30.825607 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3ade3e9f-98c9-4fb6-ae04-abff0806435c-audit-dir\") pod \"oauth-openshift-76cf47c974-872xf\" (UID: \"3ade3e9f-98c9-4fb6-ae04-abff0806435c\") " pod="openshift-authentication/oauth-openshift-76cf47c974-872xf" Oct 05 20:58:30 crc kubenswrapper[4754]: I1005 20:58:30.825656 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/3ade3e9f-98c9-4fb6-ae04-abff0806435c-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-76cf47c974-872xf\" (UID: \"3ade3e9f-98c9-4fb6-ae04-abff0806435c\") " pod="openshift-authentication/oauth-openshift-76cf47c974-872xf" Oct 05 20:58:30 crc kubenswrapper[4754]: I1005 20:58:30.825707 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lnbvq\" (UniqueName: \"kubernetes.io/projected/3ade3e9f-98c9-4fb6-ae04-abff0806435c-kube-api-access-lnbvq\") pod \"oauth-openshift-76cf47c974-872xf\" (UID: \"3ade3e9f-98c9-4fb6-ae04-abff0806435c\") " pod="openshift-authentication/oauth-openshift-76cf47c974-872xf" Oct 05 20:58:30 crc kubenswrapper[4754]: I1005 20:58:30.825734 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3ade3e9f-98c9-4fb6-ae04-abff0806435c-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-76cf47c974-872xf\" (UID: \"3ade3e9f-98c9-4fb6-ae04-abff0806435c\") " pod="openshift-authentication/oauth-openshift-76cf47c974-872xf" Oct 05 20:58:30 crc kubenswrapper[4754]: I1005 20:58:30.825840 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6c0dfaff-08a7-4152-8624-1973e4a4d148-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "6c0dfaff-08a7-4152-8624-1973e4a4d148" (UID: "6c0dfaff-08a7-4152-8624-1973e4a4d148"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 20:58:30 crc kubenswrapper[4754]: I1005 20:58:30.825866 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6c0dfaff-08a7-4152-8624-1973e4a4d148-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "6c0dfaff-08a7-4152-8624-1973e4a4d148" (UID: "6c0dfaff-08a7-4152-8624-1973e4a4d148"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 20:58:30 crc kubenswrapper[4754]: I1005 20:58:30.825918 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6c0dfaff-08a7-4152-8624-1973e4a4d148-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "6c0dfaff-08a7-4152-8624-1973e4a4d148" (UID: "6c0dfaff-08a7-4152-8624-1973e4a4d148"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 05 20:58:30 crc kubenswrapper[4754]: I1005 20:58:30.826332 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6c0dfaff-08a7-4152-8624-1973e4a4d148-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "6c0dfaff-08a7-4152-8624-1973e4a4d148" (UID: "6c0dfaff-08a7-4152-8624-1973e4a4d148"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 20:58:30 crc kubenswrapper[4754]: I1005 20:58:30.826482 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6c0dfaff-08a7-4152-8624-1973e4a4d148-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "6c0dfaff-08a7-4152-8624-1973e4a4d148" (UID: "6c0dfaff-08a7-4152-8624-1973e4a4d148"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 20:58:30 crc kubenswrapper[4754]: I1005 20:58:30.841164 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c0dfaff-08a7-4152-8624-1973e4a4d148-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "6c0dfaff-08a7-4152-8624-1973e4a4d148" (UID: "6c0dfaff-08a7-4152-8624-1973e4a4d148"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 20:58:30 crc kubenswrapper[4754]: I1005 20:58:30.841432 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c0dfaff-08a7-4152-8624-1973e4a4d148-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "6c0dfaff-08a7-4152-8624-1973e4a4d148" (UID: "6c0dfaff-08a7-4152-8624-1973e4a4d148"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 20:58:30 crc kubenswrapper[4754]: I1005 20:58:30.841781 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c0dfaff-08a7-4152-8624-1973e4a4d148-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "6c0dfaff-08a7-4152-8624-1973e4a4d148" (UID: "6c0dfaff-08a7-4152-8624-1973e4a4d148"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 20:58:30 crc kubenswrapper[4754]: I1005 20:58:30.842268 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c0dfaff-08a7-4152-8624-1973e4a4d148-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "6c0dfaff-08a7-4152-8624-1973e4a4d148" (UID: "6c0dfaff-08a7-4152-8624-1973e4a4d148"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 20:58:30 crc kubenswrapper[4754]: I1005 20:58:30.842414 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c0dfaff-08a7-4152-8624-1973e4a4d148-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "6c0dfaff-08a7-4152-8624-1973e4a4d148" (UID: "6c0dfaff-08a7-4152-8624-1973e4a4d148"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 20:58:30 crc kubenswrapper[4754]: I1005 20:58:30.842906 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c0dfaff-08a7-4152-8624-1973e4a4d148-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "6c0dfaff-08a7-4152-8624-1973e4a4d148" (UID: "6c0dfaff-08a7-4152-8624-1973e4a4d148"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 20:58:30 crc kubenswrapper[4754]: I1005 20:58:30.843186 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c0dfaff-08a7-4152-8624-1973e4a4d148-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "6c0dfaff-08a7-4152-8624-1973e4a4d148" (UID: "6c0dfaff-08a7-4152-8624-1973e4a4d148"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 20:58:30 crc kubenswrapper[4754]: I1005 20:58:30.848043 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6c0dfaff-08a7-4152-8624-1973e4a4d148-kube-api-access-mqjlq" (OuterVolumeSpecName: "kube-api-access-mqjlq") pod "6c0dfaff-08a7-4152-8624-1973e4a4d148" (UID: "6c0dfaff-08a7-4152-8624-1973e4a4d148"). InnerVolumeSpecName "kube-api-access-mqjlq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 20:58:30 crc kubenswrapper[4754]: I1005 20:58:30.848872 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c0dfaff-08a7-4152-8624-1973e4a4d148-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "6c0dfaff-08a7-4152-8624-1973e4a4d148" (UID: "6c0dfaff-08a7-4152-8624-1973e4a4d148"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 20:58:30 crc kubenswrapper[4754]: I1005 20:58:30.915555 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-l2w5p"] Oct 05 20:58:30 crc kubenswrapper[4754]: I1005 20:58:30.916344 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-l2w5p" podUID="50c8cfa3-71a9-4c5e-895c-f15457183666" containerName="registry-server" containerID="cri-o://f0f016d97f54585f98642d501e6940bba4c046c1d03d148eae9257499864ef47" gracePeriod=2 Oct 05 20:58:30 crc kubenswrapper[4754]: I1005 20:58:30.924504 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-59nwx" Oct 05 20:58:30 crc kubenswrapper[4754]: I1005 20:58:30.928485 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/3ade3e9f-98c9-4fb6-ae04-abff0806435c-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-76cf47c974-872xf\" (UID: \"3ade3e9f-98c9-4fb6-ae04-abff0806435c\") " pod="openshift-authentication/oauth-openshift-76cf47c974-872xf" Oct 05 20:58:30 crc kubenswrapper[4754]: I1005 20:58:30.928585 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lnbvq\" (UniqueName: \"kubernetes.io/projected/3ade3e9f-98c9-4fb6-ae04-abff0806435c-kube-api-access-lnbvq\") pod \"oauth-openshift-76cf47c974-872xf\" (UID: \"3ade3e9f-98c9-4fb6-ae04-abff0806435c\") " pod="openshift-authentication/oauth-openshift-76cf47c974-872xf" Oct 05 20:58:30 crc kubenswrapper[4754]: I1005 20:58:30.928609 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3ade3e9f-98c9-4fb6-ae04-abff0806435c-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-76cf47c974-872xf\" (UID: \"3ade3e9f-98c9-4fb6-ae04-abff0806435c\") " pod="openshift-authentication/oauth-openshift-76cf47c974-872xf" Oct 05 20:58:30 crc kubenswrapper[4754]: I1005 20:58:30.928653 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ade3e9f-98c9-4fb6-ae04-abff0806435c-v4-0-config-system-serving-cert\") pod \"oauth-openshift-76cf47c974-872xf\" (UID: \"3ade3e9f-98c9-4fb6-ae04-abff0806435c\") " pod="openshift-authentication/oauth-openshift-76cf47c974-872xf" Oct 05 20:58:30 crc kubenswrapper[4754]: I1005 20:58:30.928670 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/3ade3e9f-98c9-4fb6-ae04-abff0806435c-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-76cf47c974-872xf\" (UID: \"3ade3e9f-98c9-4fb6-ae04-abff0806435c\") " pod="openshift-authentication/oauth-openshift-76cf47c974-872xf" Oct 05 20:58:30 crc kubenswrapper[4754]: I1005 20:58:30.928692 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/3ade3e9f-98c9-4fb6-ae04-abff0806435c-audit-policies\") pod \"oauth-openshift-76cf47c974-872xf\" (UID: \"3ade3e9f-98c9-4fb6-ae04-abff0806435c\") " pod="openshift-authentication/oauth-openshift-76cf47c974-872xf" Oct 05 20:58:30 crc kubenswrapper[4754]: I1005 20:58:30.928718 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/3ade3e9f-98c9-4fb6-ae04-abff0806435c-v4-0-config-system-router-certs\") pod \"oauth-openshift-76cf47c974-872xf\" (UID: \"3ade3e9f-98c9-4fb6-ae04-abff0806435c\") " pod="openshift-authentication/oauth-openshift-76cf47c974-872xf" Oct 05 20:58:30 crc kubenswrapper[4754]: I1005 20:58:30.928737 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/3ade3e9f-98c9-4fb6-ae04-abff0806435c-v4-0-config-system-cliconfig\") pod \"oauth-openshift-76cf47c974-872xf\" (UID: \"3ade3e9f-98c9-4fb6-ae04-abff0806435c\") " pod="openshift-authentication/oauth-openshift-76cf47c974-872xf" Oct 05 20:58:30 crc kubenswrapper[4754]: I1005 20:58:30.928771 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/3ade3e9f-98c9-4fb6-ae04-abff0806435c-v4-0-config-system-service-ca\") pod \"oauth-openshift-76cf47c974-872xf\" (UID: \"3ade3e9f-98c9-4fb6-ae04-abff0806435c\") " pod="openshift-authentication/oauth-openshift-76cf47c974-872xf" Oct 05 20:58:30 crc kubenswrapper[4754]: I1005 20:58:30.928788 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/3ade3e9f-98c9-4fb6-ae04-abff0806435c-v4-0-config-user-template-login\") pod \"oauth-openshift-76cf47c974-872xf\" (UID: \"3ade3e9f-98c9-4fb6-ae04-abff0806435c\") " pod="openshift-authentication/oauth-openshift-76cf47c974-872xf" Oct 05 20:58:30 crc kubenswrapper[4754]: I1005 20:58:30.928807 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/3ade3e9f-98c9-4fb6-ae04-abff0806435c-v4-0-config-system-session\") pod \"oauth-openshift-76cf47c974-872xf\" (UID: \"3ade3e9f-98c9-4fb6-ae04-abff0806435c\") " pod="openshift-authentication/oauth-openshift-76cf47c974-872xf" Oct 05 20:58:30 crc kubenswrapper[4754]: I1005 20:58:30.928829 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/3ade3e9f-98c9-4fb6-ae04-abff0806435c-v4-0-config-user-template-error\") pod \"oauth-openshift-76cf47c974-872xf\" (UID: \"3ade3e9f-98c9-4fb6-ae04-abff0806435c\") " pod="openshift-authentication/oauth-openshift-76cf47c974-872xf" Oct 05 20:58:30 crc kubenswrapper[4754]: I1005 20:58:30.928847 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/3ade3e9f-98c9-4fb6-ae04-abff0806435c-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-76cf47c974-872xf\" (UID: \"3ade3e9f-98c9-4fb6-ae04-abff0806435c\") " pod="openshift-authentication/oauth-openshift-76cf47c974-872xf" Oct 05 20:58:30 crc kubenswrapper[4754]: I1005 20:58:30.928865 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3ade3e9f-98c9-4fb6-ae04-abff0806435c-audit-dir\") pod \"oauth-openshift-76cf47c974-872xf\" (UID: \"3ade3e9f-98c9-4fb6-ae04-abff0806435c\") " pod="openshift-authentication/oauth-openshift-76cf47c974-872xf" Oct 05 20:58:30 crc kubenswrapper[4754]: I1005 20:58:30.928946 4754 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/6c0dfaff-08a7-4152-8624-1973e4a4d148-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Oct 05 20:58:30 crc kubenswrapper[4754]: I1005 20:58:30.928958 4754 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/6c0dfaff-08a7-4152-8624-1973e4a4d148-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 05 20:58:30 crc kubenswrapper[4754]: I1005 20:58:30.928970 4754 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/6c0dfaff-08a7-4152-8624-1973e4a4d148-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Oct 05 20:58:30 crc kubenswrapper[4754]: I1005 20:58:30.928980 4754 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/6c0dfaff-08a7-4152-8624-1973e4a4d148-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Oct 05 20:58:30 crc kubenswrapper[4754]: I1005 20:58:30.928992 4754 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/6c0dfaff-08a7-4152-8624-1973e4a4d148-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Oct 05 20:58:30 crc kubenswrapper[4754]: I1005 20:58:30.929002 4754 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/6c0dfaff-08a7-4152-8624-1973e4a4d148-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Oct 05 20:58:30 crc kubenswrapper[4754]: I1005 20:58:30.929013 4754 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/6c0dfaff-08a7-4152-8624-1973e4a4d148-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Oct 05 20:58:30 crc kubenswrapper[4754]: I1005 20:58:30.929023 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mqjlq\" (UniqueName: \"kubernetes.io/projected/6c0dfaff-08a7-4152-8624-1973e4a4d148-kube-api-access-mqjlq\") on node \"crc\" DevicePath \"\"" Oct 05 20:58:30 crc kubenswrapper[4754]: I1005 20:58:30.929033 4754 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/6c0dfaff-08a7-4152-8624-1973e4a4d148-audit-dir\") on node \"crc\" DevicePath \"\"" Oct 05 20:58:30 crc kubenswrapper[4754]: I1005 20:58:30.929042 4754 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/6c0dfaff-08a7-4152-8624-1973e4a4d148-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Oct 05 20:58:30 crc kubenswrapper[4754]: I1005 20:58:30.929052 4754 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/6c0dfaff-08a7-4152-8624-1973e4a4d148-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 05 20:58:30 crc kubenswrapper[4754]: I1005 20:58:30.929061 4754 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6c0dfaff-08a7-4152-8624-1973e4a4d148-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 20:58:30 crc kubenswrapper[4754]: I1005 20:58:30.929071 4754 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/6c0dfaff-08a7-4152-8624-1973e4a4d148-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Oct 05 20:58:30 crc kubenswrapper[4754]: I1005 20:58:30.929082 4754 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/6c0dfaff-08a7-4152-8624-1973e4a4d148-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Oct 05 20:58:30 crc kubenswrapper[4754]: I1005 20:58:30.929777 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/3ade3e9f-98c9-4fb6-ae04-abff0806435c-v4-0-config-system-cliconfig\") pod \"oauth-openshift-76cf47c974-872xf\" (UID: \"3ade3e9f-98c9-4fb6-ae04-abff0806435c\") " pod="openshift-authentication/oauth-openshift-76cf47c974-872xf" Oct 05 20:58:30 crc kubenswrapper[4754]: I1005 20:58:30.930302 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/3ade3e9f-98c9-4fb6-ae04-abff0806435c-audit-policies\") pod \"oauth-openshift-76cf47c974-872xf\" (UID: \"3ade3e9f-98c9-4fb6-ae04-abff0806435c\") " pod="openshift-authentication/oauth-openshift-76cf47c974-872xf" Oct 05 20:58:30 crc kubenswrapper[4754]: I1005 20:58:30.930740 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3ade3e9f-98c9-4fb6-ae04-abff0806435c-audit-dir\") pod \"oauth-openshift-76cf47c974-872xf\" (UID: \"3ade3e9f-98c9-4fb6-ae04-abff0806435c\") " pod="openshift-authentication/oauth-openshift-76cf47c974-872xf" Oct 05 20:58:30 crc kubenswrapper[4754]: I1005 20:58:30.932413 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3ade3e9f-98c9-4fb6-ae04-abff0806435c-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-76cf47c974-872xf\" (UID: \"3ade3e9f-98c9-4fb6-ae04-abff0806435c\") " pod="openshift-authentication/oauth-openshift-76cf47c974-872xf" Oct 05 20:58:30 crc kubenswrapper[4754]: I1005 20:58:30.933405 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/3ade3e9f-98c9-4fb6-ae04-abff0806435c-v4-0-config-system-service-ca\") pod \"oauth-openshift-76cf47c974-872xf\" (UID: \"3ade3e9f-98c9-4fb6-ae04-abff0806435c\") " pod="openshift-authentication/oauth-openshift-76cf47c974-872xf" Oct 05 20:58:30 crc kubenswrapper[4754]: I1005 20:58:30.933612 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/3ade3e9f-98c9-4fb6-ae04-abff0806435c-v4-0-config-system-router-certs\") pod \"oauth-openshift-76cf47c974-872xf\" (UID: \"3ade3e9f-98c9-4fb6-ae04-abff0806435c\") " pod="openshift-authentication/oauth-openshift-76cf47c974-872xf" Oct 05 20:58:30 crc kubenswrapper[4754]: I1005 20:58:30.936969 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/3ade3e9f-98c9-4fb6-ae04-abff0806435c-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-76cf47c974-872xf\" (UID: \"3ade3e9f-98c9-4fb6-ae04-abff0806435c\") " pod="openshift-authentication/oauth-openshift-76cf47c974-872xf" Oct 05 20:58:30 crc kubenswrapper[4754]: I1005 20:58:30.937166 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/3ade3e9f-98c9-4fb6-ae04-abff0806435c-v4-0-config-user-template-error\") pod \"oauth-openshift-76cf47c974-872xf\" (UID: \"3ade3e9f-98c9-4fb6-ae04-abff0806435c\") " pod="openshift-authentication/oauth-openshift-76cf47c974-872xf" Oct 05 20:58:30 crc kubenswrapper[4754]: I1005 20:58:30.938084 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ade3e9f-98c9-4fb6-ae04-abff0806435c-v4-0-config-system-serving-cert\") pod \"oauth-openshift-76cf47c974-872xf\" (UID: \"3ade3e9f-98c9-4fb6-ae04-abff0806435c\") " pod="openshift-authentication/oauth-openshift-76cf47c974-872xf" Oct 05 20:58:30 crc kubenswrapper[4754]: I1005 20:58:30.938460 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/3ade3e9f-98c9-4fb6-ae04-abff0806435c-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-76cf47c974-872xf\" (UID: \"3ade3e9f-98c9-4fb6-ae04-abff0806435c\") " pod="openshift-authentication/oauth-openshift-76cf47c974-872xf" Oct 05 20:58:30 crc kubenswrapper[4754]: I1005 20:58:30.941700 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/3ade3e9f-98c9-4fb6-ae04-abff0806435c-v4-0-config-user-template-login\") pod \"oauth-openshift-76cf47c974-872xf\" (UID: \"3ade3e9f-98c9-4fb6-ae04-abff0806435c\") " pod="openshift-authentication/oauth-openshift-76cf47c974-872xf" Oct 05 20:58:30 crc kubenswrapper[4754]: I1005 20:58:30.947023 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/3ade3e9f-98c9-4fb6-ae04-abff0806435c-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-76cf47c974-872xf\" (UID: \"3ade3e9f-98c9-4fb6-ae04-abff0806435c\") " pod="openshift-authentication/oauth-openshift-76cf47c974-872xf" Oct 05 20:58:30 crc kubenswrapper[4754]: I1005 20:58:30.950236 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lnbvq\" (UniqueName: \"kubernetes.io/projected/3ade3e9f-98c9-4fb6-ae04-abff0806435c-kube-api-access-lnbvq\") pod \"oauth-openshift-76cf47c974-872xf\" (UID: \"3ade3e9f-98c9-4fb6-ae04-abff0806435c\") " pod="openshift-authentication/oauth-openshift-76cf47c974-872xf" Oct 05 20:58:30 crc kubenswrapper[4754]: I1005 20:58:30.953009 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/3ade3e9f-98c9-4fb6-ae04-abff0806435c-v4-0-config-system-session\") pod \"oauth-openshift-76cf47c974-872xf\" (UID: \"3ade3e9f-98c9-4fb6-ae04-abff0806435c\") " pod="openshift-authentication/oauth-openshift-76cf47c974-872xf" Oct 05 20:58:31 crc kubenswrapper[4754]: I1005 20:58:31.018979 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-76cf47c974-872xf" Oct 05 20:58:31 crc kubenswrapper[4754]: I1005 20:58:31.030366 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t7p85\" (UniqueName: \"kubernetes.io/projected/232ff34c-dc36-4012-9038-e10014a179cd-kube-api-access-t7p85\") pod \"232ff34c-dc36-4012-9038-e10014a179cd\" (UID: \"232ff34c-dc36-4012-9038-e10014a179cd\") " Oct 05 20:58:31 crc kubenswrapper[4754]: I1005 20:58:31.030565 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/232ff34c-dc36-4012-9038-e10014a179cd-catalog-content\") pod \"232ff34c-dc36-4012-9038-e10014a179cd\" (UID: \"232ff34c-dc36-4012-9038-e10014a179cd\") " Oct 05 20:58:31 crc kubenswrapper[4754]: I1005 20:58:31.030611 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/232ff34c-dc36-4012-9038-e10014a179cd-utilities\") pod \"232ff34c-dc36-4012-9038-e10014a179cd\" (UID: \"232ff34c-dc36-4012-9038-e10014a179cd\") " Oct 05 20:58:31 crc kubenswrapper[4754]: I1005 20:58:31.031662 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/232ff34c-dc36-4012-9038-e10014a179cd-utilities" (OuterVolumeSpecName: "utilities") pod "232ff34c-dc36-4012-9038-e10014a179cd" (UID: "232ff34c-dc36-4012-9038-e10014a179cd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 20:58:31 crc kubenswrapper[4754]: I1005 20:58:31.036248 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/232ff34c-dc36-4012-9038-e10014a179cd-kube-api-access-t7p85" (OuterVolumeSpecName: "kube-api-access-t7p85") pod "232ff34c-dc36-4012-9038-e10014a179cd" (UID: "232ff34c-dc36-4012-9038-e10014a179cd"). InnerVolumeSpecName "kube-api-access-t7p85". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 20:58:31 crc kubenswrapper[4754]: I1005 20:58:31.080392 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/232ff34c-dc36-4012-9038-e10014a179cd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "232ff34c-dc36-4012-9038-e10014a179cd" (UID: "232ff34c-dc36-4012-9038-e10014a179cd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 20:58:31 crc kubenswrapper[4754]: I1005 20:58:31.132401 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t7p85\" (UniqueName: \"kubernetes.io/projected/232ff34c-dc36-4012-9038-e10014a179cd-kube-api-access-t7p85\") on node \"crc\" DevicePath \"\"" Oct 05 20:58:31 crc kubenswrapper[4754]: I1005 20:58:31.132923 4754 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/232ff34c-dc36-4012-9038-e10014a179cd-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 20:58:31 crc kubenswrapper[4754]: I1005 20:58:31.132937 4754 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/232ff34c-dc36-4012-9038-e10014a179cd-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 20:58:31 crc kubenswrapper[4754]: I1005 20:58:31.262630 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-l2w5p" Oct 05 20:58:31 crc kubenswrapper[4754]: I1005 20:58:31.437845 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/50c8cfa3-71a9-4c5e-895c-f15457183666-catalog-content\") pod \"50c8cfa3-71a9-4c5e-895c-f15457183666\" (UID: \"50c8cfa3-71a9-4c5e-895c-f15457183666\") " Oct 05 20:58:31 crc kubenswrapper[4754]: I1005 20:58:31.437980 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/50c8cfa3-71a9-4c5e-895c-f15457183666-utilities\") pod \"50c8cfa3-71a9-4c5e-895c-f15457183666\" (UID: \"50c8cfa3-71a9-4c5e-895c-f15457183666\") " Oct 05 20:58:31 crc kubenswrapper[4754]: I1005 20:58:31.438013 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6wq48\" (UniqueName: \"kubernetes.io/projected/50c8cfa3-71a9-4c5e-895c-f15457183666-kube-api-access-6wq48\") pod \"50c8cfa3-71a9-4c5e-895c-f15457183666\" (UID: \"50c8cfa3-71a9-4c5e-895c-f15457183666\") " Oct 05 20:58:31 crc kubenswrapper[4754]: I1005 20:58:31.439222 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/50c8cfa3-71a9-4c5e-895c-f15457183666-utilities" (OuterVolumeSpecName: "utilities") pod "50c8cfa3-71a9-4c5e-895c-f15457183666" (UID: "50c8cfa3-71a9-4c5e-895c-f15457183666"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 20:58:31 crc kubenswrapper[4754]: I1005 20:58:31.445817 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/50c8cfa3-71a9-4c5e-895c-f15457183666-kube-api-access-6wq48" (OuterVolumeSpecName: "kube-api-access-6wq48") pod "50c8cfa3-71a9-4c5e-895c-f15457183666" (UID: "50c8cfa3-71a9-4c5e-895c-f15457183666"). InnerVolumeSpecName "kube-api-access-6wq48". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 20:58:31 crc kubenswrapper[4754]: I1005 20:58:31.452463 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-76cf47c974-872xf"] Oct 05 20:58:31 crc kubenswrapper[4754]: I1005 20:58:31.459362 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/50c8cfa3-71a9-4c5e-895c-f15457183666-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "50c8cfa3-71a9-4c5e-895c-f15457183666" (UID: "50c8cfa3-71a9-4c5e-895c-f15457183666"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 20:58:31 crc kubenswrapper[4754]: I1005 20:58:31.513943 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-5qjvg"] Oct 05 20:58:31 crc kubenswrapper[4754]: I1005 20:58:31.514247 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-5qjvg" podUID="d5e4795f-6f22-4cde-a939-970f1ec20119" containerName="registry-server" containerID="cri-o://c011f2109a4ff30788314d105b743e703d46bc1ffcc11537bc4d9abde8d93702" gracePeriod=2 Oct 05 20:58:31 crc kubenswrapper[4754]: I1005 20:58:31.534641 4754 generic.go:334] "Generic (PLEG): container finished" podID="50c8cfa3-71a9-4c5e-895c-f15457183666" containerID="f0f016d97f54585f98642d501e6940bba4c046c1d03d148eae9257499864ef47" exitCode=0 Oct 05 20:58:31 crc kubenswrapper[4754]: I1005 20:58:31.534750 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l2w5p" event={"ID":"50c8cfa3-71a9-4c5e-895c-f15457183666","Type":"ContainerDied","Data":"f0f016d97f54585f98642d501e6940bba4c046c1d03d148eae9257499864ef47"} Oct 05 20:58:31 crc kubenswrapper[4754]: I1005 20:58:31.534790 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l2w5p" event={"ID":"50c8cfa3-71a9-4c5e-895c-f15457183666","Type":"ContainerDied","Data":"5ef1db83196df95c7d5303633ad6f3f36fc6aaf19fd40f10bbbf747b851b55b2"} Oct 05 20:58:31 crc kubenswrapper[4754]: I1005 20:58:31.534828 4754 scope.go:117] "RemoveContainer" containerID="f0f016d97f54585f98642d501e6940bba4c046c1d03d148eae9257499864ef47" Oct 05 20:58:31 crc kubenswrapper[4754]: I1005 20:58:31.534991 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-l2w5p" Oct 05 20:58:31 crc kubenswrapper[4754]: I1005 20:58:31.539054 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-76cf47c974-872xf" event={"ID":"3ade3e9f-98c9-4fb6-ae04-abff0806435c","Type":"ContainerStarted","Data":"5be3807d71e2ed5d60ba5861285ab6a06344489fc458ad53e370d2a7c1c040ca"} Oct 05 20:58:31 crc kubenswrapper[4754]: I1005 20:58:31.540454 4754 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/50c8cfa3-71a9-4c5e-895c-f15457183666-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 20:58:31 crc kubenswrapper[4754]: I1005 20:58:31.540560 4754 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/50c8cfa3-71a9-4c5e-895c-f15457183666-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 20:58:31 crc kubenswrapper[4754]: I1005 20:58:31.540639 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6wq48\" (UniqueName: \"kubernetes.io/projected/50c8cfa3-71a9-4c5e-895c-f15457183666-kube-api-access-6wq48\") on node \"crc\" DevicePath \"\"" Oct 05 20:58:31 crc kubenswrapper[4754]: I1005 20:58:31.543002 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-gchnw" event={"ID":"6c0dfaff-08a7-4152-8624-1973e4a4d148","Type":"ContainerDied","Data":"7fbaca9acf75fd1b626ab41f3763fae48be5ed525f7498ed5c9fe6eb230a2537"} Oct 05 20:58:31 crc kubenswrapper[4754]: I1005 20:58:31.543052 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-gchnw" Oct 05 20:58:31 crc kubenswrapper[4754]: I1005 20:58:31.547405 4754 generic.go:334] "Generic (PLEG): container finished" podID="232ff34c-dc36-4012-9038-e10014a179cd" containerID="beed44416a349711a5a4660be2a8a715e97d73dfe18584e74b7b560c36f2b3a3" exitCode=0 Oct 05 20:58:31 crc kubenswrapper[4754]: I1005 20:58:31.547458 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-59nwx" event={"ID":"232ff34c-dc36-4012-9038-e10014a179cd","Type":"ContainerDied","Data":"beed44416a349711a5a4660be2a8a715e97d73dfe18584e74b7b560c36f2b3a3"} Oct 05 20:58:31 crc kubenswrapper[4754]: I1005 20:58:31.547485 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-59nwx" event={"ID":"232ff34c-dc36-4012-9038-e10014a179cd","Type":"ContainerDied","Data":"f2262d22304f374a26601238d2879127c3c5921e6d1ca7f4ce13c6b08b7174b9"} Oct 05 20:58:31 crc kubenswrapper[4754]: I1005 20:58:31.547625 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-59nwx" Oct 05 20:58:31 crc kubenswrapper[4754]: I1005 20:58:31.582188 4754 scope.go:117] "RemoveContainer" containerID="7cbaef3132e2bdb3a993930804668b94e03510605a6cd26ce96e56c8e4277b5a" Oct 05 20:58:31 crc kubenswrapper[4754]: I1005 20:58:31.590425 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-gchnw"] Oct 05 20:58:31 crc kubenswrapper[4754]: I1005 20:58:31.599725 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-gchnw"] Oct 05 20:58:31 crc kubenswrapper[4754]: I1005 20:58:31.606979 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-l2w5p"] Oct 05 20:58:31 crc kubenswrapper[4754]: I1005 20:58:31.610236 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-l2w5p"] Oct 05 20:58:31 crc kubenswrapper[4754]: I1005 20:58:31.681733 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-59nwx"] Oct 05 20:58:31 crc kubenswrapper[4754]: I1005 20:58:31.685024 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-59nwx"] Oct 05 20:58:31 crc kubenswrapper[4754]: I1005 20:58:31.687698 4754 scope.go:117] "RemoveContainer" containerID="2d164208c1e65f5aef85b54dc315b995b6c67703f87f75c98d92a28f0d7aa2ad" Oct 05 20:58:31 crc kubenswrapper[4754]: I1005 20:58:31.708719 4754 scope.go:117] "RemoveContainer" containerID="f0f016d97f54585f98642d501e6940bba4c046c1d03d148eae9257499864ef47" Oct 05 20:58:31 crc kubenswrapper[4754]: E1005 20:58:31.712939 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f0f016d97f54585f98642d501e6940bba4c046c1d03d148eae9257499864ef47\": container with ID starting with f0f016d97f54585f98642d501e6940bba4c046c1d03d148eae9257499864ef47 not found: ID does not exist" containerID="f0f016d97f54585f98642d501e6940bba4c046c1d03d148eae9257499864ef47" Oct 05 20:58:31 crc kubenswrapper[4754]: I1005 20:58:31.712979 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f0f016d97f54585f98642d501e6940bba4c046c1d03d148eae9257499864ef47"} err="failed to get container status \"f0f016d97f54585f98642d501e6940bba4c046c1d03d148eae9257499864ef47\": rpc error: code = NotFound desc = could not find container \"f0f016d97f54585f98642d501e6940bba4c046c1d03d148eae9257499864ef47\": container with ID starting with f0f016d97f54585f98642d501e6940bba4c046c1d03d148eae9257499864ef47 not found: ID does not exist" Oct 05 20:58:31 crc kubenswrapper[4754]: I1005 20:58:31.713005 4754 scope.go:117] "RemoveContainer" containerID="7cbaef3132e2bdb3a993930804668b94e03510605a6cd26ce96e56c8e4277b5a" Oct 05 20:58:31 crc kubenswrapper[4754]: E1005 20:58:31.713558 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7cbaef3132e2bdb3a993930804668b94e03510605a6cd26ce96e56c8e4277b5a\": container with ID starting with 7cbaef3132e2bdb3a993930804668b94e03510605a6cd26ce96e56c8e4277b5a not found: ID does not exist" containerID="7cbaef3132e2bdb3a993930804668b94e03510605a6cd26ce96e56c8e4277b5a" Oct 05 20:58:31 crc kubenswrapper[4754]: I1005 20:58:31.713615 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7cbaef3132e2bdb3a993930804668b94e03510605a6cd26ce96e56c8e4277b5a"} err="failed to get container status \"7cbaef3132e2bdb3a993930804668b94e03510605a6cd26ce96e56c8e4277b5a\": rpc error: code = NotFound desc = could not find container \"7cbaef3132e2bdb3a993930804668b94e03510605a6cd26ce96e56c8e4277b5a\": container with ID starting with 7cbaef3132e2bdb3a993930804668b94e03510605a6cd26ce96e56c8e4277b5a not found: ID does not exist" Oct 05 20:58:31 crc kubenswrapper[4754]: I1005 20:58:31.713661 4754 scope.go:117] "RemoveContainer" containerID="2d164208c1e65f5aef85b54dc315b995b6c67703f87f75c98d92a28f0d7aa2ad" Oct 05 20:58:31 crc kubenswrapper[4754]: E1005 20:58:31.713993 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2d164208c1e65f5aef85b54dc315b995b6c67703f87f75c98d92a28f0d7aa2ad\": container with ID starting with 2d164208c1e65f5aef85b54dc315b995b6c67703f87f75c98d92a28f0d7aa2ad not found: ID does not exist" containerID="2d164208c1e65f5aef85b54dc315b995b6c67703f87f75c98d92a28f0d7aa2ad" Oct 05 20:58:31 crc kubenswrapper[4754]: I1005 20:58:31.714018 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2d164208c1e65f5aef85b54dc315b995b6c67703f87f75c98d92a28f0d7aa2ad"} err="failed to get container status \"2d164208c1e65f5aef85b54dc315b995b6c67703f87f75c98d92a28f0d7aa2ad\": rpc error: code = NotFound desc = could not find container \"2d164208c1e65f5aef85b54dc315b995b6c67703f87f75c98d92a28f0d7aa2ad\": container with ID starting with 2d164208c1e65f5aef85b54dc315b995b6c67703f87f75c98d92a28f0d7aa2ad not found: ID does not exist" Oct 05 20:58:31 crc kubenswrapper[4754]: I1005 20:58:31.714036 4754 scope.go:117] "RemoveContainer" containerID="ef2a23bdc3d3f895d033e408cb22c61a2a1743aa1b90d1cb9e60b3bae60fa470" Oct 05 20:58:31 crc kubenswrapper[4754]: I1005 20:58:31.799882 4754 scope.go:117] "RemoveContainer" containerID="beed44416a349711a5a4660be2a8a715e97d73dfe18584e74b7b560c36f2b3a3" Oct 05 20:58:31 crc kubenswrapper[4754]: I1005 20:58:31.817109 4754 scope.go:117] "RemoveContainer" containerID="8f4390632ec901428a65e37624ed36ac020cb98eccefecc6caa6156dcd884c2e" Oct 05 20:58:31 crc kubenswrapper[4754]: I1005 20:58:31.838524 4754 scope.go:117] "RemoveContainer" containerID="d6e446ac94cd1146dd3cc09eeb151e991447a00f7179a8d9f675e578ac774d23" Oct 05 20:58:31 crc kubenswrapper[4754]: I1005 20:58:31.852710 4754 scope.go:117] "RemoveContainer" containerID="beed44416a349711a5a4660be2a8a715e97d73dfe18584e74b7b560c36f2b3a3" Oct 05 20:58:31 crc kubenswrapper[4754]: E1005 20:58:31.853251 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"beed44416a349711a5a4660be2a8a715e97d73dfe18584e74b7b560c36f2b3a3\": container with ID starting with beed44416a349711a5a4660be2a8a715e97d73dfe18584e74b7b560c36f2b3a3 not found: ID does not exist" containerID="beed44416a349711a5a4660be2a8a715e97d73dfe18584e74b7b560c36f2b3a3" Oct 05 20:58:31 crc kubenswrapper[4754]: I1005 20:58:31.853306 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"beed44416a349711a5a4660be2a8a715e97d73dfe18584e74b7b560c36f2b3a3"} err="failed to get container status \"beed44416a349711a5a4660be2a8a715e97d73dfe18584e74b7b560c36f2b3a3\": rpc error: code = NotFound desc = could not find container \"beed44416a349711a5a4660be2a8a715e97d73dfe18584e74b7b560c36f2b3a3\": container with ID starting with beed44416a349711a5a4660be2a8a715e97d73dfe18584e74b7b560c36f2b3a3 not found: ID does not exist" Oct 05 20:58:31 crc kubenswrapper[4754]: I1005 20:58:31.853342 4754 scope.go:117] "RemoveContainer" containerID="8f4390632ec901428a65e37624ed36ac020cb98eccefecc6caa6156dcd884c2e" Oct 05 20:58:31 crc kubenswrapper[4754]: E1005 20:58:31.854189 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8f4390632ec901428a65e37624ed36ac020cb98eccefecc6caa6156dcd884c2e\": container with ID starting with 8f4390632ec901428a65e37624ed36ac020cb98eccefecc6caa6156dcd884c2e not found: ID does not exist" containerID="8f4390632ec901428a65e37624ed36ac020cb98eccefecc6caa6156dcd884c2e" Oct 05 20:58:31 crc kubenswrapper[4754]: I1005 20:58:31.854225 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f4390632ec901428a65e37624ed36ac020cb98eccefecc6caa6156dcd884c2e"} err="failed to get container status \"8f4390632ec901428a65e37624ed36ac020cb98eccefecc6caa6156dcd884c2e\": rpc error: code = NotFound desc = could not find container \"8f4390632ec901428a65e37624ed36ac020cb98eccefecc6caa6156dcd884c2e\": container with ID starting with 8f4390632ec901428a65e37624ed36ac020cb98eccefecc6caa6156dcd884c2e not found: ID does not exist" Oct 05 20:58:31 crc kubenswrapper[4754]: I1005 20:58:31.854250 4754 scope.go:117] "RemoveContainer" containerID="d6e446ac94cd1146dd3cc09eeb151e991447a00f7179a8d9f675e578ac774d23" Oct 05 20:58:31 crc kubenswrapper[4754]: E1005 20:58:31.854777 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d6e446ac94cd1146dd3cc09eeb151e991447a00f7179a8d9f675e578ac774d23\": container with ID starting with d6e446ac94cd1146dd3cc09eeb151e991447a00f7179a8d9f675e578ac774d23 not found: ID does not exist" containerID="d6e446ac94cd1146dd3cc09eeb151e991447a00f7179a8d9f675e578ac774d23" Oct 05 20:58:31 crc kubenswrapper[4754]: I1005 20:58:31.854830 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d6e446ac94cd1146dd3cc09eeb151e991447a00f7179a8d9f675e578ac774d23"} err="failed to get container status \"d6e446ac94cd1146dd3cc09eeb151e991447a00f7179a8d9f675e578ac774d23\": rpc error: code = NotFound desc = could not find container \"d6e446ac94cd1146dd3cc09eeb151e991447a00f7179a8d9f675e578ac774d23\": container with ID starting with d6e446ac94cd1146dd3cc09eeb151e991447a00f7179a8d9f675e578ac774d23 not found: ID does not exist" Oct 05 20:58:31 crc kubenswrapper[4754]: I1005 20:58:31.871148 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5qjvg" Oct 05 20:58:31 crc kubenswrapper[4754]: I1005 20:58:31.946603 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d5e4795f-6f22-4cde-a939-970f1ec20119-catalog-content\") pod \"d5e4795f-6f22-4cde-a939-970f1ec20119\" (UID: \"d5e4795f-6f22-4cde-a939-970f1ec20119\") " Oct 05 20:58:31 crc kubenswrapper[4754]: I1005 20:58:31.946665 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d5e4795f-6f22-4cde-a939-970f1ec20119-utilities\") pod \"d5e4795f-6f22-4cde-a939-970f1ec20119\" (UID: \"d5e4795f-6f22-4cde-a939-970f1ec20119\") " Oct 05 20:58:31 crc kubenswrapper[4754]: I1005 20:58:31.946707 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h8xtz\" (UniqueName: \"kubernetes.io/projected/d5e4795f-6f22-4cde-a939-970f1ec20119-kube-api-access-h8xtz\") pod \"d5e4795f-6f22-4cde-a939-970f1ec20119\" (UID: \"d5e4795f-6f22-4cde-a939-970f1ec20119\") " Oct 05 20:58:31 crc kubenswrapper[4754]: I1005 20:58:31.947400 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d5e4795f-6f22-4cde-a939-970f1ec20119-utilities" (OuterVolumeSpecName: "utilities") pod "d5e4795f-6f22-4cde-a939-970f1ec20119" (UID: "d5e4795f-6f22-4cde-a939-970f1ec20119"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 20:58:31 crc kubenswrapper[4754]: I1005 20:58:31.948730 4754 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d5e4795f-6f22-4cde-a939-970f1ec20119-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 20:58:31 crc kubenswrapper[4754]: I1005 20:58:31.957943 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d5e4795f-6f22-4cde-a939-970f1ec20119-kube-api-access-h8xtz" (OuterVolumeSpecName: "kube-api-access-h8xtz") pod "d5e4795f-6f22-4cde-a939-970f1ec20119" (UID: "d5e4795f-6f22-4cde-a939-970f1ec20119"). InnerVolumeSpecName "kube-api-access-h8xtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 20:58:31 crc kubenswrapper[4754]: I1005 20:58:31.995657 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d5e4795f-6f22-4cde-a939-970f1ec20119-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d5e4795f-6f22-4cde-a939-970f1ec20119" (UID: "d5e4795f-6f22-4cde-a939-970f1ec20119"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 20:58:32 crc kubenswrapper[4754]: I1005 20:58:32.049812 4754 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d5e4795f-6f22-4cde-a939-970f1ec20119-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 20:58:32 crc kubenswrapper[4754]: I1005 20:58:32.049884 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h8xtz\" (UniqueName: \"kubernetes.io/projected/d5e4795f-6f22-4cde-a939-970f1ec20119-kube-api-access-h8xtz\") on node \"crc\" DevicePath \"\"" Oct 05 20:58:32 crc kubenswrapper[4754]: I1005 20:58:32.560605 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-76cf47c974-872xf" event={"ID":"3ade3e9f-98c9-4fb6-ae04-abff0806435c","Type":"ContainerStarted","Data":"971bab5d05014147701fa939fb023968f525729add0b901960734581a3e00747"} Oct 05 20:58:32 crc kubenswrapper[4754]: I1005 20:58:32.562301 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-76cf47c974-872xf" Oct 05 20:58:32 crc kubenswrapper[4754]: I1005 20:58:32.569265 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-76cf47c974-872xf" Oct 05 20:58:32 crc kubenswrapper[4754]: I1005 20:58:32.570973 4754 generic.go:334] "Generic (PLEG): container finished" podID="d5e4795f-6f22-4cde-a939-970f1ec20119" containerID="c011f2109a4ff30788314d105b743e703d46bc1ffcc11537bc4d9abde8d93702" exitCode=0 Oct 05 20:58:32 crc kubenswrapper[4754]: I1005 20:58:32.571048 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5qjvg" Oct 05 20:58:32 crc kubenswrapper[4754]: I1005 20:58:32.571055 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5qjvg" event={"ID":"d5e4795f-6f22-4cde-a939-970f1ec20119","Type":"ContainerDied","Data":"c011f2109a4ff30788314d105b743e703d46bc1ffcc11537bc4d9abde8d93702"} Oct 05 20:58:32 crc kubenswrapper[4754]: I1005 20:58:32.571122 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5qjvg" event={"ID":"d5e4795f-6f22-4cde-a939-970f1ec20119","Type":"ContainerDied","Data":"fffbb9b87fd079439336591f5d8cf694f7d51b2dfaf8d3b0f883da2b3b59c443"} Oct 05 20:58:32 crc kubenswrapper[4754]: I1005 20:58:32.571167 4754 scope.go:117] "RemoveContainer" containerID="c011f2109a4ff30788314d105b743e703d46bc1ffcc11537bc4d9abde8d93702" Oct 05 20:58:32 crc kubenswrapper[4754]: I1005 20:58:32.614868 4754 scope.go:117] "RemoveContainer" containerID="533f474b59c58333faac7190e98ef2bb9bf1ea468e6acef27965f264795053e0" Oct 05 20:58:32 crc kubenswrapper[4754]: I1005 20:58:32.645613 4754 scope.go:117] "RemoveContainer" containerID="d76d2da9cc0c7659052d793be6ef3070a7f1d693f236cce3327c8a5e99e14728" Oct 05 20:58:32 crc kubenswrapper[4754]: I1005 20:58:32.657364 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-76cf47c974-872xf" podStartSLOduration=27.657326621 podStartE2EDuration="27.657326621s" podCreationTimestamp="2025-10-05 20:58:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 20:58:32.601322711 +0000 UTC m=+236.505441471" watchObservedRunningTime="2025-10-05 20:58:32.657326621 +0000 UTC m=+236.561445371" Oct 05 20:58:32 crc kubenswrapper[4754]: I1005 20:58:32.683558 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-5qjvg"] Oct 05 20:58:32 crc kubenswrapper[4754]: I1005 20:58:32.689342 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-5qjvg"] Oct 05 20:58:32 crc kubenswrapper[4754]: I1005 20:58:32.717328 4754 scope.go:117] "RemoveContainer" containerID="c011f2109a4ff30788314d105b743e703d46bc1ffcc11537bc4d9abde8d93702" Oct 05 20:58:32 crc kubenswrapper[4754]: E1005 20:58:32.718030 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c011f2109a4ff30788314d105b743e703d46bc1ffcc11537bc4d9abde8d93702\": container with ID starting with c011f2109a4ff30788314d105b743e703d46bc1ffcc11537bc4d9abde8d93702 not found: ID does not exist" containerID="c011f2109a4ff30788314d105b743e703d46bc1ffcc11537bc4d9abde8d93702" Oct 05 20:58:32 crc kubenswrapper[4754]: I1005 20:58:32.718082 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c011f2109a4ff30788314d105b743e703d46bc1ffcc11537bc4d9abde8d93702"} err="failed to get container status \"c011f2109a4ff30788314d105b743e703d46bc1ffcc11537bc4d9abde8d93702\": rpc error: code = NotFound desc = could not find container \"c011f2109a4ff30788314d105b743e703d46bc1ffcc11537bc4d9abde8d93702\": container with ID starting with c011f2109a4ff30788314d105b743e703d46bc1ffcc11537bc4d9abde8d93702 not found: ID does not exist" Oct 05 20:58:32 crc kubenswrapper[4754]: I1005 20:58:32.718124 4754 scope.go:117] "RemoveContainer" containerID="533f474b59c58333faac7190e98ef2bb9bf1ea468e6acef27965f264795053e0" Oct 05 20:58:32 crc kubenswrapper[4754]: E1005 20:58:32.718413 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"533f474b59c58333faac7190e98ef2bb9bf1ea468e6acef27965f264795053e0\": container with ID starting with 533f474b59c58333faac7190e98ef2bb9bf1ea468e6acef27965f264795053e0 not found: ID does not exist" containerID="533f474b59c58333faac7190e98ef2bb9bf1ea468e6acef27965f264795053e0" Oct 05 20:58:32 crc kubenswrapper[4754]: I1005 20:58:32.718473 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"533f474b59c58333faac7190e98ef2bb9bf1ea468e6acef27965f264795053e0"} err="failed to get container status \"533f474b59c58333faac7190e98ef2bb9bf1ea468e6acef27965f264795053e0\": rpc error: code = NotFound desc = could not find container \"533f474b59c58333faac7190e98ef2bb9bf1ea468e6acef27965f264795053e0\": container with ID starting with 533f474b59c58333faac7190e98ef2bb9bf1ea468e6acef27965f264795053e0 not found: ID does not exist" Oct 05 20:58:32 crc kubenswrapper[4754]: I1005 20:58:32.718510 4754 scope.go:117] "RemoveContainer" containerID="d76d2da9cc0c7659052d793be6ef3070a7f1d693f236cce3327c8a5e99e14728" Oct 05 20:58:32 crc kubenswrapper[4754]: E1005 20:58:32.719261 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d76d2da9cc0c7659052d793be6ef3070a7f1d693f236cce3327c8a5e99e14728\": container with ID starting with d76d2da9cc0c7659052d793be6ef3070a7f1d693f236cce3327c8a5e99e14728 not found: ID does not exist" containerID="d76d2da9cc0c7659052d793be6ef3070a7f1d693f236cce3327c8a5e99e14728" Oct 05 20:58:32 crc kubenswrapper[4754]: I1005 20:58:32.719319 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d76d2da9cc0c7659052d793be6ef3070a7f1d693f236cce3327c8a5e99e14728"} err="failed to get container status \"d76d2da9cc0c7659052d793be6ef3070a7f1d693f236cce3327c8a5e99e14728\": rpc error: code = NotFound desc = could not find container \"d76d2da9cc0c7659052d793be6ef3070a7f1d693f236cce3327c8a5e99e14728\": container with ID starting with d76d2da9cc0c7659052d793be6ef3070a7f1d693f236cce3327c8a5e99e14728 not found: ID does not exist" Oct 05 20:58:32 crc kubenswrapper[4754]: I1005 20:58:32.846773 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="232ff34c-dc36-4012-9038-e10014a179cd" path="/var/lib/kubelet/pods/232ff34c-dc36-4012-9038-e10014a179cd/volumes" Oct 05 20:58:32 crc kubenswrapper[4754]: I1005 20:58:32.847685 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="50c8cfa3-71a9-4c5e-895c-f15457183666" path="/var/lib/kubelet/pods/50c8cfa3-71a9-4c5e-895c-f15457183666/volumes" Oct 05 20:58:32 crc kubenswrapper[4754]: I1005 20:58:32.848346 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6c0dfaff-08a7-4152-8624-1973e4a4d148" path="/var/lib/kubelet/pods/6c0dfaff-08a7-4152-8624-1973e4a4d148/volumes" Oct 05 20:58:32 crc kubenswrapper[4754]: I1005 20:58:32.849312 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d5e4795f-6f22-4cde-a939-970f1ec20119" path="/var/lib/kubelet/pods/d5e4795f-6f22-4cde-a939-970f1ec20119/volumes" Oct 05 20:59:03 crc kubenswrapper[4754]: I1005 20:59:03.239230 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-xknjb"] Oct 05 20:59:03 crc kubenswrapper[4754]: I1005 20:59:03.251671 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-xknjb" podUID="b5b25be1-b393-4d98-aaf0-3f9693ba89d7" containerName="registry-server" containerID="cri-o://8f06273caa8a4b6742bd4cedff3d16610bf1ecc0d97633d087c4942fbbb30f70" gracePeriod=30 Oct 05 20:59:03 crc kubenswrapper[4754]: I1005 20:59:03.253090 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zwh6t"] Oct 05 20:59:03 crc kubenswrapper[4754]: I1005 20:59:03.253527 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-zwh6t" podUID="85d11cf6-caa4-4ee2-a8ee-48c44469d4b7" containerName="registry-server" containerID="cri-o://df11ecc12c5fb6fe0d12cdd353c9d061a0f848224746d9ac5db943911be9d0ff" gracePeriod=30 Oct 05 20:59:03 crc kubenswrapper[4754]: I1005 20:59:03.265408 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-ll5jq"] Oct 05 20:59:03 crc kubenswrapper[4754]: I1005 20:59:03.266067 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-ll5jq" podUID="c1c4fc78-c305-4dba-b4d5-9f05d58a8ef6" containerName="marketplace-operator" containerID="cri-o://2bc17b5eff1449d472fbd2fb6a61eaa7f22cf8a370a4c5c170b1d3dc3ee2ef1d" gracePeriod=30 Oct 05 20:59:03 crc kubenswrapper[4754]: I1005 20:59:03.278415 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5889x"] Oct 05 20:59:03 crc kubenswrapper[4754]: I1005 20:59:03.283837 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-5889x" podUID="61cebf9f-dd0d-40f9-9bd5-1ff0dd7038f8" containerName="registry-server" containerID="cri-o://36c567a32b080330ac66b125a7e5a956a5cbe8f6902f9406a5ac19b451119129" gracePeriod=30 Oct 05 20:59:03 crc kubenswrapper[4754]: I1005 20:59:03.290383 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-xdqrj"] Oct 05 20:59:03 crc kubenswrapper[4754]: I1005 20:59:03.290730 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-xdqrj" podUID="caf99098-6640-4077-8b97-69d1bfefa99c" containerName="registry-server" containerID="cri-o://4839e09a747a0b0ba463d4adc2e850e0ab819456da4ec507586bf54f71323ba8" gracePeriod=30 Oct 05 20:59:03 crc kubenswrapper[4754]: I1005 20:59:03.310253 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-xftcm"] Oct 05 20:59:03 crc kubenswrapper[4754]: E1005 20:59:03.310471 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50c8cfa3-71a9-4c5e-895c-f15457183666" containerName="registry-server" Oct 05 20:59:03 crc kubenswrapper[4754]: I1005 20:59:03.310504 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="50c8cfa3-71a9-4c5e-895c-f15457183666" containerName="registry-server" Oct 05 20:59:03 crc kubenswrapper[4754]: E1005 20:59:03.310516 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="232ff34c-dc36-4012-9038-e10014a179cd" containerName="extract-utilities" Oct 05 20:59:03 crc kubenswrapper[4754]: I1005 20:59:03.310523 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="232ff34c-dc36-4012-9038-e10014a179cd" containerName="extract-utilities" Oct 05 20:59:03 crc kubenswrapper[4754]: E1005 20:59:03.310534 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5e4795f-6f22-4cde-a939-970f1ec20119" containerName="extract-content" Oct 05 20:59:03 crc kubenswrapper[4754]: I1005 20:59:03.310540 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5e4795f-6f22-4cde-a939-970f1ec20119" containerName="extract-content" Oct 05 20:59:03 crc kubenswrapper[4754]: E1005 20:59:03.310553 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5e4795f-6f22-4cde-a939-970f1ec20119" containerName="extract-utilities" Oct 05 20:59:03 crc kubenswrapper[4754]: I1005 20:59:03.310560 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5e4795f-6f22-4cde-a939-970f1ec20119" containerName="extract-utilities" Oct 05 20:59:03 crc kubenswrapper[4754]: E1005 20:59:03.310571 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="232ff34c-dc36-4012-9038-e10014a179cd" containerName="registry-server" Oct 05 20:59:03 crc kubenswrapper[4754]: I1005 20:59:03.310577 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="232ff34c-dc36-4012-9038-e10014a179cd" containerName="registry-server" Oct 05 20:59:03 crc kubenswrapper[4754]: E1005 20:59:03.310584 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50c8cfa3-71a9-4c5e-895c-f15457183666" containerName="extract-content" Oct 05 20:59:03 crc kubenswrapper[4754]: I1005 20:59:03.310590 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="50c8cfa3-71a9-4c5e-895c-f15457183666" containerName="extract-content" Oct 05 20:59:03 crc kubenswrapper[4754]: E1005 20:59:03.310600 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="232ff34c-dc36-4012-9038-e10014a179cd" containerName="extract-content" Oct 05 20:59:03 crc kubenswrapper[4754]: I1005 20:59:03.310606 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="232ff34c-dc36-4012-9038-e10014a179cd" containerName="extract-content" Oct 05 20:59:03 crc kubenswrapper[4754]: E1005 20:59:03.310613 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50c8cfa3-71a9-4c5e-895c-f15457183666" containerName="extract-utilities" Oct 05 20:59:03 crc kubenswrapper[4754]: I1005 20:59:03.310620 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="50c8cfa3-71a9-4c5e-895c-f15457183666" containerName="extract-utilities" Oct 05 20:59:03 crc kubenswrapper[4754]: E1005 20:59:03.310626 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5e4795f-6f22-4cde-a939-970f1ec20119" containerName="registry-server" Oct 05 20:59:03 crc kubenswrapper[4754]: I1005 20:59:03.310632 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5e4795f-6f22-4cde-a939-970f1ec20119" containerName="registry-server" Oct 05 20:59:03 crc kubenswrapper[4754]: I1005 20:59:03.310719 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="d5e4795f-6f22-4cde-a939-970f1ec20119" containerName="registry-server" Oct 05 20:59:03 crc kubenswrapper[4754]: I1005 20:59:03.310728 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="232ff34c-dc36-4012-9038-e10014a179cd" containerName="registry-server" Oct 05 20:59:03 crc kubenswrapper[4754]: I1005 20:59:03.310736 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="50c8cfa3-71a9-4c5e-895c-f15457183666" containerName="registry-server" Oct 05 20:59:03 crc kubenswrapper[4754]: I1005 20:59:03.311113 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-xftcm" Oct 05 20:59:03 crc kubenswrapper[4754]: I1005 20:59:03.332178 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-xftcm"] Oct 05 20:59:03 crc kubenswrapper[4754]: I1005 20:59:03.353726 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ntw8v\" (UniqueName: \"kubernetes.io/projected/ed62afda-bbd2-4326-8101-0abe44ecb2f5-kube-api-access-ntw8v\") pod \"marketplace-operator-79b997595-xftcm\" (UID: \"ed62afda-bbd2-4326-8101-0abe44ecb2f5\") " pod="openshift-marketplace/marketplace-operator-79b997595-xftcm" Oct 05 20:59:03 crc kubenswrapper[4754]: I1005 20:59:03.353811 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/ed62afda-bbd2-4326-8101-0abe44ecb2f5-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-xftcm\" (UID: \"ed62afda-bbd2-4326-8101-0abe44ecb2f5\") " pod="openshift-marketplace/marketplace-operator-79b997595-xftcm" Oct 05 20:59:03 crc kubenswrapper[4754]: I1005 20:59:03.353870 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ed62afda-bbd2-4326-8101-0abe44ecb2f5-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-xftcm\" (UID: \"ed62afda-bbd2-4326-8101-0abe44ecb2f5\") " pod="openshift-marketplace/marketplace-operator-79b997595-xftcm" Oct 05 20:59:03 crc kubenswrapper[4754]: I1005 20:59:03.454711 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/ed62afda-bbd2-4326-8101-0abe44ecb2f5-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-xftcm\" (UID: \"ed62afda-bbd2-4326-8101-0abe44ecb2f5\") " pod="openshift-marketplace/marketplace-operator-79b997595-xftcm" Oct 05 20:59:03 crc kubenswrapper[4754]: I1005 20:59:03.454771 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ed62afda-bbd2-4326-8101-0abe44ecb2f5-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-xftcm\" (UID: \"ed62afda-bbd2-4326-8101-0abe44ecb2f5\") " pod="openshift-marketplace/marketplace-operator-79b997595-xftcm" Oct 05 20:59:03 crc kubenswrapper[4754]: I1005 20:59:03.454822 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ntw8v\" (UniqueName: \"kubernetes.io/projected/ed62afda-bbd2-4326-8101-0abe44ecb2f5-kube-api-access-ntw8v\") pod \"marketplace-operator-79b997595-xftcm\" (UID: \"ed62afda-bbd2-4326-8101-0abe44ecb2f5\") " pod="openshift-marketplace/marketplace-operator-79b997595-xftcm" Oct 05 20:59:03 crc kubenswrapper[4754]: I1005 20:59:03.456238 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ed62afda-bbd2-4326-8101-0abe44ecb2f5-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-xftcm\" (UID: \"ed62afda-bbd2-4326-8101-0abe44ecb2f5\") " pod="openshift-marketplace/marketplace-operator-79b997595-xftcm" Oct 05 20:59:03 crc kubenswrapper[4754]: I1005 20:59:03.462678 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/ed62afda-bbd2-4326-8101-0abe44ecb2f5-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-xftcm\" (UID: \"ed62afda-bbd2-4326-8101-0abe44ecb2f5\") " pod="openshift-marketplace/marketplace-operator-79b997595-xftcm" Oct 05 20:59:03 crc kubenswrapper[4754]: I1005 20:59:03.497166 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ntw8v\" (UniqueName: \"kubernetes.io/projected/ed62afda-bbd2-4326-8101-0abe44ecb2f5-kube-api-access-ntw8v\") pod \"marketplace-operator-79b997595-xftcm\" (UID: \"ed62afda-bbd2-4326-8101-0abe44ecb2f5\") " pod="openshift-marketplace/marketplace-operator-79b997595-xftcm" Oct 05 20:59:03 crc kubenswrapper[4754]: I1005 20:59:03.626735 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-xftcm" Oct 05 20:59:03 crc kubenswrapper[4754]: I1005 20:59:03.784546 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-ll5jq" Oct 05 20:59:03 crc kubenswrapper[4754]: I1005 20:59:03.828676 4754 generic.go:334] "Generic (PLEG): container finished" podID="61cebf9f-dd0d-40f9-9bd5-1ff0dd7038f8" containerID="36c567a32b080330ac66b125a7e5a956a5cbe8f6902f9406a5ac19b451119129" exitCode=0 Oct 05 20:59:03 crc kubenswrapper[4754]: I1005 20:59:03.828745 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5889x" event={"ID":"61cebf9f-dd0d-40f9-9bd5-1ff0dd7038f8","Type":"ContainerDied","Data":"36c567a32b080330ac66b125a7e5a956a5cbe8f6902f9406a5ac19b451119129"} Oct 05 20:59:03 crc kubenswrapper[4754]: I1005 20:59:03.851437 4754 generic.go:334] "Generic (PLEG): container finished" podID="b5b25be1-b393-4d98-aaf0-3f9693ba89d7" containerID="8f06273caa8a4b6742bd4cedff3d16610bf1ecc0d97633d087c4942fbbb30f70" exitCode=0 Oct 05 20:59:03 crc kubenswrapper[4754]: I1005 20:59:03.851647 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xknjb" event={"ID":"b5b25be1-b393-4d98-aaf0-3f9693ba89d7","Type":"ContainerDied","Data":"8f06273caa8a4b6742bd4cedff3d16610bf1ecc0d97633d087c4942fbbb30f70"} Oct 05 20:59:03 crc kubenswrapper[4754]: I1005 20:59:03.856936 4754 generic.go:334] "Generic (PLEG): container finished" podID="caf99098-6640-4077-8b97-69d1bfefa99c" containerID="4839e09a747a0b0ba463d4adc2e850e0ab819456da4ec507586bf54f71323ba8" exitCode=0 Oct 05 20:59:03 crc kubenswrapper[4754]: I1005 20:59:03.857005 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xdqrj" event={"ID":"caf99098-6640-4077-8b97-69d1bfefa99c","Type":"ContainerDied","Data":"4839e09a747a0b0ba463d4adc2e850e0ab819456da4ec507586bf54f71323ba8"} Oct 05 20:59:03 crc kubenswrapper[4754]: I1005 20:59:03.860788 4754 generic.go:334] "Generic (PLEG): container finished" podID="c1c4fc78-c305-4dba-b4d5-9f05d58a8ef6" containerID="2bc17b5eff1449d472fbd2fb6a61eaa7f22cf8a370a4c5c170b1d3dc3ee2ef1d" exitCode=0 Oct 05 20:59:03 crc kubenswrapper[4754]: I1005 20:59:03.861837 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-ll5jq" Oct 05 20:59:03 crc kubenswrapper[4754]: I1005 20:59:03.862596 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-ll5jq" event={"ID":"c1c4fc78-c305-4dba-b4d5-9f05d58a8ef6","Type":"ContainerDied","Data":"2bc17b5eff1449d472fbd2fb6a61eaa7f22cf8a370a4c5c170b1d3dc3ee2ef1d"} Oct 05 20:59:03 crc kubenswrapper[4754]: I1005 20:59:03.862624 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-ll5jq" event={"ID":"c1c4fc78-c305-4dba-b4d5-9f05d58a8ef6","Type":"ContainerDied","Data":"1775ab67489ca4d3c34d5d072cb2215fe929ee07d5f7e5d8cda18d28393efd0a"} Oct 05 20:59:03 crc kubenswrapper[4754]: I1005 20:59:03.862643 4754 scope.go:117] "RemoveContainer" containerID="2bc17b5eff1449d472fbd2fb6a61eaa7f22cf8a370a4c5c170b1d3dc3ee2ef1d" Oct 05 20:59:03 crc kubenswrapper[4754]: I1005 20:59:03.863106 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z9jr9\" (UniqueName: \"kubernetes.io/projected/c1c4fc78-c305-4dba-b4d5-9f05d58a8ef6-kube-api-access-z9jr9\") pod \"c1c4fc78-c305-4dba-b4d5-9f05d58a8ef6\" (UID: \"c1c4fc78-c305-4dba-b4d5-9f05d58a8ef6\") " Oct 05 20:59:03 crc kubenswrapper[4754]: I1005 20:59:03.863213 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c1c4fc78-c305-4dba-b4d5-9f05d58a8ef6-marketplace-trusted-ca\") pod \"c1c4fc78-c305-4dba-b4d5-9f05d58a8ef6\" (UID: \"c1c4fc78-c305-4dba-b4d5-9f05d58a8ef6\") " Oct 05 20:59:03 crc kubenswrapper[4754]: I1005 20:59:03.863318 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/c1c4fc78-c305-4dba-b4d5-9f05d58a8ef6-marketplace-operator-metrics\") pod \"c1c4fc78-c305-4dba-b4d5-9f05d58a8ef6\" (UID: \"c1c4fc78-c305-4dba-b4d5-9f05d58a8ef6\") " Oct 05 20:59:03 crc kubenswrapper[4754]: I1005 20:59:03.865337 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c1c4fc78-c305-4dba-b4d5-9f05d58a8ef6-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "c1c4fc78-c305-4dba-b4d5-9f05d58a8ef6" (UID: "c1c4fc78-c305-4dba-b4d5-9f05d58a8ef6"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 20:59:03 crc kubenswrapper[4754]: I1005 20:59:03.869286 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c1c4fc78-c305-4dba-b4d5-9f05d58a8ef6-kube-api-access-z9jr9" (OuterVolumeSpecName: "kube-api-access-z9jr9") pod "c1c4fc78-c305-4dba-b4d5-9f05d58a8ef6" (UID: "c1c4fc78-c305-4dba-b4d5-9f05d58a8ef6"). InnerVolumeSpecName "kube-api-access-z9jr9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 20:59:03 crc kubenswrapper[4754]: I1005 20:59:03.887167 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xknjb" Oct 05 20:59:03 crc kubenswrapper[4754]: I1005 20:59:03.900000 4754 scope.go:117] "RemoveContainer" containerID="2bc17b5eff1449d472fbd2fb6a61eaa7f22cf8a370a4c5c170b1d3dc3ee2ef1d" Oct 05 20:59:03 crc kubenswrapper[4754]: E1005 20:59:03.901281 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2bc17b5eff1449d472fbd2fb6a61eaa7f22cf8a370a4c5c170b1d3dc3ee2ef1d\": container with ID starting with 2bc17b5eff1449d472fbd2fb6a61eaa7f22cf8a370a4c5c170b1d3dc3ee2ef1d not found: ID does not exist" containerID="2bc17b5eff1449d472fbd2fb6a61eaa7f22cf8a370a4c5c170b1d3dc3ee2ef1d" Oct 05 20:59:03 crc kubenswrapper[4754]: I1005 20:59:03.901312 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2bc17b5eff1449d472fbd2fb6a61eaa7f22cf8a370a4c5c170b1d3dc3ee2ef1d"} err="failed to get container status \"2bc17b5eff1449d472fbd2fb6a61eaa7f22cf8a370a4c5c170b1d3dc3ee2ef1d\": rpc error: code = NotFound desc = could not find container \"2bc17b5eff1449d472fbd2fb6a61eaa7f22cf8a370a4c5c170b1d3dc3ee2ef1d\": container with ID starting with 2bc17b5eff1449d472fbd2fb6a61eaa7f22cf8a370a4c5c170b1d3dc3ee2ef1d not found: ID does not exist" Oct 05 20:59:03 crc kubenswrapper[4754]: I1005 20:59:03.901738 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c1c4fc78-c305-4dba-b4d5-9f05d58a8ef6-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "c1c4fc78-c305-4dba-b4d5-9f05d58a8ef6" (UID: "c1c4fc78-c305-4dba-b4d5-9f05d58a8ef6"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 20:59:03 crc kubenswrapper[4754]: I1005 20:59:03.903526 4754 generic.go:334] "Generic (PLEG): container finished" podID="85d11cf6-caa4-4ee2-a8ee-48c44469d4b7" containerID="df11ecc12c5fb6fe0d12cdd353c9d061a0f848224746d9ac5db943911be9d0ff" exitCode=0 Oct 05 20:59:03 crc kubenswrapper[4754]: I1005 20:59:03.903570 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zwh6t" event={"ID":"85d11cf6-caa4-4ee2-a8ee-48c44469d4b7","Type":"ContainerDied","Data":"df11ecc12c5fb6fe0d12cdd353c9d061a0f848224746d9ac5db943911be9d0ff"} Oct 05 20:59:03 crc kubenswrapper[4754]: I1005 20:59:03.904569 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5889x" Oct 05 20:59:03 crc kubenswrapper[4754]: I1005 20:59:03.962587 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-xftcm"] Oct 05 20:59:03 crc kubenswrapper[4754]: I1005 20:59:03.976116 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-68wh5\" (UniqueName: \"kubernetes.io/projected/61cebf9f-dd0d-40f9-9bd5-1ff0dd7038f8-kube-api-access-68wh5\") pod \"61cebf9f-dd0d-40f9-9bd5-1ff0dd7038f8\" (UID: \"61cebf9f-dd0d-40f9-9bd5-1ff0dd7038f8\") " Oct 05 20:59:03 crc kubenswrapper[4754]: I1005 20:59:03.976295 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/61cebf9f-dd0d-40f9-9bd5-1ff0dd7038f8-utilities\") pod \"61cebf9f-dd0d-40f9-9bd5-1ff0dd7038f8\" (UID: \"61cebf9f-dd0d-40f9-9bd5-1ff0dd7038f8\") " Oct 05 20:59:03 crc kubenswrapper[4754]: I1005 20:59:03.976359 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p86kl\" (UniqueName: \"kubernetes.io/projected/b5b25be1-b393-4d98-aaf0-3f9693ba89d7-kube-api-access-p86kl\") pod \"b5b25be1-b393-4d98-aaf0-3f9693ba89d7\" (UID: \"b5b25be1-b393-4d98-aaf0-3f9693ba89d7\") " Oct 05 20:59:03 crc kubenswrapper[4754]: I1005 20:59:03.976424 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b5b25be1-b393-4d98-aaf0-3f9693ba89d7-catalog-content\") pod \"b5b25be1-b393-4d98-aaf0-3f9693ba89d7\" (UID: \"b5b25be1-b393-4d98-aaf0-3f9693ba89d7\") " Oct 05 20:59:03 crc kubenswrapper[4754]: I1005 20:59:03.976512 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/61cebf9f-dd0d-40f9-9bd5-1ff0dd7038f8-catalog-content\") pod \"61cebf9f-dd0d-40f9-9bd5-1ff0dd7038f8\" (UID: \"61cebf9f-dd0d-40f9-9bd5-1ff0dd7038f8\") " Oct 05 20:59:03 crc kubenswrapper[4754]: I1005 20:59:03.976559 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b5b25be1-b393-4d98-aaf0-3f9693ba89d7-utilities\") pod \"b5b25be1-b393-4d98-aaf0-3f9693ba89d7\" (UID: \"b5b25be1-b393-4d98-aaf0-3f9693ba89d7\") " Oct 05 20:59:03 crc kubenswrapper[4754]: I1005 20:59:03.976918 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z9jr9\" (UniqueName: \"kubernetes.io/projected/c1c4fc78-c305-4dba-b4d5-9f05d58a8ef6-kube-api-access-z9jr9\") on node \"crc\" DevicePath \"\"" Oct 05 20:59:03 crc kubenswrapper[4754]: I1005 20:59:03.976938 4754 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c1c4fc78-c305-4dba-b4d5-9f05d58a8ef6-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 05 20:59:03 crc kubenswrapper[4754]: I1005 20:59:03.976950 4754 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/c1c4fc78-c305-4dba-b4d5-9f05d58a8ef6-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Oct 05 20:59:03 crc kubenswrapper[4754]: I1005 20:59:03.977108 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/61cebf9f-dd0d-40f9-9bd5-1ff0dd7038f8-utilities" (OuterVolumeSpecName: "utilities") pod "61cebf9f-dd0d-40f9-9bd5-1ff0dd7038f8" (UID: "61cebf9f-dd0d-40f9-9bd5-1ff0dd7038f8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 20:59:03 crc kubenswrapper[4754]: I1005 20:59:03.977751 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/61cebf9f-dd0d-40f9-9bd5-1ff0dd7038f8-kube-api-access-68wh5" (OuterVolumeSpecName: "kube-api-access-68wh5") pod "61cebf9f-dd0d-40f9-9bd5-1ff0dd7038f8" (UID: "61cebf9f-dd0d-40f9-9bd5-1ff0dd7038f8"). InnerVolumeSpecName "kube-api-access-68wh5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 20:59:03 crc kubenswrapper[4754]: I1005 20:59:03.987339 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b5b25be1-b393-4d98-aaf0-3f9693ba89d7-utilities" (OuterVolumeSpecName: "utilities") pod "b5b25be1-b393-4d98-aaf0-3f9693ba89d7" (UID: "b5b25be1-b393-4d98-aaf0-3f9693ba89d7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 20:59:04 crc kubenswrapper[4754]: I1005 20:59:04.006474 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b5b25be1-b393-4d98-aaf0-3f9693ba89d7-kube-api-access-p86kl" (OuterVolumeSpecName: "kube-api-access-p86kl") pod "b5b25be1-b393-4d98-aaf0-3f9693ba89d7" (UID: "b5b25be1-b393-4d98-aaf0-3f9693ba89d7"). InnerVolumeSpecName "kube-api-access-p86kl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 20:59:04 crc kubenswrapper[4754]: I1005 20:59:04.017368 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/61cebf9f-dd0d-40f9-9bd5-1ff0dd7038f8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "61cebf9f-dd0d-40f9-9bd5-1ff0dd7038f8" (UID: "61cebf9f-dd0d-40f9-9bd5-1ff0dd7038f8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 20:59:04 crc kubenswrapper[4754]: I1005 20:59:04.033181 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b5b25be1-b393-4d98-aaf0-3f9693ba89d7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b5b25be1-b393-4d98-aaf0-3f9693ba89d7" (UID: "b5b25be1-b393-4d98-aaf0-3f9693ba89d7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 20:59:04 crc kubenswrapper[4754]: I1005 20:59:04.079298 4754 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b5b25be1-b393-4d98-aaf0-3f9693ba89d7-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 20:59:04 crc kubenswrapper[4754]: I1005 20:59:04.079342 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-68wh5\" (UniqueName: \"kubernetes.io/projected/61cebf9f-dd0d-40f9-9bd5-1ff0dd7038f8-kube-api-access-68wh5\") on node \"crc\" DevicePath \"\"" Oct 05 20:59:04 crc kubenswrapper[4754]: I1005 20:59:04.079356 4754 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/61cebf9f-dd0d-40f9-9bd5-1ff0dd7038f8-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 20:59:04 crc kubenswrapper[4754]: I1005 20:59:04.079365 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p86kl\" (UniqueName: \"kubernetes.io/projected/b5b25be1-b393-4d98-aaf0-3f9693ba89d7-kube-api-access-p86kl\") on node \"crc\" DevicePath \"\"" Oct 05 20:59:04 crc kubenswrapper[4754]: I1005 20:59:04.079375 4754 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b5b25be1-b393-4d98-aaf0-3f9693ba89d7-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 20:59:04 crc kubenswrapper[4754]: I1005 20:59:04.079385 4754 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/61cebf9f-dd0d-40f9-9bd5-1ff0dd7038f8-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 20:59:04 crc kubenswrapper[4754]: I1005 20:59:04.166103 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zwh6t" Oct 05 20:59:04 crc kubenswrapper[4754]: I1005 20:59:04.181104 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2ngdc\" (UniqueName: \"kubernetes.io/projected/85d11cf6-caa4-4ee2-a8ee-48c44469d4b7-kube-api-access-2ngdc\") pod \"85d11cf6-caa4-4ee2-a8ee-48c44469d4b7\" (UID: \"85d11cf6-caa4-4ee2-a8ee-48c44469d4b7\") " Oct 05 20:59:04 crc kubenswrapper[4754]: I1005 20:59:04.181177 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/85d11cf6-caa4-4ee2-a8ee-48c44469d4b7-catalog-content\") pod \"85d11cf6-caa4-4ee2-a8ee-48c44469d4b7\" (UID: \"85d11cf6-caa4-4ee2-a8ee-48c44469d4b7\") " Oct 05 20:59:04 crc kubenswrapper[4754]: I1005 20:59:04.181211 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/85d11cf6-caa4-4ee2-a8ee-48c44469d4b7-utilities\") pod \"85d11cf6-caa4-4ee2-a8ee-48c44469d4b7\" (UID: \"85d11cf6-caa4-4ee2-a8ee-48c44469d4b7\") " Oct 05 20:59:04 crc kubenswrapper[4754]: I1005 20:59:04.182579 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/85d11cf6-caa4-4ee2-a8ee-48c44469d4b7-utilities" (OuterVolumeSpecName: "utilities") pod "85d11cf6-caa4-4ee2-a8ee-48c44469d4b7" (UID: "85d11cf6-caa4-4ee2-a8ee-48c44469d4b7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 20:59:04 crc kubenswrapper[4754]: I1005 20:59:04.186213 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/85d11cf6-caa4-4ee2-a8ee-48c44469d4b7-kube-api-access-2ngdc" (OuterVolumeSpecName: "kube-api-access-2ngdc") pod "85d11cf6-caa4-4ee2-a8ee-48c44469d4b7" (UID: "85d11cf6-caa4-4ee2-a8ee-48c44469d4b7"). InnerVolumeSpecName "kube-api-access-2ngdc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 20:59:04 crc kubenswrapper[4754]: I1005 20:59:04.216882 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-ll5jq"] Oct 05 20:59:04 crc kubenswrapper[4754]: I1005 20:59:04.222806 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-ll5jq"] Oct 05 20:59:04 crc kubenswrapper[4754]: I1005 20:59:04.227204 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xdqrj" Oct 05 20:59:04 crc kubenswrapper[4754]: I1005 20:59:04.283004 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9dgdm\" (UniqueName: \"kubernetes.io/projected/caf99098-6640-4077-8b97-69d1bfefa99c-kube-api-access-9dgdm\") pod \"caf99098-6640-4077-8b97-69d1bfefa99c\" (UID: \"caf99098-6640-4077-8b97-69d1bfefa99c\") " Oct 05 20:59:04 crc kubenswrapper[4754]: I1005 20:59:04.283068 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/caf99098-6640-4077-8b97-69d1bfefa99c-catalog-content\") pod \"caf99098-6640-4077-8b97-69d1bfefa99c\" (UID: \"caf99098-6640-4077-8b97-69d1bfefa99c\") " Oct 05 20:59:04 crc kubenswrapper[4754]: I1005 20:59:04.283175 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/caf99098-6640-4077-8b97-69d1bfefa99c-utilities\") pod \"caf99098-6640-4077-8b97-69d1bfefa99c\" (UID: \"caf99098-6640-4077-8b97-69d1bfefa99c\") " Oct 05 20:59:04 crc kubenswrapper[4754]: I1005 20:59:04.283728 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2ngdc\" (UniqueName: \"kubernetes.io/projected/85d11cf6-caa4-4ee2-a8ee-48c44469d4b7-kube-api-access-2ngdc\") on node \"crc\" DevicePath \"\"" Oct 05 20:59:04 crc kubenswrapper[4754]: I1005 20:59:04.283742 4754 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/85d11cf6-caa4-4ee2-a8ee-48c44469d4b7-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 20:59:04 crc kubenswrapper[4754]: I1005 20:59:04.284327 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/caf99098-6640-4077-8b97-69d1bfefa99c-utilities" (OuterVolumeSpecName: "utilities") pod "caf99098-6640-4077-8b97-69d1bfefa99c" (UID: "caf99098-6640-4077-8b97-69d1bfefa99c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 20:59:04 crc kubenswrapper[4754]: I1005 20:59:04.286343 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/caf99098-6640-4077-8b97-69d1bfefa99c-kube-api-access-9dgdm" (OuterVolumeSpecName: "kube-api-access-9dgdm") pod "caf99098-6640-4077-8b97-69d1bfefa99c" (UID: "caf99098-6640-4077-8b97-69d1bfefa99c"). InnerVolumeSpecName "kube-api-access-9dgdm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 20:59:04 crc kubenswrapper[4754]: I1005 20:59:04.292540 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/85d11cf6-caa4-4ee2-a8ee-48c44469d4b7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "85d11cf6-caa4-4ee2-a8ee-48c44469d4b7" (UID: "85d11cf6-caa4-4ee2-a8ee-48c44469d4b7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 20:59:04 crc kubenswrapper[4754]: I1005 20:59:04.371296 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/caf99098-6640-4077-8b97-69d1bfefa99c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "caf99098-6640-4077-8b97-69d1bfefa99c" (UID: "caf99098-6640-4077-8b97-69d1bfefa99c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 20:59:04 crc kubenswrapper[4754]: I1005 20:59:04.385442 4754 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/caf99098-6640-4077-8b97-69d1bfefa99c-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 20:59:04 crc kubenswrapper[4754]: I1005 20:59:04.385470 4754 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/85d11cf6-caa4-4ee2-a8ee-48c44469d4b7-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 20:59:04 crc kubenswrapper[4754]: I1005 20:59:04.385485 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9dgdm\" (UniqueName: \"kubernetes.io/projected/caf99098-6640-4077-8b97-69d1bfefa99c-kube-api-access-9dgdm\") on node \"crc\" DevicePath \"\"" Oct 05 20:59:04 crc kubenswrapper[4754]: I1005 20:59:04.385509 4754 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/caf99098-6640-4077-8b97-69d1bfefa99c-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 20:59:04 crc kubenswrapper[4754]: I1005 20:59:04.847054 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c1c4fc78-c305-4dba-b4d5-9f05d58a8ef6" path="/var/lib/kubelet/pods/c1c4fc78-c305-4dba-b4d5-9f05d58a8ef6/volumes" Oct 05 20:59:04 crc kubenswrapper[4754]: I1005 20:59:04.915144 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zwh6t" Oct 05 20:59:04 crc kubenswrapper[4754]: I1005 20:59:04.915937 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zwh6t" event={"ID":"85d11cf6-caa4-4ee2-a8ee-48c44469d4b7","Type":"ContainerDied","Data":"7e2db30679f9bb3068ecaf9bb3b848ee881c5daac364bc6c5a62aeb4a367136f"} Oct 05 20:59:04 crc kubenswrapper[4754]: I1005 20:59:04.915990 4754 scope.go:117] "RemoveContainer" containerID="df11ecc12c5fb6fe0d12cdd353c9d061a0f848224746d9ac5db943911be9d0ff" Oct 05 20:59:04 crc kubenswrapper[4754]: I1005 20:59:04.928475 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5889x" Oct 05 20:59:04 crc kubenswrapper[4754]: I1005 20:59:04.928542 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5889x" event={"ID":"61cebf9f-dd0d-40f9-9bd5-1ff0dd7038f8","Type":"ContainerDied","Data":"831673c15f0c69e2fd6dc94ca3d8411e53559dcd1108802081f757966b315067"} Oct 05 20:59:04 crc kubenswrapper[4754]: I1005 20:59:04.941790 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xknjb" event={"ID":"b5b25be1-b393-4d98-aaf0-3f9693ba89d7","Type":"ContainerDied","Data":"2502a9d5b335e202508425ca7b4c43fee5b6e09a4bacb5ae9a9b9d3d07139080"} Oct 05 20:59:04 crc kubenswrapper[4754]: I1005 20:59:04.941938 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xknjb" Oct 05 20:59:04 crc kubenswrapper[4754]: I1005 20:59:04.948127 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xdqrj" event={"ID":"caf99098-6640-4077-8b97-69d1bfefa99c","Type":"ContainerDied","Data":"96cfcf8172f5b44ba90e77feec2d1e3ee1393f09d878f09c63656a9aa222713b"} Oct 05 20:59:04 crc kubenswrapper[4754]: I1005 20:59:04.948218 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xdqrj" Oct 05 20:59:04 crc kubenswrapper[4754]: I1005 20:59:04.953893 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-xftcm" event={"ID":"ed62afda-bbd2-4326-8101-0abe44ecb2f5","Type":"ContainerStarted","Data":"96a35a4ca06b18b7ed8631a800481dc9169891c8b651ab607ee9ab52bba0afbf"} Oct 05 20:59:04 crc kubenswrapper[4754]: I1005 20:59:04.954033 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-xftcm" event={"ID":"ed62afda-bbd2-4326-8101-0abe44ecb2f5","Type":"ContainerStarted","Data":"8f53f6451342f9b7769def186b0163053b405e7a71ad52031f30765e4d068a74"} Oct 05 20:59:04 crc kubenswrapper[4754]: I1005 20:59:04.954454 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-xftcm" Oct 05 20:59:04 crc kubenswrapper[4754]: I1005 20:59:04.959007 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zwh6t"] Oct 05 20:59:04 crc kubenswrapper[4754]: I1005 20:59:04.959490 4754 scope.go:117] "RemoveContainer" containerID="25bc1b4acb780b87220b6a8388386b75b9f855100762c4b16dee83e294476085" Oct 05 20:59:04 crc kubenswrapper[4754]: I1005 20:59:04.961624 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-xftcm" Oct 05 20:59:04 crc kubenswrapper[4754]: I1005 20:59:04.963786 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-zwh6t"] Oct 05 20:59:04 crc kubenswrapper[4754]: I1005 20:59:04.973790 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-xknjb"] Oct 05 20:59:04 crc kubenswrapper[4754]: I1005 20:59:04.977279 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-xknjb"] Oct 05 20:59:04 crc kubenswrapper[4754]: I1005 20:59:04.989367 4754 scope.go:117] "RemoveContainer" containerID="ed64ccc8ba22eb586a997140263b753574a123b89d48f5dbace62ef8f13c71a5" Oct 05 20:59:04 crc kubenswrapper[4754]: I1005 20:59:04.996020 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-xdqrj"] Oct 05 20:59:04 crc kubenswrapper[4754]: I1005 20:59:04.998464 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-xdqrj"] Oct 05 20:59:05 crc kubenswrapper[4754]: I1005 20:59:05.013599 4754 scope.go:117] "RemoveContainer" containerID="36c567a32b080330ac66b125a7e5a956a5cbe8f6902f9406a5ac19b451119129" Oct 05 20:59:05 crc kubenswrapper[4754]: I1005 20:59:05.017609 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5889x"] Oct 05 20:59:05 crc kubenswrapper[4754]: I1005 20:59:05.022872 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-5889x"] Oct 05 20:59:05 crc kubenswrapper[4754]: I1005 20:59:05.023992 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-xftcm" podStartSLOduration=2.023981434 podStartE2EDuration="2.023981434s" podCreationTimestamp="2025-10-05 20:59:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 20:59:05.021730476 +0000 UTC m=+268.925849186" watchObservedRunningTime="2025-10-05 20:59:05.023981434 +0000 UTC m=+268.928100144" Oct 05 20:59:05 crc kubenswrapper[4754]: I1005 20:59:05.036990 4754 scope.go:117] "RemoveContainer" containerID="85b2ac566e0435f9eb3619b63bd81c71c3f614dc0b553c34cf9826126a37aec2" Oct 05 20:59:05 crc kubenswrapper[4754]: I1005 20:59:05.062983 4754 scope.go:117] "RemoveContainer" containerID="31cf7e5dfda9011d8644759cb52e266ca933e48a1579f6fa5323b051d9a5f29e" Oct 05 20:59:05 crc kubenswrapper[4754]: I1005 20:59:05.079037 4754 scope.go:117] "RemoveContainer" containerID="8f06273caa8a4b6742bd4cedff3d16610bf1ecc0d97633d087c4942fbbb30f70" Oct 05 20:59:05 crc kubenswrapper[4754]: I1005 20:59:05.093840 4754 scope.go:117] "RemoveContainer" containerID="04d86388c6f62ae88d96a2cded7ecd37f98d5c9ff7697d137554bbef2e3aa5c2" Oct 05 20:59:05 crc kubenswrapper[4754]: I1005 20:59:05.108300 4754 scope.go:117] "RemoveContainer" containerID="8457a90a27ace934a18c57e5a39bc1fa61bf28f1cf604f816967a9d1e074208c" Oct 05 20:59:05 crc kubenswrapper[4754]: I1005 20:59:05.119710 4754 scope.go:117] "RemoveContainer" containerID="4839e09a747a0b0ba463d4adc2e850e0ab819456da4ec507586bf54f71323ba8" Oct 05 20:59:05 crc kubenswrapper[4754]: I1005 20:59:05.130182 4754 scope.go:117] "RemoveContainer" containerID="bbedaaf423efe4f7b92787043d222dab156fbeb2eeaa05812b1b89d2d2cfc01a" Oct 05 20:59:05 crc kubenswrapper[4754]: I1005 20:59:05.141191 4754 scope.go:117] "RemoveContainer" containerID="126c9a934dc7d9adf3eb6d7c0390dda2dfae5b78c59cb48672aba70080b5915f" Oct 05 20:59:05 crc kubenswrapper[4754]: I1005 20:59:05.455258 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-qdsv6"] Oct 05 20:59:05 crc kubenswrapper[4754]: E1005 20:59:05.455456 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85d11cf6-caa4-4ee2-a8ee-48c44469d4b7" containerName="extract-utilities" Oct 05 20:59:05 crc kubenswrapper[4754]: I1005 20:59:05.455469 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="85d11cf6-caa4-4ee2-a8ee-48c44469d4b7" containerName="extract-utilities" Oct 05 20:59:05 crc kubenswrapper[4754]: E1005 20:59:05.455479 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5b25be1-b393-4d98-aaf0-3f9693ba89d7" containerName="extract-utilities" Oct 05 20:59:05 crc kubenswrapper[4754]: I1005 20:59:05.455485 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5b25be1-b393-4d98-aaf0-3f9693ba89d7" containerName="extract-utilities" Oct 05 20:59:05 crc kubenswrapper[4754]: E1005 20:59:05.455510 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61cebf9f-dd0d-40f9-9bd5-1ff0dd7038f8" containerName="extract-content" Oct 05 20:59:05 crc kubenswrapper[4754]: I1005 20:59:05.455518 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="61cebf9f-dd0d-40f9-9bd5-1ff0dd7038f8" containerName="extract-content" Oct 05 20:59:05 crc kubenswrapper[4754]: E1005 20:59:05.455529 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85d11cf6-caa4-4ee2-a8ee-48c44469d4b7" containerName="extract-content" Oct 05 20:59:05 crc kubenswrapper[4754]: I1005 20:59:05.455549 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="85d11cf6-caa4-4ee2-a8ee-48c44469d4b7" containerName="extract-content" Oct 05 20:59:05 crc kubenswrapper[4754]: E1005 20:59:05.455558 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5b25be1-b393-4d98-aaf0-3f9693ba89d7" containerName="extract-content" Oct 05 20:59:05 crc kubenswrapper[4754]: I1005 20:59:05.455565 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5b25be1-b393-4d98-aaf0-3f9693ba89d7" containerName="extract-content" Oct 05 20:59:05 crc kubenswrapper[4754]: E1005 20:59:05.455586 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="caf99098-6640-4077-8b97-69d1bfefa99c" containerName="extract-utilities" Oct 05 20:59:05 crc kubenswrapper[4754]: I1005 20:59:05.455592 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="caf99098-6640-4077-8b97-69d1bfefa99c" containerName="extract-utilities" Oct 05 20:59:05 crc kubenswrapper[4754]: E1005 20:59:05.455602 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61cebf9f-dd0d-40f9-9bd5-1ff0dd7038f8" containerName="registry-server" Oct 05 20:59:05 crc kubenswrapper[4754]: I1005 20:59:05.455608 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="61cebf9f-dd0d-40f9-9bd5-1ff0dd7038f8" containerName="registry-server" Oct 05 20:59:05 crc kubenswrapper[4754]: E1005 20:59:05.455616 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61cebf9f-dd0d-40f9-9bd5-1ff0dd7038f8" containerName="extract-utilities" Oct 05 20:59:05 crc kubenswrapper[4754]: I1005 20:59:05.455622 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="61cebf9f-dd0d-40f9-9bd5-1ff0dd7038f8" containerName="extract-utilities" Oct 05 20:59:05 crc kubenswrapper[4754]: E1005 20:59:05.455629 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5b25be1-b393-4d98-aaf0-3f9693ba89d7" containerName="registry-server" Oct 05 20:59:05 crc kubenswrapper[4754]: I1005 20:59:05.455634 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5b25be1-b393-4d98-aaf0-3f9693ba89d7" containerName="registry-server" Oct 05 20:59:05 crc kubenswrapper[4754]: E1005 20:59:05.455642 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="caf99098-6640-4077-8b97-69d1bfefa99c" containerName="extract-content" Oct 05 20:59:05 crc kubenswrapper[4754]: I1005 20:59:05.455648 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="caf99098-6640-4077-8b97-69d1bfefa99c" containerName="extract-content" Oct 05 20:59:05 crc kubenswrapper[4754]: E1005 20:59:05.455655 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="caf99098-6640-4077-8b97-69d1bfefa99c" containerName="registry-server" Oct 05 20:59:05 crc kubenswrapper[4754]: I1005 20:59:05.455661 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="caf99098-6640-4077-8b97-69d1bfefa99c" containerName="registry-server" Oct 05 20:59:05 crc kubenswrapper[4754]: E1005 20:59:05.455669 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85d11cf6-caa4-4ee2-a8ee-48c44469d4b7" containerName="registry-server" Oct 05 20:59:05 crc kubenswrapper[4754]: I1005 20:59:05.455674 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="85d11cf6-caa4-4ee2-a8ee-48c44469d4b7" containerName="registry-server" Oct 05 20:59:05 crc kubenswrapper[4754]: E1005 20:59:05.455683 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1c4fc78-c305-4dba-b4d5-9f05d58a8ef6" containerName="marketplace-operator" Oct 05 20:59:05 crc kubenswrapper[4754]: I1005 20:59:05.455689 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1c4fc78-c305-4dba-b4d5-9f05d58a8ef6" containerName="marketplace-operator" Oct 05 20:59:05 crc kubenswrapper[4754]: I1005 20:59:05.455769 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="b5b25be1-b393-4d98-aaf0-3f9693ba89d7" containerName="registry-server" Oct 05 20:59:05 crc kubenswrapper[4754]: I1005 20:59:05.455777 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="c1c4fc78-c305-4dba-b4d5-9f05d58a8ef6" containerName="marketplace-operator" Oct 05 20:59:05 crc kubenswrapper[4754]: I1005 20:59:05.455788 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="61cebf9f-dd0d-40f9-9bd5-1ff0dd7038f8" containerName="registry-server" Oct 05 20:59:05 crc kubenswrapper[4754]: I1005 20:59:05.455798 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="85d11cf6-caa4-4ee2-a8ee-48c44469d4b7" containerName="registry-server" Oct 05 20:59:05 crc kubenswrapper[4754]: I1005 20:59:05.455804 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="caf99098-6640-4077-8b97-69d1bfefa99c" containerName="registry-server" Oct 05 20:59:05 crc kubenswrapper[4754]: I1005 20:59:05.456663 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qdsv6" Oct 05 20:59:05 crc kubenswrapper[4754]: I1005 20:59:05.459244 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Oct 05 20:59:05 crc kubenswrapper[4754]: I1005 20:59:05.472138 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qdsv6"] Oct 05 20:59:05 crc kubenswrapper[4754]: I1005 20:59:05.499231 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f5951a44-c13f-4cef-a2cf-6803479067f8-catalog-content\") pod \"redhat-marketplace-qdsv6\" (UID: \"f5951a44-c13f-4cef-a2cf-6803479067f8\") " pod="openshift-marketplace/redhat-marketplace-qdsv6" Oct 05 20:59:05 crc kubenswrapper[4754]: I1005 20:59:05.499343 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ffkwv\" (UniqueName: \"kubernetes.io/projected/f5951a44-c13f-4cef-a2cf-6803479067f8-kube-api-access-ffkwv\") pod \"redhat-marketplace-qdsv6\" (UID: \"f5951a44-c13f-4cef-a2cf-6803479067f8\") " pod="openshift-marketplace/redhat-marketplace-qdsv6" Oct 05 20:59:05 crc kubenswrapper[4754]: I1005 20:59:05.499395 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f5951a44-c13f-4cef-a2cf-6803479067f8-utilities\") pod \"redhat-marketplace-qdsv6\" (UID: \"f5951a44-c13f-4cef-a2cf-6803479067f8\") " pod="openshift-marketplace/redhat-marketplace-qdsv6" Oct 05 20:59:05 crc kubenswrapper[4754]: I1005 20:59:05.600798 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f5951a44-c13f-4cef-a2cf-6803479067f8-utilities\") pod \"redhat-marketplace-qdsv6\" (UID: \"f5951a44-c13f-4cef-a2cf-6803479067f8\") " pod="openshift-marketplace/redhat-marketplace-qdsv6" Oct 05 20:59:05 crc kubenswrapper[4754]: I1005 20:59:05.600878 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f5951a44-c13f-4cef-a2cf-6803479067f8-catalog-content\") pod \"redhat-marketplace-qdsv6\" (UID: \"f5951a44-c13f-4cef-a2cf-6803479067f8\") " pod="openshift-marketplace/redhat-marketplace-qdsv6" Oct 05 20:59:05 crc kubenswrapper[4754]: I1005 20:59:05.600932 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ffkwv\" (UniqueName: \"kubernetes.io/projected/f5951a44-c13f-4cef-a2cf-6803479067f8-kube-api-access-ffkwv\") pod \"redhat-marketplace-qdsv6\" (UID: \"f5951a44-c13f-4cef-a2cf-6803479067f8\") " pod="openshift-marketplace/redhat-marketplace-qdsv6" Oct 05 20:59:05 crc kubenswrapper[4754]: I1005 20:59:05.601682 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f5951a44-c13f-4cef-a2cf-6803479067f8-utilities\") pod \"redhat-marketplace-qdsv6\" (UID: \"f5951a44-c13f-4cef-a2cf-6803479067f8\") " pod="openshift-marketplace/redhat-marketplace-qdsv6" Oct 05 20:59:05 crc kubenswrapper[4754]: I1005 20:59:05.601907 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f5951a44-c13f-4cef-a2cf-6803479067f8-catalog-content\") pod \"redhat-marketplace-qdsv6\" (UID: \"f5951a44-c13f-4cef-a2cf-6803479067f8\") " pod="openshift-marketplace/redhat-marketplace-qdsv6" Oct 05 20:59:05 crc kubenswrapper[4754]: I1005 20:59:05.622222 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ffkwv\" (UniqueName: \"kubernetes.io/projected/f5951a44-c13f-4cef-a2cf-6803479067f8-kube-api-access-ffkwv\") pod \"redhat-marketplace-qdsv6\" (UID: \"f5951a44-c13f-4cef-a2cf-6803479067f8\") " pod="openshift-marketplace/redhat-marketplace-qdsv6" Oct 05 20:59:05 crc kubenswrapper[4754]: I1005 20:59:05.651940 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-m7sf7"] Oct 05 20:59:05 crc kubenswrapper[4754]: I1005 20:59:05.652945 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-m7sf7" Oct 05 20:59:05 crc kubenswrapper[4754]: I1005 20:59:05.654981 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Oct 05 20:59:05 crc kubenswrapper[4754]: I1005 20:59:05.665959 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-m7sf7"] Oct 05 20:59:05 crc kubenswrapper[4754]: I1005 20:59:05.702552 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5a34915c-4494-448a-b835-6f1276b105fa-utilities\") pod \"certified-operators-m7sf7\" (UID: \"5a34915c-4494-448a-b835-6f1276b105fa\") " pod="openshift-marketplace/certified-operators-m7sf7" Oct 05 20:59:05 crc kubenswrapper[4754]: I1005 20:59:05.702608 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5a34915c-4494-448a-b835-6f1276b105fa-catalog-content\") pod \"certified-operators-m7sf7\" (UID: \"5a34915c-4494-448a-b835-6f1276b105fa\") " pod="openshift-marketplace/certified-operators-m7sf7" Oct 05 20:59:05 crc kubenswrapper[4754]: I1005 20:59:05.702634 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ttnr7\" (UniqueName: \"kubernetes.io/projected/5a34915c-4494-448a-b835-6f1276b105fa-kube-api-access-ttnr7\") pod \"certified-operators-m7sf7\" (UID: \"5a34915c-4494-448a-b835-6f1276b105fa\") " pod="openshift-marketplace/certified-operators-m7sf7" Oct 05 20:59:05 crc kubenswrapper[4754]: I1005 20:59:05.797200 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qdsv6" Oct 05 20:59:05 crc kubenswrapper[4754]: I1005 20:59:05.803713 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5a34915c-4494-448a-b835-6f1276b105fa-utilities\") pod \"certified-operators-m7sf7\" (UID: \"5a34915c-4494-448a-b835-6f1276b105fa\") " pod="openshift-marketplace/certified-operators-m7sf7" Oct 05 20:59:05 crc kubenswrapper[4754]: I1005 20:59:05.803770 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5a34915c-4494-448a-b835-6f1276b105fa-catalog-content\") pod \"certified-operators-m7sf7\" (UID: \"5a34915c-4494-448a-b835-6f1276b105fa\") " pod="openshift-marketplace/certified-operators-m7sf7" Oct 05 20:59:05 crc kubenswrapper[4754]: I1005 20:59:05.803798 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ttnr7\" (UniqueName: \"kubernetes.io/projected/5a34915c-4494-448a-b835-6f1276b105fa-kube-api-access-ttnr7\") pod \"certified-operators-m7sf7\" (UID: \"5a34915c-4494-448a-b835-6f1276b105fa\") " pod="openshift-marketplace/certified-operators-m7sf7" Oct 05 20:59:05 crc kubenswrapper[4754]: I1005 20:59:05.804905 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5a34915c-4494-448a-b835-6f1276b105fa-catalog-content\") pod \"certified-operators-m7sf7\" (UID: \"5a34915c-4494-448a-b835-6f1276b105fa\") " pod="openshift-marketplace/certified-operators-m7sf7" Oct 05 20:59:05 crc kubenswrapper[4754]: I1005 20:59:05.805077 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5a34915c-4494-448a-b835-6f1276b105fa-utilities\") pod \"certified-operators-m7sf7\" (UID: \"5a34915c-4494-448a-b835-6f1276b105fa\") " pod="openshift-marketplace/certified-operators-m7sf7" Oct 05 20:59:05 crc kubenswrapper[4754]: I1005 20:59:05.824027 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ttnr7\" (UniqueName: \"kubernetes.io/projected/5a34915c-4494-448a-b835-6f1276b105fa-kube-api-access-ttnr7\") pod \"certified-operators-m7sf7\" (UID: \"5a34915c-4494-448a-b835-6f1276b105fa\") " pod="openshift-marketplace/certified-operators-m7sf7" Oct 05 20:59:05 crc kubenswrapper[4754]: I1005 20:59:05.982016 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-m7sf7" Oct 05 20:59:06 crc kubenswrapper[4754]: I1005 20:59:06.214946 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qdsv6"] Oct 05 20:59:06 crc kubenswrapper[4754]: W1005 20:59:06.223204 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf5951a44_c13f_4cef_a2cf_6803479067f8.slice/crio-fa45b73fe124a45cfaecafe3ea6ecaeeb9827b2d248259dcb70b5ee8a27383b5 WatchSource:0}: Error finding container fa45b73fe124a45cfaecafe3ea6ecaeeb9827b2d248259dcb70b5ee8a27383b5: Status 404 returned error can't find the container with id fa45b73fe124a45cfaecafe3ea6ecaeeb9827b2d248259dcb70b5ee8a27383b5 Oct 05 20:59:06 crc kubenswrapper[4754]: I1005 20:59:06.370754 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-m7sf7"] Oct 05 20:59:06 crc kubenswrapper[4754]: W1005 20:59:06.377353 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5a34915c_4494_448a_b835_6f1276b105fa.slice/crio-b9da28c2b78f882c80e44e80fa1a6bd280dd261953365eeabeae317463ee995c WatchSource:0}: Error finding container b9da28c2b78f882c80e44e80fa1a6bd280dd261953365eeabeae317463ee995c: Status 404 returned error can't find the container with id b9da28c2b78f882c80e44e80fa1a6bd280dd261953365eeabeae317463ee995c Oct 05 20:59:06 crc kubenswrapper[4754]: I1005 20:59:06.844197 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="61cebf9f-dd0d-40f9-9bd5-1ff0dd7038f8" path="/var/lib/kubelet/pods/61cebf9f-dd0d-40f9-9bd5-1ff0dd7038f8/volumes" Oct 05 20:59:06 crc kubenswrapper[4754]: I1005 20:59:06.845131 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="85d11cf6-caa4-4ee2-a8ee-48c44469d4b7" path="/var/lib/kubelet/pods/85d11cf6-caa4-4ee2-a8ee-48c44469d4b7/volumes" Oct 05 20:59:06 crc kubenswrapper[4754]: I1005 20:59:06.845715 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b5b25be1-b393-4d98-aaf0-3f9693ba89d7" path="/var/lib/kubelet/pods/b5b25be1-b393-4d98-aaf0-3f9693ba89d7/volumes" Oct 05 20:59:06 crc kubenswrapper[4754]: I1005 20:59:06.846772 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="caf99098-6640-4077-8b97-69d1bfefa99c" path="/var/lib/kubelet/pods/caf99098-6640-4077-8b97-69d1bfefa99c/volumes" Oct 05 20:59:06 crc kubenswrapper[4754]: I1005 20:59:06.995890 4754 generic.go:334] "Generic (PLEG): container finished" podID="5a34915c-4494-448a-b835-6f1276b105fa" containerID="72a837326d4262c9c4d574132acbae4bb4281748dcc2085fbdede619a801d89c" exitCode=0 Oct 05 20:59:06 crc kubenswrapper[4754]: I1005 20:59:06.995964 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m7sf7" event={"ID":"5a34915c-4494-448a-b835-6f1276b105fa","Type":"ContainerDied","Data":"72a837326d4262c9c4d574132acbae4bb4281748dcc2085fbdede619a801d89c"} Oct 05 20:59:06 crc kubenswrapper[4754]: I1005 20:59:06.995996 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m7sf7" event={"ID":"5a34915c-4494-448a-b835-6f1276b105fa","Type":"ContainerStarted","Data":"b9da28c2b78f882c80e44e80fa1a6bd280dd261953365eeabeae317463ee995c"} Oct 05 20:59:06 crc kubenswrapper[4754]: I1005 20:59:06.998828 4754 generic.go:334] "Generic (PLEG): container finished" podID="f5951a44-c13f-4cef-a2cf-6803479067f8" containerID="b287c2874fa4ca18d3f859736e26dd773cc3f24628a3eae23626d30be40661df" exitCode=0 Oct 05 20:59:06 crc kubenswrapper[4754]: I1005 20:59:06.999610 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qdsv6" event={"ID":"f5951a44-c13f-4cef-a2cf-6803479067f8","Type":"ContainerDied","Data":"b287c2874fa4ca18d3f859736e26dd773cc3f24628a3eae23626d30be40661df"} Oct 05 20:59:06 crc kubenswrapper[4754]: I1005 20:59:06.999657 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qdsv6" event={"ID":"f5951a44-c13f-4cef-a2cf-6803479067f8","Type":"ContainerStarted","Data":"fa45b73fe124a45cfaecafe3ea6ecaeeb9827b2d248259dcb70b5ee8a27383b5"} Oct 05 20:59:07 crc kubenswrapper[4754]: I1005 20:59:07.867056 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-2lvpm"] Oct 05 20:59:07 crc kubenswrapper[4754]: I1005 20:59:07.868332 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2lvpm" Oct 05 20:59:07 crc kubenswrapper[4754]: I1005 20:59:07.870749 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Oct 05 20:59:07 crc kubenswrapper[4754]: I1005 20:59:07.886812 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-2lvpm"] Oct 05 20:59:07 crc kubenswrapper[4754]: I1005 20:59:07.945300 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b27b18ff-2e0a-446d-b605-5b74f825260b-catalog-content\") pod \"redhat-operators-2lvpm\" (UID: \"b27b18ff-2e0a-446d-b605-5b74f825260b\") " pod="openshift-marketplace/redhat-operators-2lvpm" Oct 05 20:59:07 crc kubenswrapper[4754]: I1005 20:59:07.945342 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b27b18ff-2e0a-446d-b605-5b74f825260b-utilities\") pod \"redhat-operators-2lvpm\" (UID: \"b27b18ff-2e0a-446d-b605-5b74f825260b\") " pod="openshift-marketplace/redhat-operators-2lvpm" Oct 05 20:59:07 crc kubenswrapper[4754]: I1005 20:59:07.945387 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h5kx8\" (UniqueName: \"kubernetes.io/projected/b27b18ff-2e0a-446d-b605-5b74f825260b-kube-api-access-h5kx8\") pod \"redhat-operators-2lvpm\" (UID: \"b27b18ff-2e0a-446d-b605-5b74f825260b\") " pod="openshift-marketplace/redhat-operators-2lvpm" Oct 05 20:59:08 crc kubenswrapper[4754]: I1005 20:59:08.006154 4754 generic.go:334] "Generic (PLEG): container finished" podID="f5951a44-c13f-4cef-a2cf-6803479067f8" containerID="c272cd1ceee5dc237202cc6a6ccc16cfd5c45e04b994e6e9d81d26075fa59ef3" exitCode=0 Oct 05 20:59:08 crc kubenswrapper[4754]: I1005 20:59:08.006230 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qdsv6" event={"ID":"f5951a44-c13f-4cef-a2cf-6803479067f8","Type":"ContainerDied","Data":"c272cd1ceee5dc237202cc6a6ccc16cfd5c45e04b994e6e9d81d26075fa59ef3"} Oct 05 20:59:08 crc kubenswrapper[4754]: I1005 20:59:08.009714 4754 generic.go:334] "Generic (PLEG): container finished" podID="5a34915c-4494-448a-b835-6f1276b105fa" containerID="f82dee9e16a425d7cc54b62124e7efbd776c5d014a60ba5cc65df1e626ac92bc" exitCode=0 Oct 05 20:59:08 crc kubenswrapper[4754]: I1005 20:59:08.009756 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m7sf7" event={"ID":"5a34915c-4494-448a-b835-6f1276b105fa","Type":"ContainerDied","Data":"f82dee9e16a425d7cc54b62124e7efbd776c5d014a60ba5cc65df1e626ac92bc"} Oct 05 20:59:08 crc kubenswrapper[4754]: I1005 20:59:08.047815 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b27b18ff-2e0a-446d-b605-5b74f825260b-catalog-content\") pod \"redhat-operators-2lvpm\" (UID: \"b27b18ff-2e0a-446d-b605-5b74f825260b\") " pod="openshift-marketplace/redhat-operators-2lvpm" Oct 05 20:59:08 crc kubenswrapper[4754]: I1005 20:59:08.047875 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b27b18ff-2e0a-446d-b605-5b74f825260b-utilities\") pod \"redhat-operators-2lvpm\" (UID: \"b27b18ff-2e0a-446d-b605-5b74f825260b\") " pod="openshift-marketplace/redhat-operators-2lvpm" Oct 05 20:59:08 crc kubenswrapper[4754]: I1005 20:59:08.047906 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h5kx8\" (UniqueName: \"kubernetes.io/projected/b27b18ff-2e0a-446d-b605-5b74f825260b-kube-api-access-h5kx8\") pod \"redhat-operators-2lvpm\" (UID: \"b27b18ff-2e0a-446d-b605-5b74f825260b\") " pod="openshift-marketplace/redhat-operators-2lvpm" Oct 05 20:59:08 crc kubenswrapper[4754]: I1005 20:59:08.049718 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b27b18ff-2e0a-446d-b605-5b74f825260b-catalog-content\") pod \"redhat-operators-2lvpm\" (UID: \"b27b18ff-2e0a-446d-b605-5b74f825260b\") " pod="openshift-marketplace/redhat-operators-2lvpm" Oct 05 20:59:08 crc kubenswrapper[4754]: I1005 20:59:08.050013 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b27b18ff-2e0a-446d-b605-5b74f825260b-utilities\") pod \"redhat-operators-2lvpm\" (UID: \"b27b18ff-2e0a-446d-b605-5b74f825260b\") " pod="openshift-marketplace/redhat-operators-2lvpm" Oct 05 20:59:08 crc kubenswrapper[4754]: I1005 20:59:08.065078 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-pkf48"] Oct 05 20:59:08 crc kubenswrapper[4754]: I1005 20:59:08.066629 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pkf48" Oct 05 20:59:08 crc kubenswrapper[4754]: I1005 20:59:08.069382 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Oct 05 20:59:08 crc kubenswrapper[4754]: I1005 20:59:08.074047 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-pkf48"] Oct 05 20:59:08 crc kubenswrapper[4754]: I1005 20:59:08.101580 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h5kx8\" (UniqueName: \"kubernetes.io/projected/b27b18ff-2e0a-446d-b605-5b74f825260b-kube-api-access-h5kx8\") pod \"redhat-operators-2lvpm\" (UID: \"b27b18ff-2e0a-446d-b605-5b74f825260b\") " pod="openshift-marketplace/redhat-operators-2lvpm" Oct 05 20:59:08 crc kubenswrapper[4754]: I1005 20:59:08.149021 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/095d3b9f-145a-4bc8-adca-a1831739e193-utilities\") pod \"community-operators-pkf48\" (UID: \"095d3b9f-145a-4bc8-adca-a1831739e193\") " pod="openshift-marketplace/community-operators-pkf48" Oct 05 20:59:08 crc kubenswrapper[4754]: I1005 20:59:08.149158 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/095d3b9f-145a-4bc8-adca-a1831739e193-catalog-content\") pod \"community-operators-pkf48\" (UID: \"095d3b9f-145a-4bc8-adca-a1831739e193\") " pod="openshift-marketplace/community-operators-pkf48" Oct 05 20:59:08 crc kubenswrapper[4754]: I1005 20:59:08.149193 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mmq4k\" (UniqueName: \"kubernetes.io/projected/095d3b9f-145a-4bc8-adca-a1831739e193-kube-api-access-mmq4k\") pod \"community-operators-pkf48\" (UID: \"095d3b9f-145a-4bc8-adca-a1831739e193\") " pod="openshift-marketplace/community-operators-pkf48" Oct 05 20:59:08 crc kubenswrapper[4754]: I1005 20:59:08.183610 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2lvpm" Oct 05 20:59:08 crc kubenswrapper[4754]: I1005 20:59:08.251039 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/095d3b9f-145a-4bc8-adca-a1831739e193-utilities\") pod \"community-operators-pkf48\" (UID: \"095d3b9f-145a-4bc8-adca-a1831739e193\") " pod="openshift-marketplace/community-operators-pkf48" Oct 05 20:59:08 crc kubenswrapper[4754]: I1005 20:59:08.251127 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/095d3b9f-145a-4bc8-adca-a1831739e193-catalog-content\") pod \"community-operators-pkf48\" (UID: \"095d3b9f-145a-4bc8-adca-a1831739e193\") " pod="openshift-marketplace/community-operators-pkf48" Oct 05 20:59:08 crc kubenswrapper[4754]: I1005 20:59:08.251149 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mmq4k\" (UniqueName: \"kubernetes.io/projected/095d3b9f-145a-4bc8-adca-a1831739e193-kube-api-access-mmq4k\") pod \"community-operators-pkf48\" (UID: \"095d3b9f-145a-4bc8-adca-a1831739e193\") " pod="openshift-marketplace/community-operators-pkf48" Oct 05 20:59:08 crc kubenswrapper[4754]: I1005 20:59:08.251932 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/095d3b9f-145a-4bc8-adca-a1831739e193-utilities\") pod \"community-operators-pkf48\" (UID: \"095d3b9f-145a-4bc8-adca-a1831739e193\") " pod="openshift-marketplace/community-operators-pkf48" Oct 05 20:59:08 crc kubenswrapper[4754]: I1005 20:59:08.252191 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/095d3b9f-145a-4bc8-adca-a1831739e193-catalog-content\") pod \"community-operators-pkf48\" (UID: \"095d3b9f-145a-4bc8-adca-a1831739e193\") " pod="openshift-marketplace/community-operators-pkf48" Oct 05 20:59:08 crc kubenswrapper[4754]: I1005 20:59:08.289121 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mmq4k\" (UniqueName: \"kubernetes.io/projected/095d3b9f-145a-4bc8-adca-a1831739e193-kube-api-access-mmq4k\") pod \"community-operators-pkf48\" (UID: \"095d3b9f-145a-4bc8-adca-a1831739e193\") " pod="openshift-marketplace/community-operators-pkf48" Oct 05 20:59:08 crc kubenswrapper[4754]: I1005 20:59:08.385666 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pkf48" Oct 05 20:59:08 crc kubenswrapper[4754]: I1005 20:59:08.597857 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-2lvpm"] Oct 05 20:59:08 crc kubenswrapper[4754]: W1005 20:59:08.606526 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb27b18ff_2e0a_446d_b605_5b74f825260b.slice/crio-d4a83834289914e30a5d5cb31285d3c909b8b8e0de0b08f1e157964dbab758e9 WatchSource:0}: Error finding container d4a83834289914e30a5d5cb31285d3c909b8b8e0de0b08f1e157964dbab758e9: Status 404 returned error can't find the container with id d4a83834289914e30a5d5cb31285d3c909b8b8e0de0b08f1e157964dbab758e9 Oct 05 20:59:08 crc kubenswrapper[4754]: I1005 20:59:08.787016 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-pkf48"] Oct 05 20:59:08 crc kubenswrapper[4754]: W1005 20:59:08.798634 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod095d3b9f_145a_4bc8_adca_a1831739e193.slice/crio-da4c50f613324e13d96b46e5282d74463f3128dcc7391c2b4e935b1329154111 WatchSource:0}: Error finding container da4c50f613324e13d96b46e5282d74463f3128dcc7391c2b4e935b1329154111: Status 404 returned error can't find the container with id da4c50f613324e13d96b46e5282d74463f3128dcc7391c2b4e935b1329154111 Oct 05 20:59:09 crc kubenswrapper[4754]: I1005 20:59:09.016036 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2lvpm" event={"ID":"b27b18ff-2e0a-446d-b605-5b74f825260b","Type":"ContainerStarted","Data":"d4a83834289914e30a5d5cb31285d3c909b8b8e0de0b08f1e157964dbab758e9"} Oct 05 20:59:09 crc kubenswrapper[4754]: I1005 20:59:09.020404 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pkf48" event={"ID":"095d3b9f-145a-4bc8-adca-a1831739e193","Type":"ContainerStarted","Data":"da4c50f613324e13d96b46e5282d74463f3128dcc7391c2b4e935b1329154111"} Oct 05 20:59:09 crc kubenswrapper[4754]: I1005 20:59:09.022792 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qdsv6" event={"ID":"f5951a44-c13f-4cef-a2cf-6803479067f8","Type":"ContainerStarted","Data":"174e27d81dd4104a618b498f0df7cf81165310b34dc7e41e464ee462ab39fc13"} Oct 05 20:59:09 crc kubenswrapper[4754]: I1005 20:59:09.043204 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-qdsv6" podStartSLOduration=2.509087284 podStartE2EDuration="4.043180203s" podCreationTimestamp="2025-10-05 20:59:05 +0000 UTC" firstStartedPulling="2025-10-05 20:59:07.003390809 +0000 UTC m=+270.907509519" lastFinishedPulling="2025-10-05 20:59:08.537483728 +0000 UTC m=+272.441602438" observedRunningTime="2025-10-05 20:59:09.038437141 +0000 UTC m=+272.942555851" watchObservedRunningTime="2025-10-05 20:59:09.043180203 +0000 UTC m=+272.947298913" Oct 05 20:59:10 crc kubenswrapper[4754]: I1005 20:59:10.032760 4754 generic.go:334] "Generic (PLEG): container finished" podID="095d3b9f-145a-4bc8-adca-a1831739e193" containerID="a5c181f2c69507c44219465ebed1d0f01b07084beb0f9fddae9f071c1f90b0c2" exitCode=0 Oct 05 20:59:10 crc kubenswrapper[4754]: I1005 20:59:10.032876 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pkf48" event={"ID":"095d3b9f-145a-4bc8-adca-a1831739e193","Type":"ContainerDied","Data":"a5c181f2c69507c44219465ebed1d0f01b07084beb0f9fddae9f071c1f90b0c2"} Oct 05 20:59:10 crc kubenswrapper[4754]: I1005 20:59:10.036140 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m7sf7" event={"ID":"5a34915c-4494-448a-b835-6f1276b105fa","Type":"ContainerStarted","Data":"f9419d1254af29a6974085d95119de6429b4849fdf922d16e581ce7f5ce86a9c"} Oct 05 20:59:10 crc kubenswrapper[4754]: I1005 20:59:10.038032 4754 generic.go:334] "Generic (PLEG): container finished" podID="b27b18ff-2e0a-446d-b605-5b74f825260b" containerID="635785839ed75ef54f55d33fa47e40b0e65c751e28e32551b9dbf190c83a8ab5" exitCode=0 Oct 05 20:59:10 crc kubenswrapper[4754]: I1005 20:59:10.038291 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2lvpm" event={"ID":"b27b18ff-2e0a-446d-b605-5b74f825260b","Type":"ContainerDied","Data":"635785839ed75ef54f55d33fa47e40b0e65c751e28e32551b9dbf190c83a8ab5"} Oct 05 20:59:10 crc kubenswrapper[4754]: I1005 20:59:10.092399 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-m7sf7" podStartSLOduration=3.27314094 podStartE2EDuration="5.0923779s" podCreationTimestamp="2025-10-05 20:59:05 +0000 UTC" firstStartedPulling="2025-10-05 20:59:06.998271657 +0000 UTC m=+270.902390357" lastFinishedPulling="2025-10-05 20:59:08.817508607 +0000 UTC m=+272.721627317" observedRunningTime="2025-10-05 20:59:10.087246688 +0000 UTC m=+273.991365418" watchObservedRunningTime="2025-10-05 20:59:10.0923779 +0000 UTC m=+273.996496610" Oct 05 20:59:11 crc kubenswrapper[4754]: I1005 20:59:11.047129 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2lvpm" event={"ID":"b27b18ff-2e0a-446d-b605-5b74f825260b","Type":"ContainerStarted","Data":"ede1f7866e2c876c3df39f1ac92b0519ea1d39a2c6a3b1024be68d5a3213ee98"} Oct 05 20:59:11 crc kubenswrapper[4754]: I1005 20:59:11.052364 4754 generic.go:334] "Generic (PLEG): container finished" podID="095d3b9f-145a-4bc8-adca-a1831739e193" containerID="650645686de45db43c29e3a6ff385dfa16c45f77a7b873d8e25925f188197695" exitCode=0 Oct 05 20:59:11 crc kubenswrapper[4754]: I1005 20:59:11.053179 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pkf48" event={"ID":"095d3b9f-145a-4bc8-adca-a1831739e193","Type":"ContainerDied","Data":"650645686de45db43c29e3a6ff385dfa16c45f77a7b873d8e25925f188197695"} Oct 05 20:59:12 crc kubenswrapper[4754]: I1005 20:59:12.060313 4754 generic.go:334] "Generic (PLEG): container finished" podID="b27b18ff-2e0a-446d-b605-5b74f825260b" containerID="ede1f7866e2c876c3df39f1ac92b0519ea1d39a2c6a3b1024be68d5a3213ee98" exitCode=0 Oct 05 20:59:12 crc kubenswrapper[4754]: I1005 20:59:12.060531 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2lvpm" event={"ID":"b27b18ff-2e0a-446d-b605-5b74f825260b","Type":"ContainerDied","Data":"ede1f7866e2c876c3df39f1ac92b0519ea1d39a2c6a3b1024be68d5a3213ee98"} Oct 05 20:59:13 crc kubenswrapper[4754]: I1005 20:59:13.070385 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2lvpm" event={"ID":"b27b18ff-2e0a-446d-b605-5b74f825260b","Type":"ContainerStarted","Data":"4f86fd0577823e7995bf4da02a9b04f1dd23dfd5522c74dc00770e36b4444408"} Oct 05 20:59:13 crc kubenswrapper[4754]: I1005 20:59:13.073220 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pkf48" event={"ID":"095d3b9f-145a-4bc8-adca-a1831739e193","Type":"ContainerStarted","Data":"e88d6a5afd89fd1a292790e2c1523ba5b176e7542143a72ccd135580f3b9ab8d"} Oct 05 20:59:13 crc kubenswrapper[4754]: I1005 20:59:13.088867 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-2lvpm" podStartSLOduration=3.661060139 podStartE2EDuration="6.08884011s" podCreationTimestamp="2025-10-05 20:59:07 +0000 UTC" firstStartedPulling="2025-10-05 20:59:10.040419678 +0000 UTC m=+273.944538388" lastFinishedPulling="2025-10-05 20:59:12.468199649 +0000 UTC m=+276.372318359" observedRunningTime="2025-10-05 20:59:13.087728022 +0000 UTC m=+276.991846732" watchObservedRunningTime="2025-10-05 20:59:13.08884011 +0000 UTC m=+276.992958820" Oct 05 20:59:13 crc kubenswrapper[4754]: I1005 20:59:13.115639 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-pkf48" podStartSLOduration=3.567854577 podStartE2EDuration="5.115614447s" podCreationTimestamp="2025-10-05 20:59:08 +0000 UTC" firstStartedPulling="2025-10-05 20:59:10.037616566 +0000 UTC m=+273.941735276" lastFinishedPulling="2025-10-05 20:59:11.585376436 +0000 UTC m=+275.489495146" observedRunningTime="2025-10-05 20:59:13.113441541 +0000 UTC m=+277.017560261" watchObservedRunningTime="2025-10-05 20:59:13.115614447 +0000 UTC m=+277.019733157" Oct 05 20:59:15 crc kubenswrapper[4754]: I1005 20:59:15.798458 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-qdsv6" Oct 05 20:59:15 crc kubenswrapper[4754]: I1005 20:59:15.799063 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-qdsv6" Oct 05 20:59:15 crc kubenswrapper[4754]: I1005 20:59:15.858943 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-qdsv6" Oct 05 20:59:15 crc kubenswrapper[4754]: I1005 20:59:15.983964 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-m7sf7" Oct 05 20:59:15 crc kubenswrapper[4754]: I1005 20:59:15.984064 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-m7sf7" Oct 05 20:59:16 crc kubenswrapper[4754]: I1005 20:59:16.034522 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-m7sf7" Oct 05 20:59:16 crc kubenswrapper[4754]: I1005 20:59:16.146505 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-qdsv6" Oct 05 20:59:16 crc kubenswrapper[4754]: I1005 20:59:16.152898 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-m7sf7" Oct 05 20:59:18 crc kubenswrapper[4754]: I1005 20:59:18.184850 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-2lvpm" Oct 05 20:59:18 crc kubenswrapper[4754]: I1005 20:59:18.185410 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-2lvpm" Oct 05 20:59:18 crc kubenswrapper[4754]: I1005 20:59:18.239903 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-2lvpm" Oct 05 20:59:18 crc kubenswrapper[4754]: I1005 20:59:18.386371 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-pkf48" Oct 05 20:59:18 crc kubenswrapper[4754]: I1005 20:59:18.386811 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-pkf48" Oct 05 20:59:18 crc kubenswrapper[4754]: I1005 20:59:18.433597 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-pkf48" Oct 05 20:59:19 crc kubenswrapper[4754]: I1005 20:59:19.155055 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-2lvpm" Oct 05 20:59:19 crc kubenswrapper[4754]: I1005 20:59:19.159393 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-pkf48" Oct 05 21:00:00 crc kubenswrapper[4754]: I1005 21:00:00.162822 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29328300-gkv6p"] Oct 05 21:00:00 crc kubenswrapper[4754]: I1005 21:00:00.165177 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29328300-gkv6p" Oct 05 21:00:00 crc kubenswrapper[4754]: I1005 21:00:00.171451 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 05 21:00:00 crc kubenswrapper[4754]: I1005 21:00:00.171876 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 05 21:00:00 crc kubenswrapper[4754]: I1005 21:00:00.192907 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29328300-gkv6p"] Oct 05 21:00:00 crc kubenswrapper[4754]: I1005 21:00:00.352773 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/25e2c21c-e2af-4b3a-b990-76e7d4128b9f-secret-volume\") pod \"collect-profiles-29328300-gkv6p\" (UID: \"25e2c21c-e2af-4b3a-b990-76e7d4128b9f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29328300-gkv6p" Oct 05 21:00:00 crc kubenswrapper[4754]: I1005 21:00:00.352914 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fj4bg\" (UniqueName: \"kubernetes.io/projected/25e2c21c-e2af-4b3a-b990-76e7d4128b9f-kube-api-access-fj4bg\") pod \"collect-profiles-29328300-gkv6p\" (UID: \"25e2c21c-e2af-4b3a-b990-76e7d4128b9f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29328300-gkv6p" Oct 05 21:00:00 crc kubenswrapper[4754]: I1005 21:00:00.353024 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/25e2c21c-e2af-4b3a-b990-76e7d4128b9f-config-volume\") pod \"collect-profiles-29328300-gkv6p\" (UID: \"25e2c21c-e2af-4b3a-b990-76e7d4128b9f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29328300-gkv6p" Oct 05 21:00:00 crc kubenswrapper[4754]: I1005 21:00:00.454394 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/25e2c21c-e2af-4b3a-b990-76e7d4128b9f-config-volume\") pod \"collect-profiles-29328300-gkv6p\" (UID: \"25e2c21c-e2af-4b3a-b990-76e7d4128b9f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29328300-gkv6p" Oct 05 21:00:00 crc kubenswrapper[4754]: I1005 21:00:00.455040 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/25e2c21c-e2af-4b3a-b990-76e7d4128b9f-secret-volume\") pod \"collect-profiles-29328300-gkv6p\" (UID: \"25e2c21c-e2af-4b3a-b990-76e7d4128b9f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29328300-gkv6p" Oct 05 21:00:00 crc kubenswrapper[4754]: I1005 21:00:00.455385 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fj4bg\" (UniqueName: \"kubernetes.io/projected/25e2c21c-e2af-4b3a-b990-76e7d4128b9f-kube-api-access-fj4bg\") pod \"collect-profiles-29328300-gkv6p\" (UID: \"25e2c21c-e2af-4b3a-b990-76e7d4128b9f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29328300-gkv6p" Oct 05 21:00:00 crc kubenswrapper[4754]: I1005 21:00:00.457382 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/25e2c21c-e2af-4b3a-b990-76e7d4128b9f-config-volume\") pod \"collect-profiles-29328300-gkv6p\" (UID: \"25e2c21c-e2af-4b3a-b990-76e7d4128b9f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29328300-gkv6p" Oct 05 21:00:00 crc kubenswrapper[4754]: I1005 21:00:00.472311 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/25e2c21c-e2af-4b3a-b990-76e7d4128b9f-secret-volume\") pod \"collect-profiles-29328300-gkv6p\" (UID: \"25e2c21c-e2af-4b3a-b990-76e7d4128b9f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29328300-gkv6p" Oct 05 21:00:00 crc kubenswrapper[4754]: I1005 21:00:00.486284 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fj4bg\" (UniqueName: \"kubernetes.io/projected/25e2c21c-e2af-4b3a-b990-76e7d4128b9f-kube-api-access-fj4bg\") pod \"collect-profiles-29328300-gkv6p\" (UID: \"25e2c21c-e2af-4b3a-b990-76e7d4128b9f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29328300-gkv6p" Oct 05 21:00:00 crc kubenswrapper[4754]: I1005 21:00:00.492621 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29328300-gkv6p" Oct 05 21:00:00 crc kubenswrapper[4754]: I1005 21:00:00.744413 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29328300-gkv6p"] Oct 05 21:00:01 crc kubenswrapper[4754]: I1005 21:00:01.412990 4754 generic.go:334] "Generic (PLEG): container finished" podID="25e2c21c-e2af-4b3a-b990-76e7d4128b9f" containerID="27e07597c80288def0893b4b8459ca7506f59fa2521265c1270b55514ae6c14e" exitCode=0 Oct 05 21:00:01 crc kubenswrapper[4754]: I1005 21:00:01.413075 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29328300-gkv6p" event={"ID":"25e2c21c-e2af-4b3a-b990-76e7d4128b9f","Type":"ContainerDied","Data":"27e07597c80288def0893b4b8459ca7506f59fa2521265c1270b55514ae6c14e"} Oct 05 21:00:01 crc kubenswrapper[4754]: I1005 21:00:01.413695 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29328300-gkv6p" event={"ID":"25e2c21c-e2af-4b3a-b990-76e7d4128b9f","Type":"ContainerStarted","Data":"ae7682690ea5945964401b2e4f13a572a9c5a247fe534b37173b99a8f11fbd40"} Oct 05 21:00:02 crc kubenswrapper[4754]: I1005 21:00:02.735069 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29328300-gkv6p" Oct 05 21:00:02 crc kubenswrapper[4754]: I1005 21:00:02.796228 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/25e2c21c-e2af-4b3a-b990-76e7d4128b9f-config-volume\") pod \"25e2c21c-e2af-4b3a-b990-76e7d4128b9f\" (UID: \"25e2c21c-e2af-4b3a-b990-76e7d4128b9f\") " Oct 05 21:00:02 crc kubenswrapper[4754]: I1005 21:00:02.796323 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/25e2c21c-e2af-4b3a-b990-76e7d4128b9f-secret-volume\") pod \"25e2c21c-e2af-4b3a-b990-76e7d4128b9f\" (UID: \"25e2c21c-e2af-4b3a-b990-76e7d4128b9f\") " Oct 05 21:00:02 crc kubenswrapper[4754]: I1005 21:00:02.796474 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fj4bg\" (UniqueName: \"kubernetes.io/projected/25e2c21c-e2af-4b3a-b990-76e7d4128b9f-kube-api-access-fj4bg\") pod \"25e2c21c-e2af-4b3a-b990-76e7d4128b9f\" (UID: \"25e2c21c-e2af-4b3a-b990-76e7d4128b9f\") " Oct 05 21:00:02 crc kubenswrapper[4754]: I1005 21:00:02.797235 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e2c21c-e2af-4b3a-b990-76e7d4128b9f-config-volume" (OuterVolumeSpecName: "config-volume") pod "25e2c21c-e2af-4b3a-b990-76e7d4128b9f" (UID: "25e2c21c-e2af-4b3a-b990-76e7d4128b9f"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 21:00:02 crc kubenswrapper[4754]: I1005 21:00:02.805766 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e2c21c-e2af-4b3a-b990-76e7d4128b9f-kube-api-access-fj4bg" (OuterVolumeSpecName: "kube-api-access-fj4bg") pod "25e2c21c-e2af-4b3a-b990-76e7d4128b9f" (UID: "25e2c21c-e2af-4b3a-b990-76e7d4128b9f"). InnerVolumeSpecName "kube-api-access-fj4bg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:00:02 crc kubenswrapper[4754]: I1005 21:00:02.807738 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e2c21c-e2af-4b3a-b990-76e7d4128b9f-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "25e2c21c-e2af-4b3a-b990-76e7d4128b9f" (UID: "25e2c21c-e2af-4b3a-b990-76e7d4128b9f"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:00:02 crc kubenswrapper[4754]: I1005 21:00:02.897844 4754 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/25e2c21c-e2af-4b3a-b990-76e7d4128b9f-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 05 21:00:02 crc kubenswrapper[4754]: I1005 21:00:02.898156 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fj4bg\" (UniqueName: \"kubernetes.io/projected/25e2c21c-e2af-4b3a-b990-76e7d4128b9f-kube-api-access-fj4bg\") on node \"crc\" DevicePath \"\"" Oct 05 21:00:02 crc kubenswrapper[4754]: I1005 21:00:02.898233 4754 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/25e2c21c-e2af-4b3a-b990-76e7d4128b9f-config-volume\") on node \"crc\" DevicePath \"\"" Oct 05 21:00:03 crc kubenswrapper[4754]: I1005 21:00:03.427957 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29328300-gkv6p" event={"ID":"25e2c21c-e2af-4b3a-b990-76e7d4128b9f","Type":"ContainerDied","Data":"ae7682690ea5945964401b2e4f13a572a9c5a247fe534b37173b99a8f11fbd40"} Oct 05 21:00:03 crc kubenswrapper[4754]: I1005 21:00:03.428016 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ae7682690ea5945964401b2e4f13a572a9c5a247fe534b37173b99a8f11fbd40" Oct 05 21:00:03 crc kubenswrapper[4754]: I1005 21:00:03.428607 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29328300-gkv6p" Oct 05 21:00:37 crc kubenswrapper[4754]: I1005 21:00:35.245681 4754 patch_prober.go:28] interesting pod/machine-config-daemon-b2h9k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 21:00:37 crc kubenswrapper[4754]: I1005 21:00:35.246729 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 21:01:05 crc kubenswrapper[4754]: I1005 21:01:05.245867 4754 patch_prober.go:28] interesting pod/machine-config-daemon-b2h9k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 21:01:05 crc kubenswrapper[4754]: I1005 21:01:05.246701 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 21:01:29 crc kubenswrapper[4754]: I1005 21:01:29.700636 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-wqk66"] Oct 05 21:01:29 crc kubenswrapper[4754]: E1005 21:01:29.701351 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25e2c21c-e2af-4b3a-b990-76e7d4128b9f" containerName="collect-profiles" Oct 05 21:01:29 crc kubenswrapper[4754]: I1005 21:01:29.701363 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="25e2c21c-e2af-4b3a-b990-76e7d4128b9f" containerName="collect-profiles" Oct 05 21:01:29 crc kubenswrapper[4754]: I1005 21:01:29.701457 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="25e2c21c-e2af-4b3a-b990-76e7d4128b9f" containerName="collect-profiles" Oct 05 21:01:29 crc kubenswrapper[4754]: I1005 21:01:29.701860 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-wqk66" Oct 05 21:01:29 crc kubenswrapper[4754]: I1005 21:01:29.723913 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-wqk66"] Oct 05 21:01:29 crc kubenswrapper[4754]: I1005 21:01:29.865517 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-wqk66\" (UID: \"8f22dd67-2a2e-442b-bd31-7a9c22478999\") " pod="openshift-image-registry/image-registry-66df7c8f76-wqk66" Oct 05 21:01:29 crc kubenswrapper[4754]: I1005 21:01:29.865561 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f22dd67-2a2e-442b-bd31-7a9c22478999-installation-pull-secrets\") pod \"image-registry-66df7c8f76-wqk66\" (UID: \"8f22dd67-2a2e-442b-bd31-7a9c22478999\") " pod="openshift-image-registry/image-registry-66df7c8f76-wqk66" Oct 05 21:01:29 crc kubenswrapper[4754]: I1005 21:01:29.865578 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f22dd67-2a2e-442b-bd31-7a9c22478999-trusted-ca\") pod \"image-registry-66df7c8f76-wqk66\" (UID: \"8f22dd67-2a2e-442b-bd31-7a9c22478999\") " pod="openshift-image-registry/image-registry-66df7c8f76-wqk66" Oct 05 21:01:29 crc kubenswrapper[4754]: I1005 21:01:29.865606 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g67tv\" (UniqueName: \"kubernetes.io/projected/8f22dd67-2a2e-442b-bd31-7a9c22478999-kube-api-access-g67tv\") pod \"image-registry-66df7c8f76-wqk66\" (UID: \"8f22dd67-2a2e-442b-bd31-7a9c22478999\") " pod="openshift-image-registry/image-registry-66df7c8f76-wqk66" Oct 05 21:01:29 crc kubenswrapper[4754]: I1005 21:01:29.865626 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f22dd67-2a2e-442b-bd31-7a9c22478999-ca-trust-extracted\") pod \"image-registry-66df7c8f76-wqk66\" (UID: \"8f22dd67-2a2e-442b-bd31-7a9c22478999\") " pod="openshift-image-registry/image-registry-66df7c8f76-wqk66" Oct 05 21:01:29 crc kubenswrapper[4754]: I1005 21:01:29.865771 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f22dd67-2a2e-442b-bd31-7a9c22478999-registry-certificates\") pod \"image-registry-66df7c8f76-wqk66\" (UID: \"8f22dd67-2a2e-442b-bd31-7a9c22478999\") " pod="openshift-image-registry/image-registry-66df7c8f76-wqk66" Oct 05 21:01:29 crc kubenswrapper[4754]: I1005 21:01:29.865906 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f22dd67-2a2e-442b-bd31-7a9c22478999-bound-sa-token\") pod \"image-registry-66df7c8f76-wqk66\" (UID: \"8f22dd67-2a2e-442b-bd31-7a9c22478999\") " pod="openshift-image-registry/image-registry-66df7c8f76-wqk66" Oct 05 21:01:29 crc kubenswrapper[4754]: I1005 21:01:29.866004 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f22dd67-2a2e-442b-bd31-7a9c22478999-registry-tls\") pod \"image-registry-66df7c8f76-wqk66\" (UID: \"8f22dd67-2a2e-442b-bd31-7a9c22478999\") " pod="openshift-image-registry/image-registry-66df7c8f76-wqk66" Oct 05 21:01:29 crc kubenswrapper[4754]: I1005 21:01:29.889313 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-wqk66\" (UID: \"8f22dd67-2a2e-442b-bd31-7a9c22478999\") " pod="openshift-image-registry/image-registry-66df7c8f76-wqk66" Oct 05 21:01:29 crc kubenswrapper[4754]: I1005 21:01:29.968390 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f22dd67-2a2e-442b-bd31-7a9c22478999-installation-pull-secrets\") pod \"image-registry-66df7c8f76-wqk66\" (UID: \"8f22dd67-2a2e-442b-bd31-7a9c22478999\") " pod="openshift-image-registry/image-registry-66df7c8f76-wqk66" Oct 05 21:01:29 crc kubenswrapper[4754]: I1005 21:01:29.968860 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f22dd67-2a2e-442b-bd31-7a9c22478999-trusted-ca\") pod \"image-registry-66df7c8f76-wqk66\" (UID: \"8f22dd67-2a2e-442b-bd31-7a9c22478999\") " pod="openshift-image-registry/image-registry-66df7c8f76-wqk66" Oct 05 21:01:29 crc kubenswrapper[4754]: I1005 21:01:29.968989 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g67tv\" (UniqueName: \"kubernetes.io/projected/8f22dd67-2a2e-442b-bd31-7a9c22478999-kube-api-access-g67tv\") pod \"image-registry-66df7c8f76-wqk66\" (UID: \"8f22dd67-2a2e-442b-bd31-7a9c22478999\") " pod="openshift-image-registry/image-registry-66df7c8f76-wqk66" Oct 05 21:01:29 crc kubenswrapper[4754]: I1005 21:01:29.969030 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f22dd67-2a2e-442b-bd31-7a9c22478999-ca-trust-extracted\") pod \"image-registry-66df7c8f76-wqk66\" (UID: \"8f22dd67-2a2e-442b-bd31-7a9c22478999\") " pod="openshift-image-registry/image-registry-66df7c8f76-wqk66" Oct 05 21:01:29 crc kubenswrapper[4754]: I1005 21:01:29.969104 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f22dd67-2a2e-442b-bd31-7a9c22478999-registry-certificates\") pod \"image-registry-66df7c8f76-wqk66\" (UID: \"8f22dd67-2a2e-442b-bd31-7a9c22478999\") " pod="openshift-image-registry/image-registry-66df7c8f76-wqk66" Oct 05 21:01:29 crc kubenswrapper[4754]: I1005 21:01:29.969141 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f22dd67-2a2e-442b-bd31-7a9c22478999-bound-sa-token\") pod \"image-registry-66df7c8f76-wqk66\" (UID: \"8f22dd67-2a2e-442b-bd31-7a9c22478999\") " pod="openshift-image-registry/image-registry-66df7c8f76-wqk66" Oct 05 21:01:29 crc kubenswrapper[4754]: I1005 21:01:29.969208 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f22dd67-2a2e-442b-bd31-7a9c22478999-registry-tls\") pod \"image-registry-66df7c8f76-wqk66\" (UID: \"8f22dd67-2a2e-442b-bd31-7a9c22478999\") " pod="openshift-image-registry/image-registry-66df7c8f76-wqk66" Oct 05 21:01:29 crc kubenswrapper[4754]: I1005 21:01:29.970357 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f22dd67-2a2e-442b-bd31-7a9c22478999-ca-trust-extracted\") pod \"image-registry-66df7c8f76-wqk66\" (UID: \"8f22dd67-2a2e-442b-bd31-7a9c22478999\") " pod="openshift-image-registry/image-registry-66df7c8f76-wqk66" Oct 05 21:01:29 crc kubenswrapper[4754]: I1005 21:01:29.971568 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f22dd67-2a2e-442b-bd31-7a9c22478999-trusted-ca\") pod \"image-registry-66df7c8f76-wqk66\" (UID: \"8f22dd67-2a2e-442b-bd31-7a9c22478999\") " pod="openshift-image-registry/image-registry-66df7c8f76-wqk66" Oct 05 21:01:29 crc kubenswrapper[4754]: I1005 21:01:29.972578 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f22dd67-2a2e-442b-bd31-7a9c22478999-registry-certificates\") pod \"image-registry-66df7c8f76-wqk66\" (UID: \"8f22dd67-2a2e-442b-bd31-7a9c22478999\") " pod="openshift-image-registry/image-registry-66df7c8f76-wqk66" Oct 05 21:01:29 crc kubenswrapper[4754]: I1005 21:01:29.975383 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f22dd67-2a2e-442b-bd31-7a9c22478999-installation-pull-secrets\") pod \"image-registry-66df7c8f76-wqk66\" (UID: \"8f22dd67-2a2e-442b-bd31-7a9c22478999\") " pod="openshift-image-registry/image-registry-66df7c8f76-wqk66" Oct 05 21:01:29 crc kubenswrapper[4754]: I1005 21:01:29.977575 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f22dd67-2a2e-442b-bd31-7a9c22478999-registry-tls\") pod \"image-registry-66df7c8f76-wqk66\" (UID: \"8f22dd67-2a2e-442b-bd31-7a9c22478999\") " pod="openshift-image-registry/image-registry-66df7c8f76-wqk66" Oct 05 21:01:29 crc kubenswrapper[4754]: I1005 21:01:29.999309 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f22dd67-2a2e-442b-bd31-7a9c22478999-bound-sa-token\") pod \"image-registry-66df7c8f76-wqk66\" (UID: \"8f22dd67-2a2e-442b-bd31-7a9c22478999\") " pod="openshift-image-registry/image-registry-66df7c8f76-wqk66" Oct 05 21:01:30 crc kubenswrapper[4754]: I1005 21:01:30.003578 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g67tv\" (UniqueName: \"kubernetes.io/projected/8f22dd67-2a2e-442b-bd31-7a9c22478999-kube-api-access-g67tv\") pod \"image-registry-66df7c8f76-wqk66\" (UID: \"8f22dd67-2a2e-442b-bd31-7a9c22478999\") " pod="openshift-image-registry/image-registry-66df7c8f76-wqk66" Oct 05 21:01:30 crc kubenswrapper[4754]: I1005 21:01:30.017333 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-wqk66" Oct 05 21:01:30 crc kubenswrapper[4754]: I1005 21:01:30.308148 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-wqk66"] Oct 05 21:01:31 crc kubenswrapper[4754]: I1005 21:01:31.090337 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-wqk66" event={"ID":"8f22dd67-2a2e-442b-bd31-7a9c22478999","Type":"ContainerStarted","Data":"67e04d9170889f066fbeadd64018fcc5e69eb3863586b3ea4c7d4c31b5f7e362"} Oct 05 21:01:31 crc kubenswrapper[4754]: I1005 21:01:31.091098 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-wqk66" event={"ID":"8f22dd67-2a2e-442b-bd31-7a9c22478999","Type":"ContainerStarted","Data":"d342a3aa6d276d81d3a0dbad1c4b0b51094dc5474036bcb20cc256a0a42368e7"} Oct 05 21:01:31 crc kubenswrapper[4754]: I1005 21:01:31.091136 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-wqk66" Oct 05 21:01:31 crc kubenswrapper[4754]: I1005 21:01:31.143982 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-wqk66" podStartSLOduration=2.143957455 podStartE2EDuration="2.143957455s" podCreationTimestamp="2025-10-05 21:01:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 21:01:31.112766828 +0000 UTC m=+415.016885608" watchObservedRunningTime="2025-10-05 21:01:31.143957455 +0000 UTC m=+415.048076165" Oct 05 21:01:35 crc kubenswrapper[4754]: I1005 21:01:35.244874 4754 patch_prober.go:28] interesting pod/machine-config-daemon-b2h9k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 21:01:35 crc kubenswrapper[4754]: I1005 21:01:35.245483 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 21:01:35 crc kubenswrapper[4754]: I1005 21:01:35.245603 4754 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" Oct 05 21:01:35 crc kubenswrapper[4754]: I1005 21:01:35.246754 4754 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1748c85d7a12794d4f4515c3318cf93b10e679b7d52cfbe7c2fbc83011ee324c"} pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 05 21:01:35 crc kubenswrapper[4754]: I1005 21:01:35.246829 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" containerName="machine-config-daemon" containerID="cri-o://1748c85d7a12794d4f4515c3318cf93b10e679b7d52cfbe7c2fbc83011ee324c" gracePeriod=600 Oct 05 21:01:36 crc kubenswrapper[4754]: I1005 21:01:36.133922 4754 generic.go:334] "Generic (PLEG): container finished" podID="be95a413-6a59-45b4-84b7-b43cae694a26" containerID="1748c85d7a12794d4f4515c3318cf93b10e679b7d52cfbe7c2fbc83011ee324c" exitCode=0 Oct 05 21:01:36 crc kubenswrapper[4754]: I1005 21:01:36.134040 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" event={"ID":"be95a413-6a59-45b4-84b7-b43cae694a26","Type":"ContainerDied","Data":"1748c85d7a12794d4f4515c3318cf93b10e679b7d52cfbe7c2fbc83011ee324c"} Oct 05 21:01:36 crc kubenswrapper[4754]: I1005 21:01:36.135061 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" event={"ID":"be95a413-6a59-45b4-84b7-b43cae694a26","Type":"ContainerStarted","Data":"0b09d5fa150c23a262d03541134e26faa0bf59c28b4299e50ad59b83bfa2351d"} Oct 05 21:01:36 crc kubenswrapper[4754]: I1005 21:01:36.135114 4754 scope.go:117] "RemoveContainer" containerID="4bb8c5903e8b70dd19eebb111097d96f725e17de44e06c3e74c067376f6c7360" Oct 05 21:01:50 crc kubenswrapper[4754]: I1005 21:01:50.026672 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-wqk66" Oct 05 21:01:50 crc kubenswrapper[4754]: I1005 21:01:50.114374 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-5sqj9"] Oct 05 21:02:15 crc kubenswrapper[4754]: I1005 21:02:15.171440 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-5sqj9" podUID="c64d57c5-4bd4-48b5-85fc-9d613aea29c4" containerName="registry" containerID="cri-o://591cdabaa80ef7bd9d0e08d68cf15ca0ca3b3a4940c23598ecfc3cecf551244d" gracePeriod=30 Oct 05 21:02:15 crc kubenswrapper[4754]: I1005 21:02:15.432024 4754 generic.go:334] "Generic (PLEG): container finished" podID="c64d57c5-4bd4-48b5-85fc-9d613aea29c4" containerID="591cdabaa80ef7bd9d0e08d68cf15ca0ca3b3a4940c23598ecfc3cecf551244d" exitCode=0 Oct 05 21:02:15 crc kubenswrapper[4754]: I1005 21:02:15.432153 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-5sqj9" event={"ID":"c64d57c5-4bd4-48b5-85fc-9d613aea29c4","Type":"ContainerDied","Data":"591cdabaa80ef7bd9d0e08d68cf15ca0ca3b3a4940c23598ecfc3cecf551244d"} Oct 05 21:02:15 crc kubenswrapper[4754]: I1005 21:02:15.595456 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-5sqj9" Oct 05 21:02:15 crc kubenswrapper[4754]: I1005 21:02:15.608035 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/c64d57c5-4bd4-48b5-85fc-9d613aea29c4-registry-certificates\") pod \"c64d57c5-4bd4-48b5-85fc-9d613aea29c4\" (UID: \"c64d57c5-4bd4-48b5-85fc-9d613aea29c4\") " Oct 05 21:02:15 crc kubenswrapper[4754]: I1005 21:02:15.608140 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c64d57c5-4bd4-48b5-85fc-9d613aea29c4-bound-sa-token\") pod \"c64d57c5-4bd4-48b5-85fc-9d613aea29c4\" (UID: \"c64d57c5-4bd4-48b5-85fc-9d613aea29c4\") " Oct 05 21:02:15 crc kubenswrapper[4754]: I1005 21:02:15.608190 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7qtr7\" (UniqueName: \"kubernetes.io/projected/c64d57c5-4bd4-48b5-85fc-9d613aea29c4-kube-api-access-7qtr7\") pod \"c64d57c5-4bd4-48b5-85fc-9d613aea29c4\" (UID: \"c64d57c5-4bd4-48b5-85fc-9d613aea29c4\") " Oct 05 21:02:15 crc kubenswrapper[4754]: I1005 21:02:15.608236 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/c64d57c5-4bd4-48b5-85fc-9d613aea29c4-registry-tls\") pod \"c64d57c5-4bd4-48b5-85fc-9d613aea29c4\" (UID: \"c64d57c5-4bd4-48b5-85fc-9d613aea29c4\") " Oct 05 21:02:15 crc kubenswrapper[4754]: I1005 21:02:15.608271 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/c64d57c5-4bd4-48b5-85fc-9d613aea29c4-ca-trust-extracted\") pod \"c64d57c5-4bd4-48b5-85fc-9d613aea29c4\" (UID: \"c64d57c5-4bd4-48b5-85fc-9d613aea29c4\") " Oct 05 21:02:15 crc kubenswrapper[4754]: I1005 21:02:15.608309 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c64d57c5-4bd4-48b5-85fc-9d613aea29c4-trusted-ca\") pod \"c64d57c5-4bd4-48b5-85fc-9d613aea29c4\" (UID: \"c64d57c5-4bd4-48b5-85fc-9d613aea29c4\") " Oct 05 21:02:15 crc kubenswrapper[4754]: I1005 21:02:15.608339 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/c64d57c5-4bd4-48b5-85fc-9d613aea29c4-installation-pull-secrets\") pod \"c64d57c5-4bd4-48b5-85fc-9d613aea29c4\" (UID: \"c64d57c5-4bd4-48b5-85fc-9d613aea29c4\") " Oct 05 21:02:15 crc kubenswrapper[4754]: I1005 21:02:15.608651 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"c64d57c5-4bd4-48b5-85fc-9d613aea29c4\" (UID: \"c64d57c5-4bd4-48b5-85fc-9d613aea29c4\") " Oct 05 21:02:15 crc kubenswrapper[4754]: I1005 21:02:15.618237 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c64d57c5-4bd4-48b5-85fc-9d613aea29c4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "c64d57c5-4bd4-48b5-85fc-9d613aea29c4" (UID: "c64d57c5-4bd4-48b5-85fc-9d613aea29c4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 21:02:15 crc kubenswrapper[4754]: I1005 21:02:15.618258 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c64d57c5-4bd4-48b5-85fc-9d613aea29c4-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "c64d57c5-4bd4-48b5-85fc-9d613aea29c4" (UID: "c64d57c5-4bd4-48b5-85fc-9d613aea29c4"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 21:02:15 crc kubenswrapper[4754]: I1005 21:02:15.630633 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c64d57c5-4bd4-48b5-85fc-9d613aea29c4-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "c64d57c5-4bd4-48b5-85fc-9d613aea29c4" (UID: "c64d57c5-4bd4-48b5-85fc-9d613aea29c4"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:02:15 crc kubenswrapper[4754]: I1005 21:02:15.631033 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c64d57c5-4bd4-48b5-85fc-9d613aea29c4-kube-api-access-7qtr7" (OuterVolumeSpecName: "kube-api-access-7qtr7") pod "c64d57c5-4bd4-48b5-85fc-9d613aea29c4" (UID: "c64d57c5-4bd4-48b5-85fc-9d613aea29c4"). InnerVolumeSpecName "kube-api-access-7qtr7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:02:15 crc kubenswrapper[4754]: I1005 21:02:15.637590 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c64d57c5-4bd4-48b5-85fc-9d613aea29c4-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "c64d57c5-4bd4-48b5-85fc-9d613aea29c4" (UID: "c64d57c5-4bd4-48b5-85fc-9d613aea29c4"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:02:15 crc kubenswrapper[4754]: I1005 21:02:15.638026 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c64d57c5-4bd4-48b5-85fc-9d613aea29c4-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "c64d57c5-4bd4-48b5-85fc-9d613aea29c4" (UID: "c64d57c5-4bd4-48b5-85fc-9d613aea29c4"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:02:15 crc kubenswrapper[4754]: I1005 21:02:15.641222 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "c64d57c5-4bd4-48b5-85fc-9d613aea29c4" (UID: "c64d57c5-4bd4-48b5-85fc-9d613aea29c4"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 05 21:02:15 crc kubenswrapper[4754]: I1005 21:02:15.648337 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c64d57c5-4bd4-48b5-85fc-9d613aea29c4-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "c64d57c5-4bd4-48b5-85fc-9d613aea29c4" (UID: "c64d57c5-4bd4-48b5-85fc-9d613aea29c4"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 21:02:15 crc kubenswrapper[4754]: I1005 21:02:15.710143 4754 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/c64d57c5-4bd4-48b5-85fc-9d613aea29c4-registry-certificates\") on node \"crc\" DevicePath \"\"" Oct 05 21:02:15 crc kubenswrapper[4754]: I1005 21:02:15.710407 4754 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c64d57c5-4bd4-48b5-85fc-9d613aea29c4-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 05 21:02:15 crc kubenswrapper[4754]: I1005 21:02:15.710483 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7qtr7\" (UniqueName: \"kubernetes.io/projected/c64d57c5-4bd4-48b5-85fc-9d613aea29c4-kube-api-access-7qtr7\") on node \"crc\" DevicePath \"\"" Oct 05 21:02:15 crc kubenswrapper[4754]: I1005 21:02:15.710576 4754 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/c64d57c5-4bd4-48b5-85fc-9d613aea29c4-registry-tls\") on node \"crc\" DevicePath \"\"" Oct 05 21:02:15 crc kubenswrapper[4754]: I1005 21:02:15.710663 4754 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/c64d57c5-4bd4-48b5-85fc-9d613aea29c4-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Oct 05 21:02:15 crc kubenswrapper[4754]: I1005 21:02:15.710731 4754 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c64d57c5-4bd4-48b5-85fc-9d613aea29c4-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 05 21:02:15 crc kubenswrapper[4754]: I1005 21:02:15.710799 4754 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/c64d57c5-4bd4-48b5-85fc-9d613aea29c4-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Oct 05 21:02:16 crc kubenswrapper[4754]: I1005 21:02:16.446239 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-5sqj9" event={"ID":"c64d57c5-4bd4-48b5-85fc-9d613aea29c4","Type":"ContainerDied","Data":"aa4820d4875b0b4449cb2c7a512e82f56a182188f58561773f130af26b6b0ab5"} Oct 05 21:02:16 crc kubenswrapper[4754]: I1005 21:02:16.446313 4754 scope.go:117] "RemoveContainer" containerID="591cdabaa80ef7bd9d0e08d68cf15ca0ca3b3a4940c23598ecfc3cecf551244d" Oct 05 21:02:16 crc kubenswrapper[4754]: I1005 21:02:16.446342 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-5sqj9" Oct 05 21:02:16 crc kubenswrapper[4754]: I1005 21:02:16.486184 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-5sqj9"] Oct 05 21:02:16 crc kubenswrapper[4754]: I1005 21:02:16.503568 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-5sqj9"] Oct 05 21:02:16 crc kubenswrapper[4754]: I1005 21:02:16.850851 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c64d57c5-4bd4-48b5-85fc-9d613aea29c4" path="/var/lib/kubelet/pods/c64d57c5-4bd4-48b5-85fc-9d613aea29c4/volumes" Oct 05 21:03:35 crc kubenswrapper[4754]: I1005 21:03:35.244761 4754 patch_prober.go:28] interesting pod/machine-config-daemon-b2h9k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 21:03:35 crc kubenswrapper[4754]: I1005 21:03:35.245575 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 21:04:05 crc kubenswrapper[4754]: I1005 21:04:05.252979 4754 patch_prober.go:28] interesting pod/machine-config-daemon-b2h9k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 21:04:05 crc kubenswrapper[4754]: I1005 21:04:05.253826 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 21:04:20 crc kubenswrapper[4754]: I1005 21:04:20.616486 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-pqw5s"] Oct 05 21:04:20 crc kubenswrapper[4754]: E1005 21:04:20.617296 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c64d57c5-4bd4-48b5-85fc-9d613aea29c4" containerName="registry" Oct 05 21:04:20 crc kubenswrapper[4754]: I1005 21:04:20.617309 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="c64d57c5-4bd4-48b5-85fc-9d613aea29c4" containerName="registry" Oct 05 21:04:20 crc kubenswrapper[4754]: I1005 21:04:20.617400 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="c64d57c5-4bd4-48b5-85fc-9d613aea29c4" containerName="registry" Oct 05 21:04:20 crc kubenswrapper[4754]: I1005 21:04:20.617828 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-pqw5s" Oct 05 21:04:20 crc kubenswrapper[4754]: I1005 21:04:20.621171 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Oct 05 21:04:20 crc kubenswrapper[4754]: I1005 21:04:20.621318 4754 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-g6l6c" Oct 05 21:04:20 crc kubenswrapper[4754]: I1005 21:04:20.621385 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Oct 05 21:04:20 crc kubenswrapper[4754]: I1005 21:04:20.629269 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-5b446d88c5-c5gxb"] Oct 05 21:04:20 crc kubenswrapper[4754]: I1005 21:04:20.629981 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-c5gxb" Oct 05 21:04:20 crc kubenswrapper[4754]: I1005 21:04:20.636590 4754 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-bbw2k" Oct 05 21:04:20 crc kubenswrapper[4754]: I1005 21:04:20.646573 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-r2txk"] Oct 05 21:04:20 crc kubenswrapper[4754]: I1005 21:04:20.647432 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-r2txk" Oct 05 21:04:20 crc kubenswrapper[4754]: I1005 21:04:20.652186 4754 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-7jq9n" Oct 05 21:04:20 crc kubenswrapper[4754]: I1005 21:04:20.668144 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-r2txk"] Oct 05 21:04:20 crc kubenswrapper[4754]: I1005 21:04:20.682303 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-pqw5s"] Oct 05 21:04:20 crc kubenswrapper[4754]: I1005 21:04:20.698629 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-c5gxb"] Oct 05 21:04:20 crc kubenswrapper[4754]: I1005 21:04:20.793354 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t7rr5\" (UniqueName: \"kubernetes.io/projected/f060400d-9b95-4e44-b94a-ba61d23e307c-kube-api-access-t7rr5\") pod \"cert-manager-cainjector-7f985d654d-pqw5s\" (UID: \"f060400d-9b95-4e44-b94a-ba61d23e307c\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-pqw5s" Oct 05 21:04:20 crc kubenswrapper[4754]: I1005 21:04:20.793514 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s8h2x\" (UniqueName: \"kubernetes.io/projected/e31db265-82f2-431c-af44-9a44bcb4762c-kube-api-access-s8h2x\") pod \"cert-manager-5b446d88c5-c5gxb\" (UID: \"e31db265-82f2-431c-af44-9a44bcb4762c\") " pod="cert-manager/cert-manager-5b446d88c5-c5gxb" Oct 05 21:04:20 crc kubenswrapper[4754]: I1005 21:04:20.793744 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qpc6q\" (UniqueName: \"kubernetes.io/projected/577f8ef5-2c2c-4b67-91b5-55ce19fc52d2-kube-api-access-qpc6q\") pod \"cert-manager-webhook-5655c58dd6-r2txk\" (UID: \"577f8ef5-2c2c-4b67-91b5-55ce19fc52d2\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-r2txk" Oct 05 21:04:20 crc kubenswrapper[4754]: I1005 21:04:20.894883 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s8h2x\" (UniqueName: \"kubernetes.io/projected/e31db265-82f2-431c-af44-9a44bcb4762c-kube-api-access-s8h2x\") pod \"cert-manager-5b446d88c5-c5gxb\" (UID: \"e31db265-82f2-431c-af44-9a44bcb4762c\") " pod="cert-manager/cert-manager-5b446d88c5-c5gxb" Oct 05 21:04:20 crc kubenswrapper[4754]: I1005 21:04:20.894958 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qpc6q\" (UniqueName: \"kubernetes.io/projected/577f8ef5-2c2c-4b67-91b5-55ce19fc52d2-kube-api-access-qpc6q\") pod \"cert-manager-webhook-5655c58dd6-r2txk\" (UID: \"577f8ef5-2c2c-4b67-91b5-55ce19fc52d2\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-r2txk" Oct 05 21:04:20 crc kubenswrapper[4754]: I1005 21:04:20.895006 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t7rr5\" (UniqueName: \"kubernetes.io/projected/f060400d-9b95-4e44-b94a-ba61d23e307c-kube-api-access-t7rr5\") pod \"cert-manager-cainjector-7f985d654d-pqw5s\" (UID: \"f060400d-9b95-4e44-b94a-ba61d23e307c\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-pqw5s" Oct 05 21:04:20 crc kubenswrapper[4754]: I1005 21:04:20.913831 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qpc6q\" (UniqueName: \"kubernetes.io/projected/577f8ef5-2c2c-4b67-91b5-55ce19fc52d2-kube-api-access-qpc6q\") pod \"cert-manager-webhook-5655c58dd6-r2txk\" (UID: \"577f8ef5-2c2c-4b67-91b5-55ce19fc52d2\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-r2txk" Oct 05 21:04:20 crc kubenswrapper[4754]: I1005 21:04:20.915141 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s8h2x\" (UniqueName: \"kubernetes.io/projected/e31db265-82f2-431c-af44-9a44bcb4762c-kube-api-access-s8h2x\") pod \"cert-manager-5b446d88c5-c5gxb\" (UID: \"e31db265-82f2-431c-af44-9a44bcb4762c\") " pod="cert-manager/cert-manager-5b446d88c5-c5gxb" Oct 05 21:04:20 crc kubenswrapper[4754]: I1005 21:04:20.917819 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t7rr5\" (UniqueName: \"kubernetes.io/projected/f060400d-9b95-4e44-b94a-ba61d23e307c-kube-api-access-t7rr5\") pod \"cert-manager-cainjector-7f985d654d-pqw5s\" (UID: \"f060400d-9b95-4e44-b94a-ba61d23e307c\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-pqw5s" Oct 05 21:04:20 crc kubenswrapper[4754]: I1005 21:04:20.933756 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-pqw5s" Oct 05 21:04:20 crc kubenswrapper[4754]: I1005 21:04:20.942042 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-c5gxb" Oct 05 21:04:20 crc kubenswrapper[4754]: I1005 21:04:20.960196 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-r2txk" Oct 05 21:04:21 crc kubenswrapper[4754]: I1005 21:04:21.170296 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-r2txk"] Oct 05 21:04:21 crc kubenswrapper[4754]: I1005 21:04:21.182631 4754 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 05 21:04:21 crc kubenswrapper[4754]: I1005 21:04:21.211351 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-pqw5s"] Oct 05 21:04:21 crc kubenswrapper[4754]: W1005 21:04:21.226203 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf060400d_9b95_4e44_b94a_ba61d23e307c.slice/crio-0150289145debb4434160c5518b18d98418f674646a3dc6f9a3d69ab0a766cea WatchSource:0}: Error finding container 0150289145debb4434160c5518b18d98418f674646a3dc6f9a3d69ab0a766cea: Status 404 returned error can't find the container with id 0150289145debb4434160c5518b18d98418f674646a3dc6f9a3d69ab0a766cea Oct 05 21:04:21 crc kubenswrapper[4754]: I1005 21:04:21.258861 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-c5gxb"] Oct 05 21:04:21 crc kubenswrapper[4754]: W1005 21:04:21.261265 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode31db265_82f2_431c_af44_9a44bcb4762c.slice/crio-4d3d79add6339a78623f2c0c581c68e4a2917eac7bcfc8bf0dd6d1715b02d5f6 WatchSource:0}: Error finding container 4d3d79add6339a78623f2c0c581c68e4a2917eac7bcfc8bf0dd6d1715b02d5f6: Status 404 returned error can't find the container with id 4d3d79add6339a78623f2c0c581c68e4a2917eac7bcfc8bf0dd6d1715b02d5f6 Oct 05 21:04:21 crc kubenswrapper[4754]: I1005 21:04:21.364765 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-r2txk" event={"ID":"577f8ef5-2c2c-4b67-91b5-55ce19fc52d2","Type":"ContainerStarted","Data":"807a407068f2a58d8daf882a657a6d71fd3fe506542325714dd77c9b4bf797bd"} Oct 05 21:04:21 crc kubenswrapper[4754]: I1005 21:04:21.365825 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-pqw5s" event={"ID":"f060400d-9b95-4e44-b94a-ba61d23e307c","Type":"ContainerStarted","Data":"0150289145debb4434160c5518b18d98418f674646a3dc6f9a3d69ab0a766cea"} Oct 05 21:04:21 crc kubenswrapper[4754]: I1005 21:04:21.366864 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-c5gxb" event={"ID":"e31db265-82f2-431c-af44-9a44bcb4762c","Type":"ContainerStarted","Data":"4d3d79add6339a78623f2c0c581c68e4a2917eac7bcfc8bf0dd6d1715b02d5f6"} Oct 05 21:04:25 crc kubenswrapper[4754]: I1005 21:04:25.394229 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-r2txk" event={"ID":"577f8ef5-2c2c-4b67-91b5-55ce19fc52d2","Type":"ContainerStarted","Data":"1e4ac857c7a654be541c5a931d5e8225425e630cf7abc0eefa6606aff5f2d5e2"} Oct 05 21:04:25 crc kubenswrapper[4754]: I1005 21:04:25.399816 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-5655c58dd6-r2txk" Oct 05 21:04:25 crc kubenswrapper[4754]: I1005 21:04:25.400561 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-pqw5s" event={"ID":"f060400d-9b95-4e44-b94a-ba61d23e307c","Type":"ContainerStarted","Data":"5642a5c02bc54cab564ccbb922455937930d0ac6bbcbc8ba6a84d47300082f28"} Oct 05 21:04:25 crc kubenswrapper[4754]: I1005 21:04:25.400693 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-c5gxb" event={"ID":"e31db265-82f2-431c-af44-9a44bcb4762c","Type":"ContainerStarted","Data":"964a3aa44c5b77bfb5abb1e830b8558346abe2a93d1a826f86dc7b5fbbd1ddc9"} Oct 05 21:04:25 crc kubenswrapper[4754]: I1005 21:04:25.414918 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-5655c58dd6-r2txk" podStartSLOduration=1.548049916 podStartE2EDuration="5.41489715s" podCreationTimestamp="2025-10-05 21:04:20 +0000 UTC" firstStartedPulling="2025-10-05 21:04:21.182213581 +0000 UTC m=+585.086332291" lastFinishedPulling="2025-10-05 21:04:25.049060795 +0000 UTC m=+588.953179525" observedRunningTime="2025-10-05 21:04:25.412740466 +0000 UTC m=+589.316859196" watchObservedRunningTime="2025-10-05 21:04:25.41489715 +0000 UTC m=+589.319015870" Oct 05 21:04:25 crc kubenswrapper[4754]: I1005 21:04:25.431379 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7f985d654d-pqw5s" podStartSLOduration=1.620651381 podStartE2EDuration="5.431353549s" podCreationTimestamp="2025-10-05 21:04:20 +0000 UTC" firstStartedPulling="2025-10-05 21:04:21.229192229 +0000 UTC m=+585.133310939" lastFinishedPulling="2025-10-05 21:04:25.039894377 +0000 UTC m=+588.944013107" observedRunningTime="2025-10-05 21:04:25.428300533 +0000 UTC m=+589.332419263" watchObservedRunningTime="2025-10-05 21:04:25.431353549 +0000 UTC m=+589.335472279" Oct 05 21:04:25 crc kubenswrapper[4754]: I1005 21:04:25.446797 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-5b446d88c5-c5gxb" podStartSLOduration=1.6475709090000001 podStartE2EDuration="5.446775762s" podCreationTimestamp="2025-10-05 21:04:20 +0000 UTC" firstStartedPulling="2025-10-05 21:04:21.263402239 +0000 UTC m=+585.167520949" lastFinishedPulling="2025-10-05 21:04:25.062607092 +0000 UTC m=+588.966725802" observedRunningTime="2025-10-05 21:04:25.446149477 +0000 UTC m=+589.350268217" watchObservedRunningTime="2025-10-05 21:04:25.446775762 +0000 UTC m=+589.350894472" Oct 05 21:04:30 crc kubenswrapper[4754]: I1005 21:04:30.814183 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-7rhps"] Oct 05 21:04:30 crc kubenswrapper[4754]: I1005 21:04:30.815477 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" podUID="b44790fb-fecb-4ec3-9816-8e23dd72ab33" containerName="ovn-controller" containerID="cri-o://822223f515b5add6d71d193fd5bc77df029b8d37a1cc07be4145b53c7226bf61" gracePeriod=30 Oct 05 21:04:30 crc kubenswrapper[4754]: I1005 21:04:30.815568 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" podUID="b44790fb-fecb-4ec3-9816-8e23dd72ab33" containerName="nbdb" containerID="cri-o://c3a63451ffaa14cacc3b78d11340f4ea81445fa3af96178e9b172d063d9697d5" gracePeriod=30 Oct 05 21:04:30 crc kubenswrapper[4754]: I1005 21:04:30.815713 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" podUID="b44790fb-fecb-4ec3-9816-8e23dd72ab33" containerName="northd" containerID="cri-o://f3cfd3f2274a8a46fa36660e94ac6c9c298c791283f12d52330c5594ac131998" gracePeriod=30 Oct 05 21:04:30 crc kubenswrapper[4754]: I1005 21:04:30.815745 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" podUID="b44790fb-fecb-4ec3-9816-8e23dd72ab33" containerName="sbdb" containerID="cri-o://a2acf9637fa25e5d3ffeb05224c11218c47defc679568fb928a498869613de4e" gracePeriod=30 Oct 05 21:04:30 crc kubenswrapper[4754]: I1005 21:04:30.815815 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" podUID="b44790fb-fecb-4ec3-9816-8e23dd72ab33" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://dfdd8b1e542ab676576c311e2923fb02861f92de91fe1f4b3e4b968c549c375d" gracePeriod=30 Oct 05 21:04:30 crc kubenswrapper[4754]: I1005 21:04:30.815851 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" podUID="b44790fb-fecb-4ec3-9816-8e23dd72ab33" containerName="ovn-acl-logging" containerID="cri-o://6f626cb108716da0b595844ea0251b7a04d49b5908a1d6dd0aef5cb2d402c4c9" gracePeriod=30 Oct 05 21:04:30 crc kubenswrapper[4754]: I1005 21:04:30.815914 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" podUID="b44790fb-fecb-4ec3-9816-8e23dd72ab33" containerName="kube-rbac-proxy-node" containerID="cri-o://5a7dc53e7f96e48213d724a2cb4859514df925005644108ae3a0c33b14dcf52d" gracePeriod=30 Oct 05 21:04:30 crc kubenswrapper[4754]: I1005 21:04:30.879166 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" podUID="b44790fb-fecb-4ec3-9816-8e23dd72ab33" containerName="ovnkube-controller" containerID="cri-o://40eef1aa41324f51a533acc8eb81eb76ae6eea135681050b6b79cef39fd3446f" gracePeriod=30 Oct 05 21:04:30 crc kubenswrapper[4754]: I1005 21:04:30.966263 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-5655c58dd6-r2txk" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.221947 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7rhps_b44790fb-fecb-4ec3-9816-8e23dd72ab33/ovnkube-controller/3.log" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.225163 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7rhps_b44790fb-fecb-4ec3-9816-8e23dd72ab33/ovn-acl-logging/0.log" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.225718 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7rhps_b44790fb-fecb-4ec3-9816-8e23dd72ab33/ovn-controller/0.log" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.226290 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.317665 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-s2jfr"] Oct 05 21:04:31 crc kubenswrapper[4754]: E1005 21:04:31.317906 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b44790fb-fecb-4ec3-9816-8e23dd72ab33" containerName="kube-rbac-proxy-node" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.317924 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="b44790fb-fecb-4ec3-9816-8e23dd72ab33" containerName="kube-rbac-proxy-node" Oct 05 21:04:31 crc kubenswrapper[4754]: E1005 21:04:31.317937 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b44790fb-fecb-4ec3-9816-8e23dd72ab33" containerName="ovnkube-controller" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.317945 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="b44790fb-fecb-4ec3-9816-8e23dd72ab33" containerName="ovnkube-controller" Oct 05 21:04:31 crc kubenswrapper[4754]: E1005 21:04:31.317953 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b44790fb-fecb-4ec3-9816-8e23dd72ab33" containerName="nbdb" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.317960 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="b44790fb-fecb-4ec3-9816-8e23dd72ab33" containerName="nbdb" Oct 05 21:04:31 crc kubenswrapper[4754]: E1005 21:04:31.317969 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b44790fb-fecb-4ec3-9816-8e23dd72ab33" containerName="ovnkube-controller" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.317977 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="b44790fb-fecb-4ec3-9816-8e23dd72ab33" containerName="ovnkube-controller" Oct 05 21:04:31 crc kubenswrapper[4754]: E1005 21:04:31.317989 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b44790fb-fecb-4ec3-9816-8e23dd72ab33" containerName="sbdb" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.317996 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="b44790fb-fecb-4ec3-9816-8e23dd72ab33" containerName="sbdb" Oct 05 21:04:31 crc kubenswrapper[4754]: E1005 21:04:31.318026 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b44790fb-fecb-4ec3-9816-8e23dd72ab33" containerName="kube-rbac-proxy-ovn-metrics" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.318035 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="b44790fb-fecb-4ec3-9816-8e23dd72ab33" containerName="kube-rbac-proxy-ovn-metrics" Oct 05 21:04:31 crc kubenswrapper[4754]: E1005 21:04:31.318050 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b44790fb-fecb-4ec3-9816-8e23dd72ab33" containerName="ovnkube-controller" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.318057 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="b44790fb-fecb-4ec3-9816-8e23dd72ab33" containerName="ovnkube-controller" Oct 05 21:04:31 crc kubenswrapper[4754]: E1005 21:04:31.318066 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b44790fb-fecb-4ec3-9816-8e23dd72ab33" containerName="ovn-acl-logging" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.318073 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="b44790fb-fecb-4ec3-9816-8e23dd72ab33" containerName="ovn-acl-logging" Oct 05 21:04:31 crc kubenswrapper[4754]: E1005 21:04:31.318083 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b44790fb-fecb-4ec3-9816-8e23dd72ab33" containerName="kubecfg-setup" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.318089 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="b44790fb-fecb-4ec3-9816-8e23dd72ab33" containerName="kubecfg-setup" Oct 05 21:04:31 crc kubenswrapper[4754]: E1005 21:04:31.318099 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b44790fb-fecb-4ec3-9816-8e23dd72ab33" containerName="northd" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.318106 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="b44790fb-fecb-4ec3-9816-8e23dd72ab33" containerName="northd" Oct 05 21:04:31 crc kubenswrapper[4754]: E1005 21:04:31.318119 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b44790fb-fecb-4ec3-9816-8e23dd72ab33" containerName="ovn-controller" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.318125 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="b44790fb-fecb-4ec3-9816-8e23dd72ab33" containerName="ovn-controller" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.318235 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="b44790fb-fecb-4ec3-9816-8e23dd72ab33" containerName="northd" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.318250 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="b44790fb-fecb-4ec3-9816-8e23dd72ab33" containerName="ovnkube-controller" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.318260 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="b44790fb-fecb-4ec3-9816-8e23dd72ab33" containerName="kube-rbac-proxy-ovn-metrics" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.318270 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="b44790fb-fecb-4ec3-9816-8e23dd72ab33" containerName="kube-rbac-proxy-node" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.318278 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="b44790fb-fecb-4ec3-9816-8e23dd72ab33" containerName="ovnkube-controller" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.318286 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="b44790fb-fecb-4ec3-9816-8e23dd72ab33" containerName="nbdb" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.318296 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="b44790fb-fecb-4ec3-9816-8e23dd72ab33" containerName="ovn-controller" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.318303 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="b44790fb-fecb-4ec3-9816-8e23dd72ab33" containerName="ovnkube-controller" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.318311 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="b44790fb-fecb-4ec3-9816-8e23dd72ab33" containerName="sbdb" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.318322 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="b44790fb-fecb-4ec3-9816-8e23dd72ab33" containerName="ovnkube-controller" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.318333 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="b44790fb-fecb-4ec3-9816-8e23dd72ab33" containerName="ovnkube-controller" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.318341 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="b44790fb-fecb-4ec3-9816-8e23dd72ab33" containerName="ovn-acl-logging" Oct 05 21:04:31 crc kubenswrapper[4754]: E1005 21:04:31.318443 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b44790fb-fecb-4ec3-9816-8e23dd72ab33" containerName="ovnkube-controller" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.318451 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="b44790fb-fecb-4ec3-9816-8e23dd72ab33" containerName="ovnkube-controller" Oct 05 21:04:31 crc kubenswrapper[4754]: E1005 21:04:31.318466 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b44790fb-fecb-4ec3-9816-8e23dd72ab33" containerName="ovnkube-controller" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.318473 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="b44790fb-fecb-4ec3-9816-8e23dd72ab33" containerName="ovnkube-controller" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.320392 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-s2jfr" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.365318 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/b44790fb-fecb-4ec3-9816-8e23dd72ab33-host-cni-bin\") pod \"b44790fb-fecb-4ec3-9816-8e23dd72ab33\" (UID: \"b44790fb-fecb-4ec3-9816-8e23dd72ab33\") " Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.365408 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b44790fb-fecb-4ec3-9816-8e23dd72ab33-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "b44790fb-fecb-4ec3-9816-8e23dd72ab33" (UID: "b44790fb-fecb-4ec3-9816-8e23dd72ab33"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.365416 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/b44790fb-fecb-4ec3-9816-8e23dd72ab33-systemd-units\") pod \"b44790fb-fecb-4ec3-9816-8e23dd72ab33\" (UID: \"b44790fb-fecb-4ec3-9816-8e23dd72ab33\") " Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.365486 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b44790fb-fecb-4ec3-9816-8e23dd72ab33-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "b44790fb-fecb-4ec3-9816-8e23dd72ab33" (UID: "b44790fb-fecb-4ec3-9816-8e23dd72ab33"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.365817 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/b44790fb-fecb-4ec3-9816-8e23dd72ab33-ovnkube-script-lib\") pod \"b44790fb-fecb-4ec3-9816-8e23dd72ab33\" (UID: \"b44790fb-fecb-4ec3-9816-8e23dd72ab33\") " Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.365853 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/b44790fb-fecb-4ec3-9816-8e23dd72ab33-host-run-netns\") pod \"b44790fb-fecb-4ec3-9816-8e23dd72ab33\" (UID: \"b44790fb-fecb-4ec3-9816-8e23dd72ab33\") " Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.365878 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/b44790fb-fecb-4ec3-9816-8e23dd72ab33-node-log\") pod \"b44790fb-fecb-4ec3-9816-8e23dd72ab33\" (UID: \"b44790fb-fecb-4ec3-9816-8e23dd72ab33\") " Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.365900 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/b44790fb-fecb-4ec3-9816-8e23dd72ab33-ovnkube-config\") pod \"b44790fb-fecb-4ec3-9816-8e23dd72ab33\" (UID: \"b44790fb-fecb-4ec3-9816-8e23dd72ab33\") " Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.365940 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b44790fb-fecb-4ec3-9816-8e23dd72ab33-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "b44790fb-fecb-4ec3-9816-8e23dd72ab33" (UID: "b44790fb-fecb-4ec3-9816-8e23dd72ab33"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.365945 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b44790fb-fecb-4ec3-9816-8e23dd72ab33-host-var-lib-cni-networks-ovn-kubernetes\") pod \"b44790fb-fecb-4ec3-9816-8e23dd72ab33\" (UID: \"b44790fb-fecb-4ec3-9816-8e23dd72ab33\") " Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.365971 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b44790fb-fecb-4ec3-9816-8e23dd72ab33-node-log" (OuterVolumeSpecName: "node-log") pod "b44790fb-fecb-4ec3-9816-8e23dd72ab33" (UID: "b44790fb-fecb-4ec3-9816-8e23dd72ab33"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.365975 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b44790fb-fecb-4ec3-9816-8e23dd72ab33-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "b44790fb-fecb-4ec3-9816-8e23dd72ab33" (UID: "b44790fb-fecb-4ec3-9816-8e23dd72ab33"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.366014 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/b44790fb-fecb-4ec3-9816-8e23dd72ab33-env-overrides\") pod \"b44790fb-fecb-4ec3-9816-8e23dd72ab33\" (UID: \"b44790fb-fecb-4ec3-9816-8e23dd72ab33\") " Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.366059 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/b44790fb-fecb-4ec3-9816-8e23dd72ab33-run-ovn\") pod \"b44790fb-fecb-4ec3-9816-8e23dd72ab33\" (UID: \"b44790fb-fecb-4ec3-9816-8e23dd72ab33\") " Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.366100 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/b44790fb-fecb-4ec3-9816-8e23dd72ab33-run-systemd\") pod \"b44790fb-fecb-4ec3-9816-8e23dd72ab33\" (UID: \"b44790fb-fecb-4ec3-9816-8e23dd72ab33\") " Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.366152 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b44790fb-fecb-4ec3-9816-8e23dd72ab33-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "b44790fb-fecb-4ec3-9816-8e23dd72ab33" (UID: "b44790fb-fecb-4ec3-9816-8e23dd72ab33"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.366121 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b44790fb-fecb-4ec3-9816-8e23dd72ab33-etc-openvswitch\") pod \"b44790fb-fecb-4ec3-9816-8e23dd72ab33\" (UID: \"b44790fb-fecb-4ec3-9816-8e23dd72ab33\") " Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.366183 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b44790fb-fecb-4ec3-9816-8e23dd72ab33-run-openvswitch\") pod \"b44790fb-fecb-4ec3-9816-8e23dd72ab33\" (UID: \"b44790fb-fecb-4ec3-9816-8e23dd72ab33\") " Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.366193 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b44790fb-fecb-4ec3-9816-8e23dd72ab33-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "b44790fb-fecb-4ec3-9816-8e23dd72ab33" (UID: "b44790fb-fecb-4ec3-9816-8e23dd72ab33"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.366202 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/b44790fb-fecb-4ec3-9816-8e23dd72ab33-ovn-node-metrics-cert\") pod \"b44790fb-fecb-4ec3-9816-8e23dd72ab33\" (UID: \"b44790fb-fecb-4ec3-9816-8e23dd72ab33\") " Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.366231 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/b44790fb-fecb-4ec3-9816-8e23dd72ab33-host-slash\") pod \"b44790fb-fecb-4ec3-9816-8e23dd72ab33\" (UID: \"b44790fb-fecb-4ec3-9816-8e23dd72ab33\") " Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.366240 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b44790fb-fecb-4ec3-9816-8e23dd72ab33-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "b44790fb-fecb-4ec3-9816-8e23dd72ab33" (UID: "b44790fb-fecb-4ec3-9816-8e23dd72ab33"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.366253 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/b44790fb-fecb-4ec3-9816-8e23dd72ab33-log-socket\") pod \"b44790fb-fecb-4ec3-9816-8e23dd72ab33\" (UID: \"b44790fb-fecb-4ec3-9816-8e23dd72ab33\") " Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.366261 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b44790fb-fecb-4ec3-9816-8e23dd72ab33-host-slash" (OuterVolumeSpecName: "host-slash") pod "b44790fb-fecb-4ec3-9816-8e23dd72ab33" (UID: "b44790fb-fecb-4ec3-9816-8e23dd72ab33"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.366272 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b44790fb-fecb-4ec3-9816-8e23dd72ab33-host-run-ovn-kubernetes\") pod \"b44790fb-fecb-4ec3-9816-8e23dd72ab33\" (UID: \"b44790fb-fecb-4ec3-9816-8e23dd72ab33\") " Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.366278 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b44790fb-fecb-4ec3-9816-8e23dd72ab33-log-socket" (OuterVolumeSpecName: "log-socket") pod "b44790fb-fecb-4ec3-9816-8e23dd72ab33" (UID: "b44790fb-fecb-4ec3-9816-8e23dd72ab33"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.366295 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b44790fb-fecb-4ec3-9816-8e23dd72ab33-var-lib-openvswitch\") pod \"b44790fb-fecb-4ec3-9816-8e23dd72ab33\" (UID: \"b44790fb-fecb-4ec3-9816-8e23dd72ab33\") " Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.366322 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/b44790fb-fecb-4ec3-9816-8e23dd72ab33-host-cni-netd\") pod \"b44790fb-fecb-4ec3-9816-8e23dd72ab33\" (UID: \"b44790fb-fecb-4ec3-9816-8e23dd72ab33\") " Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.366342 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/b44790fb-fecb-4ec3-9816-8e23dd72ab33-host-kubelet\") pod \"b44790fb-fecb-4ec3-9816-8e23dd72ab33\" (UID: \"b44790fb-fecb-4ec3-9816-8e23dd72ab33\") " Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.366341 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b44790fb-fecb-4ec3-9816-8e23dd72ab33-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "b44790fb-fecb-4ec3-9816-8e23dd72ab33" (UID: "b44790fb-fecb-4ec3-9816-8e23dd72ab33"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.366359 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b44790fb-fecb-4ec3-9816-8e23dd72ab33-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "b44790fb-fecb-4ec3-9816-8e23dd72ab33" (UID: "b44790fb-fecb-4ec3-9816-8e23dd72ab33"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.366368 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gnzsc\" (UniqueName: \"kubernetes.io/projected/b44790fb-fecb-4ec3-9816-8e23dd72ab33-kube-api-access-gnzsc\") pod \"b44790fb-fecb-4ec3-9816-8e23dd72ab33\" (UID: \"b44790fb-fecb-4ec3-9816-8e23dd72ab33\") " Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.366381 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b44790fb-fecb-4ec3-9816-8e23dd72ab33-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "b44790fb-fecb-4ec3-9816-8e23dd72ab33" (UID: "b44790fb-fecb-4ec3-9816-8e23dd72ab33"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.366383 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b44790fb-fecb-4ec3-9816-8e23dd72ab33-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "b44790fb-fecb-4ec3-9816-8e23dd72ab33" (UID: "b44790fb-fecb-4ec3-9816-8e23dd72ab33"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.366388 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b44790fb-fecb-4ec3-9816-8e23dd72ab33-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "b44790fb-fecb-4ec3-9816-8e23dd72ab33" (UID: "b44790fb-fecb-4ec3-9816-8e23dd72ab33"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.366406 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b44790fb-fecb-4ec3-9816-8e23dd72ab33-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "b44790fb-fecb-4ec3-9816-8e23dd72ab33" (UID: "b44790fb-fecb-4ec3-9816-8e23dd72ab33"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.366450 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/e3e264b2-489c-4b52-9915-22da8bbe4653-env-overrides\") pod \"ovnkube-node-s2jfr\" (UID: \"e3e264b2-489c-4b52-9915-22da8bbe4653\") " pod="openshift-ovn-kubernetes/ovnkube-node-s2jfr" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.366479 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/e3e264b2-489c-4b52-9915-22da8bbe4653-node-log\") pod \"ovnkube-node-s2jfr\" (UID: \"e3e264b2-489c-4b52-9915-22da8bbe4653\") " pod="openshift-ovn-kubernetes/ovnkube-node-s2jfr" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.366531 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b44790fb-fecb-4ec3-9816-8e23dd72ab33-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "b44790fb-fecb-4ec3-9816-8e23dd72ab33" (UID: "b44790fb-fecb-4ec3-9816-8e23dd72ab33"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.366554 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/e3e264b2-489c-4b52-9915-22da8bbe4653-host-cni-bin\") pod \"ovnkube-node-s2jfr\" (UID: \"e3e264b2-489c-4b52-9915-22da8bbe4653\") " pod="openshift-ovn-kubernetes/ovnkube-node-s2jfr" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.366587 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/e3e264b2-489c-4b52-9915-22da8bbe4653-ovn-node-metrics-cert\") pod \"ovnkube-node-s2jfr\" (UID: \"e3e264b2-489c-4b52-9915-22da8bbe4653\") " pod="openshift-ovn-kubernetes/ovnkube-node-s2jfr" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.366616 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/e3e264b2-489c-4b52-9915-22da8bbe4653-host-cni-netd\") pod \"ovnkube-node-s2jfr\" (UID: \"e3e264b2-489c-4b52-9915-22da8bbe4653\") " pod="openshift-ovn-kubernetes/ovnkube-node-s2jfr" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.366640 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e3e264b2-489c-4b52-9915-22da8bbe4653-etc-openvswitch\") pod \"ovnkube-node-s2jfr\" (UID: \"e3e264b2-489c-4b52-9915-22da8bbe4653\") " pod="openshift-ovn-kubernetes/ovnkube-node-s2jfr" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.366750 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hpwfh\" (UniqueName: \"kubernetes.io/projected/e3e264b2-489c-4b52-9915-22da8bbe4653-kube-api-access-hpwfh\") pod \"ovnkube-node-s2jfr\" (UID: \"e3e264b2-489c-4b52-9915-22da8bbe4653\") " pod="openshift-ovn-kubernetes/ovnkube-node-s2jfr" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.366772 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/e3e264b2-489c-4b52-9915-22da8bbe4653-host-run-netns\") pod \"ovnkube-node-s2jfr\" (UID: \"e3e264b2-489c-4b52-9915-22da8bbe4653\") " pod="openshift-ovn-kubernetes/ovnkube-node-s2jfr" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.366792 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/e3e264b2-489c-4b52-9915-22da8bbe4653-log-socket\") pod \"ovnkube-node-s2jfr\" (UID: \"e3e264b2-489c-4b52-9915-22da8bbe4653\") " pod="openshift-ovn-kubernetes/ovnkube-node-s2jfr" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.366809 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/e3e264b2-489c-4b52-9915-22da8bbe4653-ovnkube-script-lib\") pod \"ovnkube-node-s2jfr\" (UID: \"e3e264b2-489c-4b52-9915-22da8bbe4653\") " pod="openshift-ovn-kubernetes/ovnkube-node-s2jfr" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.366825 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/e3e264b2-489c-4b52-9915-22da8bbe4653-host-kubelet\") pod \"ovnkube-node-s2jfr\" (UID: \"e3e264b2-489c-4b52-9915-22da8bbe4653\") " pod="openshift-ovn-kubernetes/ovnkube-node-s2jfr" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.366869 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/e3e264b2-489c-4b52-9915-22da8bbe4653-ovnkube-config\") pod \"ovnkube-node-s2jfr\" (UID: \"e3e264b2-489c-4b52-9915-22da8bbe4653\") " pod="openshift-ovn-kubernetes/ovnkube-node-s2jfr" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.366965 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e3e264b2-489c-4b52-9915-22da8bbe4653-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-s2jfr\" (UID: \"e3e264b2-489c-4b52-9915-22da8bbe4653\") " pod="openshift-ovn-kubernetes/ovnkube-node-s2jfr" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.367059 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/e3e264b2-489c-4b52-9915-22da8bbe4653-run-ovn\") pod \"ovnkube-node-s2jfr\" (UID: \"e3e264b2-489c-4b52-9915-22da8bbe4653\") " pod="openshift-ovn-kubernetes/ovnkube-node-s2jfr" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.367087 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e3e264b2-489c-4b52-9915-22da8bbe4653-var-lib-openvswitch\") pod \"ovnkube-node-s2jfr\" (UID: \"e3e264b2-489c-4b52-9915-22da8bbe4653\") " pod="openshift-ovn-kubernetes/ovnkube-node-s2jfr" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.367145 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e3e264b2-489c-4b52-9915-22da8bbe4653-host-run-ovn-kubernetes\") pod \"ovnkube-node-s2jfr\" (UID: \"e3e264b2-489c-4b52-9915-22da8bbe4653\") " pod="openshift-ovn-kubernetes/ovnkube-node-s2jfr" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.367173 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/e3e264b2-489c-4b52-9915-22da8bbe4653-host-slash\") pod \"ovnkube-node-s2jfr\" (UID: \"e3e264b2-489c-4b52-9915-22da8bbe4653\") " pod="openshift-ovn-kubernetes/ovnkube-node-s2jfr" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.367234 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/e3e264b2-489c-4b52-9915-22da8bbe4653-run-systemd\") pod \"ovnkube-node-s2jfr\" (UID: \"e3e264b2-489c-4b52-9915-22da8bbe4653\") " pod="openshift-ovn-kubernetes/ovnkube-node-s2jfr" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.367277 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e3e264b2-489c-4b52-9915-22da8bbe4653-run-openvswitch\") pod \"ovnkube-node-s2jfr\" (UID: \"e3e264b2-489c-4b52-9915-22da8bbe4653\") " pod="openshift-ovn-kubernetes/ovnkube-node-s2jfr" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.367300 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/e3e264b2-489c-4b52-9915-22da8bbe4653-systemd-units\") pod \"ovnkube-node-s2jfr\" (UID: \"e3e264b2-489c-4b52-9915-22da8bbe4653\") " pod="openshift-ovn-kubernetes/ovnkube-node-s2jfr" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.367359 4754 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b44790fb-fecb-4ec3-9816-8e23dd72ab33-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.367375 4754 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b44790fb-fecb-4ec3-9816-8e23dd72ab33-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.367388 4754 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/b44790fb-fecb-4ec3-9816-8e23dd72ab33-host-cni-netd\") on node \"crc\" DevicePath \"\"" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.367400 4754 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/b44790fb-fecb-4ec3-9816-8e23dd72ab33-host-kubelet\") on node \"crc\" DevicePath \"\"" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.367413 4754 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/b44790fb-fecb-4ec3-9816-8e23dd72ab33-host-cni-bin\") on node \"crc\" DevicePath \"\"" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.367426 4754 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/b44790fb-fecb-4ec3-9816-8e23dd72ab33-systemd-units\") on node \"crc\" DevicePath \"\"" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.367437 4754 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/b44790fb-fecb-4ec3-9816-8e23dd72ab33-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.367451 4754 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/b44790fb-fecb-4ec3-9816-8e23dd72ab33-host-run-netns\") on node \"crc\" DevicePath \"\"" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.367462 4754 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/b44790fb-fecb-4ec3-9816-8e23dd72ab33-node-log\") on node \"crc\" DevicePath \"\"" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.367473 4754 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/b44790fb-fecb-4ec3-9816-8e23dd72ab33-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.367489 4754 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b44790fb-fecb-4ec3-9816-8e23dd72ab33-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.367524 4754 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/b44790fb-fecb-4ec3-9816-8e23dd72ab33-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.367536 4754 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/b44790fb-fecb-4ec3-9816-8e23dd72ab33-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.367546 4754 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b44790fb-fecb-4ec3-9816-8e23dd72ab33-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.367556 4754 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b44790fb-fecb-4ec3-9816-8e23dd72ab33-run-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.367566 4754 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/b44790fb-fecb-4ec3-9816-8e23dd72ab33-host-slash\") on node \"crc\" DevicePath \"\"" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.367576 4754 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/b44790fb-fecb-4ec3-9816-8e23dd72ab33-log-socket\") on node \"crc\" DevicePath \"\"" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.372033 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b44790fb-fecb-4ec3-9816-8e23dd72ab33-kube-api-access-gnzsc" (OuterVolumeSpecName: "kube-api-access-gnzsc") pod "b44790fb-fecb-4ec3-9816-8e23dd72ab33" (UID: "b44790fb-fecb-4ec3-9816-8e23dd72ab33"). InnerVolumeSpecName "kube-api-access-gnzsc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.372148 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b44790fb-fecb-4ec3-9816-8e23dd72ab33-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "b44790fb-fecb-4ec3-9816-8e23dd72ab33" (UID: "b44790fb-fecb-4ec3-9816-8e23dd72ab33"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.382154 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b44790fb-fecb-4ec3-9816-8e23dd72ab33-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "b44790fb-fecb-4ec3-9816-8e23dd72ab33" (UID: "b44790fb-fecb-4ec3-9816-8e23dd72ab33"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.445526 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-cn76s_02d5e3f9-73c2-4496-9aca-0787184aef19/kube-multus/2.log" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.446107 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-cn76s_02d5e3f9-73c2-4496-9aca-0787184aef19/kube-multus/1.log" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.446148 4754 generic.go:334] "Generic (PLEG): container finished" podID="02d5e3f9-73c2-4496-9aca-0787184aef19" containerID="b31d3403864365b964b38f903edb9d1f9d1567917a71eafd5a7120faf14feb99" exitCode=2 Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.446202 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-cn76s" event={"ID":"02d5e3f9-73c2-4496-9aca-0787184aef19","Type":"ContainerDied","Data":"b31d3403864365b964b38f903edb9d1f9d1567917a71eafd5a7120faf14feb99"} Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.446238 4754 scope.go:117] "RemoveContainer" containerID="5ed23b9cd9b57bb169a8ebbf16b7ff8d196eb14330df0edf2b9fcd2a8c9e2c88" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.446730 4754 scope.go:117] "RemoveContainer" containerID="b31d3403864365b964b38f903edb9d1f9d1567917a71eafd5a7120faf14feb99" Oct 05 21:04:31 crc kubenswrapper[4754]: E1005 21:04:31.446888 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-cn76s_openshift-multus(02d5e3f9-73c2-4496-9aca-0787184aef19)\"" pod="openshift-multus/multus-cn76s" podUID="02d5e3f9-73c2-4496-9aca-0787184aef19" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.452559 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7rhps_b44790fb-fecb-4ec3-9816-8e23dd72ab33/ovnkube-controller/3.log" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.456430 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7rhps_b44790fb-fecb-4ec3-9816-8e23dd72ab33/ovn-acl-logging/0.log" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.456865 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7rhps_b44790fb-fecb-4ec3-9816-8e23dd72ab33/ovn-controller/0.log" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.458443 4754 generic.go:334] "Generic (PLEG): container finished" podID="b44790fb-fecb-4ec3-9816-8e23dd72ab33" containerID="40eef1aa41324f51a533acc8eb81eb76ae6eea135681050b6b79cef39fd3446f" exitCode=0 Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.458465 4754 generic.go:334] "Generic (PLEG): container finished" podID="b44790fb-fecb-4ec3-9816-8e23dd72ab33" containerID="a2acf9637fa25e5d3ffeb05224c11218c47defc679568fb928a498869613de4e" exitCode=0 Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.458475 4754 generic.go:334] "Generic (PLEG): container finished" podID="b44790fb-fecb-4ec3-9816-8e23dd72ab33" containerID="c3a63451ffaa14cacc3b78d11340f4ea81445fa3af96178e9b172d063d9697d5" exitCode=0 Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.458482 4754 generic.go:334] "Generic (PLEG): container finished" podID="b44790fb-fecb-4ec3-9816-8e23dd72ab33" containerID="f3cfd3f2274a8a46fa36660e94ac6c9c298c791283f12d52330c5594ac131998" exitCode=0 Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.458510 4754 generic.go:334] "Generic (PLEG): container finished" podID="b44790fb-fecb-4ec3-9816-8e23dd72ab33" containerID="dfdd8b1e542ab676576c311e2923fb02861f92de91fe1f4b3e4b968c549c375d" exitCode=0 Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.458522 4754 generic.go:334] "Generic (PLEG): container finished" podID="b44790fb-fecb-4ec3-9816-8e23dd72ab33" containerID="5a7dc53e7f96e48213d724a2cb4859514df925005644108ae3a0c33b14dcf52d" exitCode=0 Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.458530 4754 generic.go:334] "Generic (PLEG): container finished" podID="b44790fb-fecb-4ec3-9816-8e23dd72ab33" containerID="6f626cb108716da0b595844ea0251b7a04d49b5908a1d6dd0aef5cb2d402c4c9" exitCode=143 Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.458538 4754 generic.go:334] "Generic (PLEG): container finished" podID="b44790fb-fecb-4ec3-9816-8e23dd72ab33" containerID="822223f515b5add6d71d193fd5bc77df029b8d37a1cc07be4145b53c7226bf61" exitCode=143 Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.458561 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" event={"ID":"b44790fb-fecb-4ec3-9816-8e23dd72ab33","Type":"ContainerDied","Data":"40eef1aa41324f51a533acc8eb81eb76ae6eea135681050b6b79cef39fd3446f"} Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.458589 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" event={"ID":"b44790fb-fecb-4ec3-9816-8e23dd72ab33","Type":"ContainerDied","Data":"a2acf9637fa25e5d3ffeb05224c11218c47defc679568fb928a498869613de4e"} Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.458604 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" event={"ID":"b44790fb-fecb-4ec3-9816-8e23dd72ab33","Type":"ContainerDied","Data":"c3a63451ffaa14cacc3b78d11340f4ea81445fa3af96178e9b172d063d9697d5"} Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.458616 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" event={"ID":"b44790fb-fecb-4ec3-9816-8e23dd72ab33","Type":"ContainerDied","Data":"f3cfd3f2274a8a46fa36660e94ac6c9c298c791283f12d52330c5594ac131998"} Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.458629 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" event={"ID":"b44790fb-fecb-4ec3-9816-8e23dd72ab33","Type":"ContainerDied","Data":"dfdd8b1e542ab676576c311e2923fb02861f92de91fe1f4b3e4b968c549c375d"} Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.458642 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" event={"ID":"b44790fb-fecb-4ec3-9816-8e23dd72ab33","Type":"ContainerDied","Data":"5a7dc53e7f96e48213d724a2cb4859514df925005644108ae3a0c33b14dcf52d"} Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.458661 4754 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"40eef1aa41324f51a533acc8eb81eb76ae6eea135681050b6b79cef39fd3446f"} Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.458671 4754 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7c990164cb7213726a9086a630586e38da00d62fdb57aed8418bbec265a54705"} Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.458677 4754 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a2acf9637fa25e5d3ffeb05224c11218c47defc679568fb928a498869613de4e"} Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.458682 4754 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c3a63451ffaa14cacc3b78d11340f4ea81445fa3af96178e9b172d063d9697d5"} Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.458687 4754 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f3cfd3f2274a8a46fa36660e94ac6c9c298c791283f12d52330c5594ac131998"} Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.458692 4754 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"dfdd8b1e542ab676576c311e2923fb02861f92de91fe1f4b3e4b968c549c375d"} Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.458698 4754 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5a7dc53e7f96e48213d724a2cb4859514df925005644108ae3a0c33b14dcf52d"} Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.458703 4754 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6f626cb108716da0b595844ea0251b7a04d49b5908a1d6dd0aef5cb2d402c4c9"} Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.458707 4754 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"822223f515b5add6d71d193fd5bc77df029b8d37a1cc07be4145b53c7226bf61"} Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.458712 4754 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"35ebeb7a87d5d47b6567917d38eca38a3e622c497fba58971296c4609c287423"} Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.458718 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" event={"ID":"b44790fb-fecb-4ec3-9816-8e23dd72ab33","Type":"ContainerDied","Data":"6f626cb108716da0b595844ea0251b7a04d49b5908a1d6dd0aef5cb2d402c4c9"} Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.458644 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.458726 4754 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"40eef1aa41324f51a533acc8eb81eb76ae6eea135681050b6b79cef39fd3446f"} Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.458837 4754 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7c990164cb7213726a9086a630586e38da00d62fdb57aed8418bbec265a54705"} Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.458849 4754 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a2acf9637fa25e5d3ffeb05224c11218c47defc679568fb928a498869613de4e"} Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.458856 4754 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c3a63451ffaa14cacc3b78d11340f4ea81445fa3af96178e9b172d063d9697d5"} Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.458862 4754 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f3cfd3f2274a8a46fa36660e94ac6c9c298c791283f12d52330c5594ac131998"} Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.458867 4754 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"dfdd8b1e542ab676576c311e2923fb02861f92de91fe1f4b3e4b968c549c375d"} Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.458872 4754 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5a7dc53e7f96e48213d724a2cb4859514df925005644108ae3a0c33b14dcf52d"} Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.458877 4754 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6f626cb108716da0b595844ea0251b7a04d49b5908a1d6dd0aef5cb2d402c4c9"} Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.458882 4754 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"822223f515b5add6d71d193fd5bc77df029b8d37a1cc07be4145b53c7226bf61"} Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.458887 4754 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"35ebeb7a87d5d47b6567917d38eca38a3e622c497fba58971296c4609c287423"} Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.458899 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" event={"ID":"b44790fb-fecb-4ec3-9816-8e23dd72ab33","Type":"ContainerDied","Data":"822223f515b5add6d71d193fd5bc77df029b8d37a1cc07be4145b53c7226bf61"} Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.458912 4754 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"40eef1aa41324f51a533acc8eb81eb76ae6eea135681050b6b79cef39fd3446f"} Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.458921 4754 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7c990164cb7213726a9086a630586e38da00d62fdb57aed8418bbec265a54705"} Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.458927 4754 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a2acf9637fa25e5d3ffeb05224c11218c47defc679568fb928a498869613de4e"} Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.458934 4754 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c3a63451ffaa14cacc3b78d11340f4ea81445fa3af96178e9b172d063d9697d5"} Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.458940 4754 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f3cfd3f2274a8a46fa36660e94ac6c9c298c791283f12d52330c5594ac131998"} Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.458947 4754 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"dfdd8b1e542ab676576c311e2923fb02861f92de91fe1f4b3e4b968c549c375d"} Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.458952 4754 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5a7dc53e7f96e48213d724a2cb4859514df925005644108ae3a0c33b14dcf52d"} Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.458957 4754 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6f626cb108716da0b595844ea0251b7a04d49b5908a1d6dd0aef5cb2d402c4c9"} Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.458963 4754 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"822223f515b5add6d71d193fd5bc77df029b8d37a1cc07be4145b53c7226bf61"} Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.458970 4754 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"35ebeb7a87d5d47b6567917d38eca38a3e622c497fba58971296c4609c287423"} Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.458980 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7rhps" event={"ID":"b44790fb-fecb-4ec3-9816-8e23dd72ab33","Type":"ContainerDied","Data":"fd436f3a6ea66b669ec762475a28d06906ed8dbaaff8e697b5cf00cf82b92886"} Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.458992 4754 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"40eef1aa41324f51a533acc8eb81eb76ae6eea135681050b6b79cef39fd3446f"} Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.459001 4754 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7c990164cb7213726a9086a630586e38da00d62fdb57aed8418bbec265a54705"} Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.459008 4754 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a2acf9637fa25e5d3ffeb05224c11218c47defc679568fb928a498869613de4e"} Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.459016 4754 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c3a63451ffaa14cacc3b78d11340f4ea81445fa3af96178e9b172d063d9697d5"} Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.459023 4754 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f3cfd3f2274a8a46fa36660e94ac6c9c298c791283f12d52330c5594ac131998"} Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.459030 4754 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"dfdd8b1e542ab676576c311e2923fb02861f92de91fe1f4b3e4b968c549c375d"} Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.459039 4754 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5a7dc53e7f96e48213d724a2cb4859514df925005644108ae3a0c33b14dcf52d"} Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.459045 4754 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6f626cb108716da0b595844ea0251b7a04d49b5908a1d6dd0aef5cb2d402c4c9"} Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.459053 4754 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"822223f515b5add6d71d193fd5bc77df029b8d37a1cc07be4145b53c7226bf61"} Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.459060 4754 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"35ebeb7a87d5d47b6567917d38eca38a3e622c497fba58971296c4609c287423"} Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.474711 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e3e264b2-489c-4b52-9915-22da8bbe4653-etc-openvswitch\") pod \"ovnkube-node-s2jfr\" (UID: \"e3e264b2-489c-4b52-9915-22da8bbe4653\") " pod="openshift-ovn-kubernetes/ovnkube-node-s2jfr" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.474845 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hpwfh\" (UniqueName: \"kubernetes.io/projected/e3e264b2-489c-4b52-9915-22da8bbe4653-kube-api-access-hpwfh\") pod \"ovnkube-node-s2jfr\" (UID: \"e3e264b2-489c-4b52-9915-22da8bbe4653\") " pod="openshift-ovn-kubernetes/ovnkube-node-s2jfr" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.474919 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/e3e264b2-489c-4b52-9915-22da8bbe4653-host-run-netns\") pod \"ovnkube-node-s2jfr\" (UID: \"e3e264b2-489c-4b52-9915-22da8bbe4653\") " pod="openshift-ovn-kubernetes/ovnkube-node-s2jfr" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.474973 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/e3e264b2-489c-4b52-9915-22da8bbe4653-log-socket\") pod \"ovnkube-node-s2jfr\" (UID: \"e3e264b2-489c-4b52-9915-22da8bbe4653\") " pod="openshift-ovn-kubernetes/ovnkube-node-s2jfr" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.474995 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/e3e264b2-489c-4b52-9915-22da8bbe4653-ovnkube-script-lib\") pod \"ovnkube-node-s2jfr\" (UID: \"e3e264b2-489c-4b52-9915-22da8bbe4653\") " pod="openshift-ovn-kubernetes/ovnkube-node-s2jfr" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.475101 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/e3e264b2-489c-4b52-9915-22da8bbe4653-host-kubelet\") pod \"ovnkube-node-s2jfr\" (UID: \"e3e264b2-489c-4b52-9915-22da8bbe4653\") " pod="openshift-ovn-kubernetes/ovnkube-node-s2jfr" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.475124 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/e3e264b2-489c-4b52-9915-22da8bbe4653-ovnkube-config\") pod \"ovnkube-node-s2jfr\" (UID: \"e3e264b2-489c-4b52-9915-22da8bbe4653\") " pod="openshift-ovn-kubernetes/ovnkube-node-s2jfr" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.475240 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e3e264b2-489c-4b52-9915-22da8bbe4653-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-s2jfr\" (UID: \"e3e264b2-489c-4b52-9915-22da8bbe4653\") " pod="openshift-ovn-kubernetes/ovnkube-node-s2jfr" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.475297 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/e3e264b2-489c-4b52-9915-22da8bbe4653-run-ovn\") pod \"ovnkube-node-s2jfr\" (UID: \"e3e264b2-489c-4b52-9915-22da8bbe4653\") " pod="openshift-ovn-kubernetes/ovnkube-node-s2jfr" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.475381 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e3e264b2-489c-4b52-9915-22da8bbe4653-var-lib-openvswitch\") pod \"ovnkube-node-s2jfr\" (UID: \"e3e264b2-489c-4b52-9915-22da8bbe4653\") " pod="openshift-ovn-kubernetes/ovnkube-node-s2jfr" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.475432 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e3e264b2-489c-4b52-9915-22da8bbe4653-host-run-ovn-kubernetes\") pod \"ovnkube-node-s2jfr\" (UID: \"e3e264b2-489c-4b52-9915-22da8bbe4653\") " pod="openshift-ovn-kubernetes/ovnkube-node-s2jfr" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.475633 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/e3e264b2-489c-4b52-9915-22da8bbe4653-host-slash\") pod \"ovnkube-node-s2jfr\" (UID: \"e3e264b2-489c-4b52-9915-22da8bbe4653\") " pod="openshift-ovn-kubernetes/ovnkube-node-s2jfr" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.475700 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/e3e264b2-489c-4b52-9915-22da8bbe4653-run-systemd\") pod \"ovnkube-node-s2jfr\" (UID: \"e3e264b2-489c-4b52-9915-22da8bbe4653\") " pod="openshift-ovn-kubernetes/ovnkube-node-s2jfr" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.475833 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e3e264b2-489c-4b52-9915-22da8bbe4653-run-openvswitch\") pod \"ovnkube-node-s2jfr\" (UID: \"e3e264b2-489c-4b52-9915-22da8bbe4653\") " pod="openshift-ovn-kubernetes/ovnkube-node-s2jfr" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.475859 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/e3e264b2-489c-4b52-9915-22da8bbe4653-systemd-units\") pod \"ovnkube-node-s2jfr\" (UID: \"e3e264b2-489c-4b52-9915-22da8bbe4653\") " pod="openshift-ovn-kubernetes/ovnkube-node-s2jfr" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.475931 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/e3e264b2-489c-4b52-9915-22da8bbe4653-env-overrides\") pod \"ovnkube-node-s2jfr\" (UID: \"e3e264b2-489c-4b52-9915-22da8bbe4653\") " pod="openshift-ovn-kubernetes/ovnkube-node-s2jfr" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.475967 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/e3e264b2-489c-4b52-9915-22da8bbe4653-node-log\") pod \"ovnkube-node-s2jfr\" (UID: \"e3e264b2-489c-4b52-9915-22da8bbe4653\") " pod="openshift-ovn-kubernetes/ovnkube-node-s2jfr" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.475991 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/e3e264b2-489c-4b52-9915-22da8bbe4653-host-cni-bin\") pod \"ovnkube-node-s2jfr\" (UID: \"e3e264b2-489c-4b52-9915-22da8bbe4653\") " pod="openshift-ovn-kubernetes/ovnkube-node-s2jfr" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.476015 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/e3e264b2-489c-4b52-9915-22da8bbe4653-ovn-node-metrics-cert\") pod \"ovnkube-node-s2jfr\" (UID: \"e3e264b2-489c-4b52-9915-22da8bbe4653\") " pod="openshift-ovn-kubernetes/ovnkube-node-s2jfr" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.476075 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/e3e264b2-489c-4b52-9915-22da8bbe4653-host-cni-netd\") pod \"ovnkube-node-s2jfr\" (UID: \"e3e264b2-489c-4b52-9915-22da8bbe4653\") " pod="openshift-ovn-kubernetes/ovnkube-node-s2jfr" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.477005 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gnzsc\" (UniqueName: \"kubernetes.io/projected/b44790fb-fecb-4ec3-9816-8e23dd72ab33-kube-api-access-gnzsc\") on node \"crc\" DevicePath \"\"" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.477986 4754 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/b44790fb-fecb-4ec3-9816-8e23dd72ab33-run-systemd\") on node \"crc\" DevicePath \"\"" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.478107 4754 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/b44790fb-fecb-4ec3-9816-8e23dd72ab33-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.478187 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/e3e264b2-489c-4b52-9915-22da8bbe4653-host-cni-netd\") pod \"ovnkube-node-s2jfr\" (UID: \"e3e264b2-489c-4b52-9915-22da8bbe4653\") " pod="openshift-ovn-kubernetes/ovnkube-node-s2jfr" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.478241 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e3e264b2-489c-4b52-9915-22da8bbe4653-etc-openvswitch\") pod \"ovnkube-node-s2jfr\" (UID: \"e3e264b2-489c-4b52-9915-22da8bbe4653\") " pod="openshift-ovn-kubernetes/ovnkube-node-s2jfr" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.479191 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/e3e264b2-489c-4b52-9915-22da8bbe4653-host-run-netns\") pod \"ovnkube-node-s2jfr\" (UID: \"e3e264b2-489c-4b52-9915-22da8bbe4653\") " pod="openshift-ovn-kubernetes/ovnkube-node-s2jfr" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.479249 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/e3e264b2-489c-4b52-9915-22da8bbe4653-log-socket\") pod \"ovnkube-node-s2jfr\" (UID: \"e3e264b2-489c-4b52-9915-22da8bbe4653\") " pod="openshift-ovn-kubernetes/ovnkube-node-s2jfr" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.480157 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/e3e264b2-489c-4b52-9915-22da8bbe4653-ovnkube-script-lib\") pod \"ovnkube-node-s2jfr\" (UID: \"e3e264b2-489c-4b52-9915-22da8bbe4653\") " pod="openshift-ovn-kubernetes/ovnkube-node-s2jfr" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.480231 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/e3e264b2-489c-4b52-9915-22da8bbe4653-host-kubelet\") pod \"ovnkube-node-s2jfr\" (UID: \"e3e264b2-489c-4b52-9915-22da8bbe4653\") " pod="openshift-ovn-kubernetes/ovnkube-node-s2jfr" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.481344 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/e3e264b2-489c-4b52-9915-22da8bbe4653-ovnkube-config\") pod \"ovnkube-node-s2jfr\" (UID: \"e3e264b2-489c-4b52-9915-22da8bbe4653\") " pod="openshift-ovn-kubernetes/ovnkube-node-s2jfr" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.481404 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e3e264b2-489c-4b52-9915-22da8bbe4653-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-s2jfr\" (UID: \"e3e264b2-489c-4b52-9915-22da8bbe4653\") " pod="openshift-ovn-kubernetes/ovnkube-node-s2jfr" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.481438 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/e3e264b2-489c-4b52-9915-22da8bbe4653-run-ovn\") pod \"ovnkube-node-s2jfr\" (UID: \"e3e264b2-489c-4b52-9915-22da8bbe4653\") " pod="openshift-ovn-kubernetes/ovnkube-node-s2jfr" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.481912 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/e3e264b2-489c-4b52-9915-22da8bbe4653-systemd-units\") pod \"ovnkube-node-s2jfr\" (UID: \"e3e264b2-489c-4b52-9915-22da8bbe4653\") " pod="openshift-ovn-kubernetes/ovnkube-node-s2jfr" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.481977 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e3e264b2-489c-4b52-9915-22da8bbe4653-run-openvswitch\") pod \"ovnkube-node-s2jfr\" (UID: \"e3e264b2-489c-4b52-9915-22da8bbe4653\") " pod="openshift-ovn-kubernetes/ovnkube-node-s2jfr" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.482186 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e3e264b2-489c-4b52-9915-22da8bbe4653-var-lib-openvswitch\") pod \"ovnkube-node-s2jfr\" (UID: \"e3e264b2-489c-4b52-9915-22da8bbe4653\") " pod="openshift-ovn-kubernetes/ovnkube-node-s2jfr" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.482274 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e3e264b2-489c-4b52-9915-22da8bbe4653-host-run-ovn-kubernetes\") pod \"ovnkube-node-s2jfr\" (UID: \"e3e264b2-489c-4b52-9915-22da8bbe4653\") " pod="openshift-ovn-kubernetes/ovnkube-node-s2jfr" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.482304 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/e3e264b2-489c-4b52-9915-22da8bbe4653-node-log\") pod \"ovnkube-node-s2jfr\" (UID: \"e3e264b2-489c-4b52-9915-22da8bbe4653\") " pod="openshift-ovn-kubernetes/ovnkube-node-s2jfr" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.482672 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/e3e264b2-489c-4b52-9915-22da8bbe4653-host-cni-bin\") pod \"ovnkube-node-s2jfr\" (UID: \"e3e264b2-489c-4b52-9915-22da8bbe4653\") " pod="openshift-ovn-kubernetes/ovnkube-node-s2jfr" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.482694 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/e3e264b2-489c-4b52-9915-22da8bbe4653-host-slash\") pod \"ovnkube-node-s2jfr\" (UID: \"e3e264b2-489c-4b52-9915-22da8bbe4653\") " pod="openshift-ovn-kubernetes/ovnkube-node-s2jfr" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.482916 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/e3e264b2-489c-4b52-9915-22da8bbe4653-run-systemd\") pod \"ovnkube-node-s2jfr\" (UID: \"e3e264b2-489c-4b52-9915-22da8bbe4653\") " pod="openshift-ovn-kubernetes/ovnkube-node-s2jfr" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.483117 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/e3e264b2-489c-4b52-9915-22da8bbe4653-env-overrides\") pod \"ovnkube-node-s2jfr\" (UID: \"e3e264b2-489c-4b52-9915-22da8bbe4653\") " pod="openshift-ovn-kubernetes/ovnkube-node-s2jfr" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.485993 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/e3e264b2-489c-4b52-9915-22da8bbe4653-ovn-node-metrics-cert\") pod \"ovnkube-node-s2jfr\" (UID: \"e3e264b2-489c-4b52-9915-22da8bbe4653\") " pod="openshift-ovn-kubernetes/ovnkube-node-s2jfr" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.510314 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-7rhps"] Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.511044 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hpwfh\" (UniqueName: \"kubernetes.io/projected/e3e264b2-489c-4b52-9915-22da8bbe4653-kube-api-access-hpwfh\") pod \"ovnkube-node-s2jfr\" (UID: \"e3e264b2-489c-4b52-9915-22da8bbe4653\") " pod="openshift-ovn-kubernetes/ovnkube-node-s2jfr" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.512634 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-7rhps"] Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.548324 4754 scope.go:117] "RemoveContainer" containerID="40eef1aa41324f51a533acc8eb81eb76ae6eea135681050b6b79cef39fd3446f" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.564123 4754 scope.go:117] "RemoveContainer" containerID="7c990164cb7213726a9086a630586e38da00d62fdb57aed8418bbec265a54705" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.585679 4754 scope.go:117] "RemoveContainer" containerID="a2acf9637fa25e5d3ffeb05224c11218c47defc679568fb928a498869613de4e" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.602352 4754 scope.go:117] "RemoveContainer" containerID="c3a63451ffaa14cacc3b78d11340f4ea81445fa3af96178e9b172d063d9697d5" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.619913 4754 scope.go:117] "RemoveContainer" containerID="f3cfd3f2274a8a46fa36660e94ac6c9c298c791283f12d52330c5594ac131998" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.635051 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-s2jfr" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.635123 4754 scope.go:117] "RemoveContainer" containerID="dfdd8b1e542ab676576c311e2923fb02861f92de91fe1f4b3e4b968c549c375d" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.649011 4754 scope.go:117] "RemoveContainer" containerID="5a7dc53e7f96e48213d724a2cb4859514df925005644108ae3a0c33b14dcf52d" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.668718 4754 scope.go:117] "RemoveContainer" containerID="6f626cb108716da0b595844ea0251b7a04d49b5908a1d6dd0aef5cb2d402c4c9" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.694576 4754 scope.go:117] "RemoveContainer" containerID="822223f515b5add6d71d193fd5bc77df029b8d37a1cc07be4145b53c7226bf61" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.713719 4754 scope.go:117] "RemoveContainer" containerID="35ebeb7a87d5d47b6567917d38eca38a3e622c497fba58971296c4609c287423" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.737357 4754 scope.go:117] "RemoveContainer" containerID="40eef1aa41324f51a533acc8eb81eb76ae6eea135681050b6b79cef39fd3446f" Oct 05 21:04:31 crc kubenswrapper[4754]: E1005 21:04:31.737881 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"40eef1aa41324f51a533acc8eb81eb76ae6eea135681050b6b79cef39fd3446f\": container with ID starting with 40eef1aa41324f51a533acc8eb81eb76ae6eea135681050b6b79cef39fd3446f not found: ID does not exist" containerID="40eef1aa41324f51a533acc8eb81eb76ae6eea135681050b6b79cef39fd3446f" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.737994 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"40eef1aa41324f51a533acc8eb81eb76ae6eea135681050b6b79cef39fd3446f"} err="failed to get container status \"40eef1aa41324f51a533acc8eb81eb76ae6eea135681050b6b79cef39fd3446f\": rpc error: code = NotFound desc = could not find container \"40eef1aa41324f51a533acc8eb81eb76ae6eea135681050b6b79cef39fd3446f\": container with ID starting with 40eef1aa41324f51a533acc8eb81eb76ae6eea135681050b6b79cef39fd3446f not found: ID does not exist" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.738123 4754 scope.go:117] "RemoveContainer" containerID="7c990164cb7213726a9086a630586e38da00d62fdb57aed8418bbec265a54705" Oct 05 21:04:31 crc kubenswrapper[4754]: E1005 21:04:31.738574 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7c990164cb7213726a9086a630586e38da00d62fdb57aed8418bbec265a54705\": container with ID starting with 7c990164cb7213726a9086a630586e38da00d62fdb57aed8418bbec265a54705 not found: ID does not exist" containerID="7c990164cb7213726a9086a630586e38da00d62fdb57aed8418bbec265a54705" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.738635 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7c990164cb7213726a9086a630586e38da00d62fdb57aed8418bbec265a54705"} err="failed to get container status \"7c990164cb7213726a9086a630586e38da00d62fdb57aed8418bbec265a54705\": rpc error: code = NotFound desc = could not find container \"7c990164cb7213726a9086a630586e38da00d62fdb57aed8418bbec265a54705\": container with ID starting with 7c990164cb7213726a9086a630586e38da00d62fdb57aed8418bbec265a54705 not found: ID does not exist" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.738676 4754 scope.go:117] "RemoveContainer" containerID="a2acf9637fa25e5d3ffeb05224c11218c47defc679568fb928a498869613de4e" Oct 05 21:04:31 crc kubenswrapper[4754]: E1005 21:04:31.739063 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a2acf9637fa25e5d3ffeb05224c11218c47defc679568fb928a498869613de4e\": container with ID starting with a2acf9637fa25e5d3ffeb05224c11218c47defc679568fb928a498869613de4e not found: ID does not exist" containerID="a2acf9637fa25e5d3ffeb05224c11218c47defc679568fb928a498869613de4e" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.739151 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a2acf9637fa25e5d3ffeb05224c11218c47defc679568fb928a498869613de4e"} err="failed to get container status \"a2acf9637fa25e5d3ffeb05224c11218c47defc679568fb928a498869613de4e\": rpc error: code = NotFound desc = could not find container \"a2acf9637fa25e5d3ffeb05224c11218c47defc679568fb928a498869613de4e\": container with ID starting with a2acf9637fa25e5d3ffeb05224c11218c47defc679568fb928a498869613de4e not found: ID does not exist" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.739215 4754 scope.go:117] "RemoveContainer" containerID="c3a63451ffaa14cacc3b78d11340f4ea81445fa3af96178e9b172d063d9697d5" Oct 05 21:04:31 crc kubenswrapper[4754]: E1005 21:04:31.740650 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c3a63451ffaa14cacc3b78d11340f4ea81445fa3af96178e9b172d063d9697d5\": container with ID starting with c3a63451ffaa14cacc3b78d11340f4ea81445fa3af96178e9b172d063d9697d5 not found: ID does not exist" containerID="c3a63451ffaa14cacc3b78d11340f4ea81445fa3af96178e9b172d063d9697d5" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.740698 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c3a63451ffaa14cacc3b78d11340f4ea81445fa3af96178e9b172d063d9697d5"} err="failed to get container status \"c3a63451ffaa14cacc3b78d11340f4ea81445fa3af96178e9b172d063d9697d5\": rpc error: code = NotFound desc = could not find container \"c3a63451ffaa14cacc3b78d11340f4ea81445fa3af96178e9b172d063d9697d5\": container with ID starting with c3a63451ffaa14cacc3b78d11340f4ea81445fa3af96178e9b172d063d9697d5 not found: ID does not exist" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.740726 4754 scope.go:117] "RemoveContainer" containerID="f3cfd3f2274a8a46fa36660e94ac6c9c298c791283f12d52330c5594ac131998" Oct 05 21:04:31 crc kubenswrapper[4754]: E1005 21:04:31.741156 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f3cfd3f2274a8a46fa36660e94ac6c9c298c791283f12d52330c5594ac131998\": container with ID starting with f3cfd3f2274a8a46fa36660e94ac6c9c298c791283f12d52330c5594ac131998 not found: ID does not exist" containerID="f3cfd3f2274a8a46fa36660e94ac6c9c298c791283f12d52330c5594ac131998" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.741228 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f3cfd3f2274a8a46fa36660e94ac6c9c298c791283f12d52330c5594ac131998"} err="failed to get container status \"f3cfd3f2274a8a46fa36660e94ac6c9c298c791283f12d52330c5594ac131998\": rpc error: code = NotFound desc = could not find container \"f3cfd3f2274a8a46fa36660e94ac6c9c298c791283f12d52330c5594ac131998\": container with ID starting with f3cfd3f2274a8a46fa36660e94ac6c9c298c791283f12d52330c5594ac131998 not found: ID does not exist" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.741292 4754 scope.go:117] "RemoveContainer" containerID="dfdd8b1e542ab676576c311e2923fb02861f92de91fe1f4b3e4b968c549c375d" Oct 05 21:04:31 crc kubenswrapper[4754]: E1005 21:04:31.741650 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dfdd8b1e542ab676576c311e2923fb02861f92de91fe1f4b3e4b968c549c375d\": container with ID starting with dfdd8b1e542ab676576c311e2923fb02861f92de91fe1f4b3e4b968c549c375d not found: ID does not exist" containerID="dfdd8b1e542ab676576c311e2923fb02861f92de91fe1f4b3e4b968c549c375d" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.741739 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dfdd8b1e542ab676576c311e2923fb02861f92de91fe1f4b3e4b968c549c375d"} err="failed to get container status \"dfdd8b1e542ab676576c311e2923fb02861f92de91fe1f4b3e4b968c549c375d\": rpc error: code = NotFound desc = could not find container \"dfdd8b1e542ab676576c311e2923fb02861f92de91fe1f4b3e4b968c549c375d\": container with ID starting with dfdd8b1e542ab676576c311e2923fb02861f92de91fe1f4b3e4b968c549c375d not found: ID does not exist" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.741773 4754 scope.go:117] "RemoveContainer" containerID="5a7dc53e7f96e48213d724a2cb4859514df925005644108ae3a0c33b14dcf52d" Oct 05 21:04:31 crc kubenswrapper[4754]: E1005 21:04:31.742052 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5a7dc53e7f96e48213d724a2cb4859514df925005644108ae3a0c33b14dcf52d\": container with ID starting with 5a7dc53e7f96e48213d724a2cb4859514df925005644108ae3a0c33b14dcf52d not found: ID does not exist" containerID="5a7dc53e7f96e48213d724a2cb4859514df925005644108ae3a0c33b14dcf52d" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.742149 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5a7dc53e7f96e48213d724a2cb4859514df925005644108ae3a0c33b14dcf52d"} err="failed to get container status \"5a7dc53e7f96e48213d724a2cb4859514df925005644108ae3a0c33b14dcf52d\": rpc error: code = NotFound desc = could not find container \"5a7dc53e7f96e48213d724a2cb4859514df925005644108ae3a0c33b14dcf52d\": container with ID starting with 5a7dc53e7f96e48213d724a2cb4859514df925005644108ae3a0c33b14dcf52d not found: ID does not exist" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.742213 4754 scope.go:117] "RemoveContainer" containerID="6f626cb108716da0b595844ea0251b7a04d49b5908a1d6dd0aef5cb2d402c4c9" Oct 05 21:04:31 crc kubenswrapper[4754]: E1005 21:04:31.742570 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6f626cb108716da0b595844ea0251b7a04d49b5908a1d6dd0aef5cb2d402c4c9\": container with ID starting with 6f626cb108716da0b595844ea0251b7a04d49b5908a1d6dd0aef5cb2d402c4c9 not found: ID does not exist" containerID="6f626cb108716da0b595844ea0251b7a04d49b5908a1d6dd0aef5cb2d402c4c9" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.742612 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6f626cb108716da0b595844ea0251b7a04d49b5908a1d6dd0aef5cb2d402c4c9"} err="failed to get container status \"6f626cb108716da0b595844ea0251b7a04d49b5908a1d6dd0aef5cb2d402c4c9\": rpc error: code = NotFound desc = could not find container \"6f626cb108716da0b595844ea0251b7a04d49b5908a1d6dd0aef5cb2d402c4c9\": container with ID starting with 6f626cb108716da0b595844ea0251b7a04d49b5908a1d6dd0aef5cb2d402c4c9 not found: ID does not exist" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.742640 4754 scope.go:117] "RemoveContainer" containerID="822223f515b5add6d71d193fd5bc77df029b8d37a1cc07be4145b53c7226bf61" Oct 05 21:04:31 crc kubenswrapper[4754]: E1005 21:04:31.742853 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"822223f515b5add6d71d193fd5bc77df029b8d37a1cc07be4145b53c7226bf61\": container with ID starting with 822223f515b5add6d71d193fd5bc77df029b8d37a1cc07be4145b53c7226bf61 not found: ID does not exist" containerID="822223f515b5add6d71d193fd5bc77df029b8d37a1cc07be4145b53c7226bf61" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.742928 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"822223f515b5add6d71d193fd5bc77df029b8d37a1cc07be4145b53c7226bf61"} err="failed to get container status \"822223f515b5add6d71d193fd5bc77df029b8d37a1cc07be4145b53c7226bf61\": rpc error: code = NotFound desc = could not find container \"822223f515b5add6d71d193fd5bc77df029b8d37a1cc07be4145b53c7226bf61\": container with ID starting with 822223f515b5add6d71d193fd5bc77df029b8d37a1cc07be4145b53c7226bf61 not found: ID does not exist" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.742994 4754 scope.go:117] "RemoveContainer" containerID="35ebeb7a87d5d47b6567917d38eca38a3e622c497fba58971296c4609c287423" Oct 05 21:04:31 crc kubenswrapper[4754]: E1005 21:04:31.743254 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"35ebeb7a87d5d47b6567917d38eca38a3e622c497fba58971296c4609c287423\": container with ID starting with 35ebeb7a87d5d47b6567917d38eca38a3e622c497fba58971296c4609c287423 not found: ID does not exist" containerID="35ebeb7a87d5d47b6567917d38eca38a3e622c497fba58971296c4609c287423" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.743336 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"35ebeb7a87d5d47b6567917d38eca38a3e622c497fba58971296c4609c287423"} err="failed to get container status \"35ebeb7a87d5d47b6567917d38eca38a3e622c497fba58971296c4609c287423\": rpc error: code = NotFound desc = could not find container \"35ebeb7a87d5d47b6567917d38eca38a3e622c497fba58971296c4609c287423\": container with ID starting with 35ebeb7a87d5d47b6567917d38eca38a3e622c497fba58971296c4609c287423 not found: ID does not exist" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.743397 4754 scope.go:117] "RemoveContainer" containerID="40eef1aa41324f51a533acc8eb81eb76ae6eea135681050b6b79cef39fd3446f" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.745838 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"40eef1aa41324f51a533acc8eb81eb76ae6eea135681050b6b79cef39fd3446f"} err="failed to get container status \"40eef1aa41324f51a533acc8eb81eb76ae6eea135681050b6b79cef39fd3446f\": rpc error: code = NotFound desc = could not find container \"40eef1aa41324f51a533acc8eb81eb76ae6eea135681050b6b79cef39fd3446f\": container with ID starting with 40eef1aa41324f51a533acc8eb81eb76ae6eea135681050b6b79cef39fd3446f not found: ID does not exist" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.745884 4754 scope.go:117] "RemoveContainer" containerID="7c990164cb7213726a9086a630586e38da00d62fdb57aed8418bbec265a54705" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.746525 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7c990164cb7213726a9086a630586e38da00d62fdb57aed8418bbec265a54705"} err="failed to get container status \"7c990164cb7213726a9086a630586e38da00d62fdb57aed8418bbec265a54705\": rpc error: code = NotFound desc = could not find container \"7c990164cb7213726a9086a630586e38da00d62fdb57aed8418bbec265a54705\": container with ID starting with 7c990164cb7213726a9086a630586e38da00d62fdb57aed8418bbec265a54705 not found: ID does not exist" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.746635 4754 scope.go:117] "RemoveContainer" containerID="a2acf9637fa25e5d3ffeb05224c11218c47defc679568fb928a498869613de4e" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.746966 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a2acf9637fa25e5d3ffeb05224c11218c47defc679568fb928a498869613de4e"} err="failed to get container status \"a2acf9637fa25e5d3ffeb05224c11218c47defc679568fb928a498869613de4e\": rpc error: code = NotFound desc = could not find container \"a2acf9637fa25e5d3ffeb05224c11218c47defc679568fb928a498869613de4e\": container with ID starting with a2acf9637fa25e5d3ffeb05224c11218c47defc679568fb928a498869613de4e not found: ID does not exist" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.746989 4754 scope.go:117] "RemoveContainer" containerID="c3a63451ffaa14cacc3b78d11340f4ea81445fa3af96178e9b172d063d9697d5" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.747259 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c3a63451ffaa14cacc3b78d11340f4ea81445fa3af96178e9b172d063d9697d5"} err="failed to get container status \"c3a63451ffaa14cacc3b78d11340f4ea81445fa3af96178e9b172d063d9697d5\": rpc error: code = NotFound desc = could not find container \"c3a63451ffaa14cacc3b78d11340f4ea81445fa3af96178e9b172d063d9697d5\": container with ID starting with c3a63451ffaa14cacc3b78d11340f4ea81445fa3af96178e9b172d063d9697d5 not found: ID does not exist" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.747339 4754 scope.go:117] "RemoveContainer" containerID="f3cfd3f2274a8a46fa36660e94ac6c9c298c791283f12d52330c5594ac131998" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.747628 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f3cfd3f2274a8a46fa36660e94ac6c9c298c791283f12d52330c5594ac131998"} err="failed to get container status \"f3cfd3f2274a8a46fa36660e94ac6c9c298c791283f12d52330c5594ac131998\": rpc error: code = NotFound desc = could not find container \"f3cfd3f2274a8a46fa36660e94ac6c9c298c791283f12d52330c5594ac131998\": container with ID starting with f3cfd3f2274a8a46fa36660e94ac6c9c298c791283f12d52330c5594ac131998 not found: ID does not exist" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.747650 4754 scope.go:117] "RemoveContainer" containerID="dfdd8b1e542ab676576c311e2923fb02861f92de91fe1f4b3e4b968c549c375d" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.747975 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dfdd8b1e542ab676576c311e2923fb02861f92de91fe1f4b3e4b968c549c375d"} err="failed to get container status \"dfdd8b1e542ab676576c311e2923fb02861f92de91fe1f4b3e4b968c549c375d\": rpc error: code = NotFound desc = could not find container \"dfdd8b1e542ab676576c311e2923fb02861f92de91fe1f4b3e4b968c549c375d\": container with ID starting with dfdd8b1e542ab676576c311e2923fb02861f92de91fe1f4b3e4b968c549c375d not found: ID does not exist" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.748143 4754 scope.go:117] "RemoveContainer" containerID="5a7dc53e7f96e48213d724a2cb4859514df925005644108ae3a0c33b14dcf52d" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.748479 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5a7dc53e7f96e48213d724a2cb4859514df925005644108ae3a0c33b14dcf52d"} err="failed to get container status \"5a7dc53e7f96e48213d724a2cb4859514df925005644108ae3a0c33b14dcf52d\": rpc error: code = NotFound desc = could not find container \"5a7dc53e7f96e48213d724a2cb4859514df925005644108ae3a0c33b14dcf52d\": container with ID starting with 5a7dc53e7f96e48213d724a2cb4859514df925005644108ae3a0c33b14dcf52d not found: ID does not exist" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.748560 4754 scope.go:117] "RemoveContainer" containerID="6f626cb108716da0b595844ea0251b7a04d49b5908a1d6dd0aef5cb2d402c4c9" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.748896 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6f626cb108716da0b595844ea0251b7a04d49b5908a1d6dd0aef5cb2d402c4c9"} err="failed to get container status \"6f626cb108716da0b595844ea0251b7a04d49b5908a1d6dd0aef5cb2d402c4c9\": rpc error: code = NotFound desc = could not find container \"6f626cb108716da0b595844ea0251b7a04d49b5908a1d6dd0aef5cb2d402c4c9\": container with ID starting with 6f626cb108716da0b595844ea0251b7a04d49b5908a1d6dd0aef5cb2d402c4c9 not found: ID does not exist" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.748938 4754 scope.go:117] "RemoveContainer" containerID="822223f515b5add6d71d193fd5bc77df029b8d37a1cc07be4145b53c7226bf61" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.749233 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"822223f515b5add6d71d193fd5bc77df029b8d37a1cc07be4145b53c7226bf61"} err="failed to get container status \"822223f515b5add6d71d193fd5bc77df029b8d37a1cc07be4145b53c7226bf61\": rpc error: code = NotFound desc = could not find container \"822223f515b5add6d71d193fd5bc77df029b8d37a1cc07be4145b53c7226bf61\": container with ID starting with 822223f515b5add6d71d193fd5bc77df029b8d37a1cc07be4145b53c7226bf61 not found: ID does not exist" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.749318 4754 scope.go:117] "RemoveContainer" containerID="35ebeb7a87d5d47b6567917d38eca38a3e622c497fba58971296c4609c287423" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.749679 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"35ebeb7a87d5d47b6567917d38eca38a3e622c497fba58971296c4609c287423"} err="failed to get container status \"35ebeb7a87d5d47b6567917d38eca38a3e622c497fba58971296c4609c287423\": rpc error: code = NotFound desc = could not find container \"35ebeb7a87d5d47b6567917d38eca38a3e622c497fba58971296c4609c287423\": container with ID starting with 35ebeb7a87d5d47b6567917d38eca38a3e622c497fba58971296c4609c287423 not found: ID does not exist" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.749716 4754 scope.go:117] "RemoveContainer" containerID="40eef1aa41324f51a533acc8eb81eb76ae6eea135681050b6b79cef39fd3446f" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.750046 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"40eef1aa41324f51a533acc8eb81eb76ae6eea135681050b6b79cef39fd3446f"} err="failed to get container status \"40eef1aa41324f51a533acc8eb81eb76ae6eea135681050b6b79cef39fd3446f\": rpc error: code = NotFound desc = could not find container \"40eef1aa41324f51a533acc8eb81eb76ae6eea135681050b6b79cef39fd3446f\": container with ID starting with 40eef1aa41324f51a533acc8eb81eb76ae6eea135681050b6b79cef39fd3446f not found: ID does not exist" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.750083 4754 scope.go:117] "RemoveContainer" containerID="7c990164cb7213726a9086a630586e38da00d62fdb57aed8418bbec265a54705" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.750332 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7c990164cb7213726a9086a630586e38da00d62fdb57aed8418bbec265a54705"} err="failed to get container status \"7c990164cb7213726a9086a630586e38da00d62fdb57aed8418bbec265a54705\": rpc error: code = NotFound desc = could not find container \"7c990164cb7213726a9086a630586e38da00d62fdb57aed8418bbec265a54705\": container with ID starting with 7c990164cb7213726a9086a630586e38da00d62fdb57aed8418bbec265a54705 not found: ID does not exist" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.750411 4754 scope.go:117] "RemoveContainer" containerID="a2acf9637fa25e5d3ffeb05224c11218c47defc679568fb928a498869613de4e" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.750894 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a2acf9637fa25e5d3ffeb05224c11218c47defc679568fb928a498869613de4e"} err="failed to get container status \"a2acf9637fa25e5d3ffeb05224c11218c47defc679568fb928a498869613de4e\": rpc error: code = NotFound desc = could not find container \"a2acf9637fa25e5d3ffeb05224c11218c47defc679568fb928a498869613de4e\": container with ID starting with a2acf9637fa25e5d3ffeb05224c11218c47defc679568fb928a498869613de4e not found: ID does not exist" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.750938 4754 scope.go:117] "RemoveContainer" containerID="c3a63451ffaa14cacc3b78d11340f4ea81445fa3af96178e9b172d063d9697d5" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.751214 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c3a63451ffaa14cacc3b78d11340f4ea81445fa3af96178e9b172d063d9697d5"} err="failed to get container status \"c3a63451ffaa14cacc3b78d11340f4ea81445fa3af96178e9b172d063d9697d5\": rpc error: code = NotFound desc = could not find container \"c3a63451ffaa14cacc3b78d11340f4ea81445fa3af96178e9b172d063d9697d5\": container with ID starting with c3a63451ffaa14cacc3b78d11340f4ea81445fa3af96178e9b172d063d9697d5 not found: ID does not exist" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.751246 4754 scope.go:117] "RemoveContainer" containerID="f3cfd3f2274a8a46fa36660e94ac6c9c298c791283f12d52330c5594ac131998" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.751660 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f3cfd3f2274a8a46fa36660e94ac6c9c298c791283f12d52330c5594ac131998"} err="failed to get container status \"f3cfd3f2274a8a46fa36660e94ac6c9c298c791283f12d52330c5594ac131998\": rpc error: code = NotFound desc = could not find container \"f3cfd3f2274a8a46fa36660e94ac6c9c298c791283f12d52330c5594ac131998\": container with ID starting with f3cfd3f2274a8a46fa36660e94ac6c9c298c791283f12d52330c5594ac131998 not found: ID does not exist" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.751752 4754 scope.go:117] "RemoveContainer" containerID="dfdd8b1e542ab676576c311e2923fb02861f92de91fe1f4b3e4b968c549c375d" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.752141 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dfdd8b1e542ab676576c311e2923fb02861f92de91fe1f4b3e4b968c549c375d"} err="failed to get container status \"dfdd8b1e542ab676576c311e2923fb02861f92de91fe1f4b3e4b968c549c375d\": rpc error: code = NotFound desc = could not find container \"dfdd8b1e542ab676576c311e2923fb02861f92de91fe1f4b3e4b968c549c375d\": container with ID starting with dfdd8b1e542ab676576c311e2923fb02861f92de91fe1f4b3e4b968c549c375d not found: ID does not exist" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.752186 4754 scope.go:117] "RemoveContainer" containerID="5a7dc53e7f96e48213d724a2cb4859514df925005644108ae3a0c33b14dcf52d" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.752434 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5a7dc53e7f96e48213d724a2cb4859514df925005644108ae3a0c33b14dcf52d"} err="failed to get container status \"5a7dc53e7f96e48213d724a2cb4859514df925005644108ae3a0c33b14dcf52d\": rpc error: code = NotFound desc = could not find container \"5a7dc53e7f96e48213d724a2cb4859514df925005644108ae3a0c33b14dcf52d\": container with ID starting with 5a7dc53e7f96e48213d724a2cb4859514df925005644108ae3a0c33b14dcf52d not found: ID does not exist" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.752528 4754 scope.go:117] "RemoveContainer" containerID="6f626cb108716da0b595844ea0251b7a04d49b5908a1d6dd0aef5cb2d402c4c9" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.752863 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6f626cb108716da0b595844ea0251b7a04d49b5908a1d6dd0aef5cb2d402c4c9"} err="failed to get container status \"6f626cb108716da0b595844ea0251b7a04d49b5908a1d6dd0aef5cb2d402c4c9\": rpc error: code = NotFound desc = could not find container \"6f626cb108716da0b595844ea0251b7a04d49b5908a1d6dd0aef5cb2d402c4c9\": container with ID starting with 6f626cb108716da0b595844ea0251b7a04d49b5908a1d6dd0aef5cb2d402c4c9 not found: ID does not exist" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.752945 4754 scope.go:117] "RemoveContainer" containerID="822223f515b5add6d71d193fd5bc77df029b8d37a1cc07be4145b53c7226bf61" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.753301 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"822223f515b5add6d71d193fd5bc77df029b8d37a1cc07be4145b53c7226bf61"} err="failed to get container status \"822223f515b5add6d71d193fd5bc77df029b8d37a1cc07be4145b53c7226bf61\": rpc error: code = NotFound desc = could not find container \"822223f515b5add6d71d193fd5bc77df029b8d37a1cc07be4145b53c7226bf61\": container with ID starting with 822223f515b5add6d71d193fd5bc77df029b8d37a1cc07be4145b53c7226bf61 not found: ID does not exist" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.753345 4754 scope.go:117] "RemoveContainer" containerID="35ebeb7a87d5d47b6567917d38eca38a3e622c497fba58971296c4609c287423" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.753590 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"35ebeb7a87d5d47b6567917d38eca38a3e622c497fba58971296c4609c287423"} err="failed to get container status \"35ebeb7a87d5d47b6567917d38eca38a3e622c497fba58971296c4609c287423\": rpc error: code = NotFound desc = could not find container \"35ebeb7a87d5d47b6567917d38eca38a3e622c497fba58971296c4609c287423\": container with ID starting with 35ebeb7a87d5d47b6567917d38eca38a3e622c497fba58971296c4609c287423 not found: ID does not exist" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.753667 4754 scope.go:117] "RemoveContainer" containerID="40eef1aa41324f51a533acc8eb81eb76ae6eea135681050b6b79cef39fd3446f" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.754052 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"40eef1aa41324f51a533acc8eb81eb76ae6eea135681050b6b79cef39fd3446f"} err="failed to get container status \"40eef1aa41324f51a533acc8eb81eb76ae6eea135681050b6b79cef39fd3446f\": rpc error: code = NotFound desc = could not find container \"40eef1aa41324f51a533acc8eb81eb76ae6eea135681050b6b79cef39fd3446f\": container with ID starting with 40eef1aa41324f51a533acc8eb81eb76ae6eea135681050b6b79cef39fd3446f not found: ID does not exist" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.754078 4754 scope.go:117] "RemoveContainer" containerID="7c990164cb7213726a9086a630586e38da00d62fdb57aed8418bbec265a54705" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.754339 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7c990164cb7213726a9086a630586e38da00d62fdb57aed8418bbec265a54705"} err="failed to get container status \"7c990164cb7213726a9086a630586e38da00d62fdb57aed8418bbec265a54705\": rpc error: code = NotFound desc = could not find container \"7c990164cb7213726a9086a630586e38da00d62fdb57aed8418bbec265a54705\": container with ID starting with 7c990164cb7213726a9086a630586e38da00d62fdb57aed8418bbec265a54705 not found: ID does not exist" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.754377 4754 scope.go:117] "RemoveContainer" containerID="a2acf9637fa25e5d3ffeb05224c11218c47defc679568fb928a498869613de4e" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.754689 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a2acf9637fa25e5d3ffeb05224c11218c47defc679568fb928a498869613de4e"} err="failed to get container status \"a2acf9637fa25e5d3ffeb05224c11218c47defc679568fb928a498869613de4e\": rpc error: code = NotFound desc = could not find container \"a2acf9637fa25e5d3ffeb05224c11218c47defc679568fb928a498869613de4e\": container with ID starting with a2acf9637fa25e5d3ffeb05224c11218c47defc679568fb928a498869613de4e not found: ID does not exist" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.754785 4754 scope.go:117] "RemoveContainer" containerID="c3a63451ffaa14cacc3b78d11340f4ea81445fa3af96178e9b172d063d9697d5" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.755060 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c3a63451ffaa14cacc3b78d11340f4ea81445fa3af96178e9b172d063d9697d5"} err="failed to get container status \"c3a63451ffaa14cacc3b78d11340f4ea81445fa3af96178e9b172d063d9697d5\": rpc error: code = NotFound desc = could not find container \"c3a63451ffaa14cacc3b78d11340f4ea81445fa3af96178e9b172d063d9697d5\": container with ID starting with c3a63451ffaa14cacc3b78d11340f4ea81445fa3af96178e9b172d063d9697d5 not found: ID does not exist" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.755139 4754 scope.go:117] "RemoveContainer" containerID="f3cfd3f2274a8a46fa36660e94ac6c9c298c791283f12d52330c5594ac131998" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.755482 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f3cfd3f2274a8a46fa36660e94ac6c9c298c791283f12d52330c5594ac131998"} err="failed to get container status \"f3cfd3f2274a8a46fa36660e94ac6c9c298c791283f12d52330c5594ac131998\": rpc error: code = NotFound desc = could not find container \"f3cfd3f2274a8a46fa36660e94ac6c9c298c791283f12d52330c5594ac131998\": container with ID starting with f3cfd3f2274a8a46fa36660e94ac6c9c298c791283f12d52330c5594ac131998 not found: ID does not exist" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.755554 4754 scope.go:117] "RemoveContainer" containerID="dfdd8b1e542ab676576c311e2923fb02861f92de91fe1f4b3e4b968c549c375d" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.755825 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dfdd8b1e542ab676576c311e2923fb02861f92de91fe1f4b3e4b968c549c375d"} err="failed to get container status \"dfdd8b1e542ab676576c311e2923fb02861f92de91fe1f4b3e4b968c549c375d\": rpc error: code = NotFound desc = could not find container \"dfdd8b1e542ab676576c311e2923fb02861f92de91fe1f4b3e4b968c549c375d\": container with ID starting with dfdd8b1e542ab676576c311e2923fb02861f92de91fe1f4b3e4b968c549c375d not found: ID does not exist" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.755911 4754 scope.go:117] "RemoveContainer" containerID="5a7dc53e7f96e48213d724a2cb4859514df925005644108ae3a0c33b14dcf52d" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.756582 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5a7dc53e7f96e48213d724a2cb4859514df925005644108ae3a0c33b14dcf52d"} err="failed to get container status \"5a7dc53e7f96e48213d724a2cb4859514df925005644108ae3a0c33b14dcf52d\": rpc error: code = NotFound desc = could not find container \"5a7dc53e7f96e48213d724a2cb4859514df925005644108ae3a0c33b14dcf52d\": container with ID starting with 5a7dc53e7f96e48213d724a2cb4859514df925005644108ae3a0c33b14dcf52d not found: ID does not exist" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.756627 4754 scope.go:117] "RemoveContainer" containerID="6f626cb108716da0b595844ea0251b7a04d49b5908a1d6dd0aef5cb2d402c4c9" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.757037 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6f626cb108716da0b595844ea0251b7a04d49b5908a1d6dd0aef5cb2d402c4c9"} err="failed to get container status \"6f626cb108716da0b595844ea0251b7a04d49b5908a1d6dd0aef5cb2d402c4c9\": rpc error: code = NotFound desc = could not find container \"6f626cb108716da0b595844ea0251b7a04d49b5908a1d6dd0aef5cb2d402c4c9\": container with ID starting with 6f626cb108716da0b595844ea0251b7a04d49b5908a1d6dd0aef5cb2d402c4c9 not found: ID does not exist" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.757118 4754 scope.go:117] "RemoveContainer" containerID="822223f515b5add6d71d193fd5bc77df029b8d37a1cc07be4145b53c7226bf61" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.757434 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"822223f515b5add6d71d193fd5bc77df029b8d37a1cc07be4145b53c7226bf61"} err="failed to get container status \"822223f515b5add6d71d193fd5bc77df029b8d37a1cc07be4145b53c7226bf61\": rpc error: code = NotFound desc = could not find container \"822223f515b5add6d71d193fd5bc77df029b8d37a1cc07be4145b53c7226bf61\": container with ID starting with 822223f515b5add6d71d193fd5bc77df029b8d37a1cc07be4145b53c7226bf61 not found: ID does not exist" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.757474 4754 scope.go:117] "RemoveContainer" containerID="35ebeb7a87d5d47b6567917d38eca38a3e622c497fba58971296c4609c287423" Oct 05 21:04:31 crc kubenswrapper[4754]: I1005 21:04:31.757875 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"35ebeb7a87d5d47b6567917d38eca38a3e622c497fba58971296c4609c287423"} err="failed to get container status \"35ebeb7a87d5d47b6567917d38eca38a3e622c497fba58971296c4609c287423\": rpc error: code = NotFound desc = could not find container \"35ebeb7a87d5d47b6567917d38eca38a3e622c497fba58971296c4609c287423\": container with ID starting with 35ebeb7a87d5d47b6567917d38eca38a3e622c497fba58971296c4609c287423 not found: ID does not exist" Oct 05 21:04:32 crc kubenswrapper[4754]: I1005 21:04:32.468262 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-cn76s_02d5e3f9-73c2-4496-9aca-0787184aef19/kube-multus/2.log" Oct 05 21:04:32 crc kubenswrapper[4754]: I1005 21:04:32.470784 4754 generic.go:334] "Generic (PLEG): container finished" podID="e3e264b2-489c-4b52-9915-22da8bbe4653" containerID="017c0ec4b91ef29f989124549e614dceefcdef6ae8186041f2ec9cd810c02ff2" exitCode=0 Oct 05 21:04:32 crc kubenswrapper[4754]: I1005 21:04:32.470832 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s2jfr" event={"ID":"e3e264b2-489c-4b52-9915-22da8bbe4653","Type":"ContainerDied","Data":"017c0ec4b91ef29f989124549e614dceefcdef6ae8186041f2ec9cd810c02ff2"} Oct 05 21:04:32 crc kubenswrapper[4754]: I1005 21:04:32.470866 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s2jfr" event={"ID":"e3e264b2-489c-4b52-9915-22da8bbe4653","Type":"ContainerStarted","Data":"0285646bd92911803d20e7359db7e1d06408a84ada7ddb8d990e63b2e49babd0"} Oct 05 21:04:32 crc kubenswrapper[4754]: I1005 21:04:32.845818 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b44790fb-fecb-4ec3-9816-8e23dd72ab33" path="/var/lib/kubelet/pods/b44790fb-fecb-4ec3-9816-8e23dd72ab33/volumes" Oct 05 21:04:33 crc kubenswrapper[4754]: I1005 21:04:33.482823 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s2jfr" event={"ID":"e3e264b2-489c-4b52-9915-22da8bbe4653","Type":"ContainerStarted","Data":"43d4bf9cf5049234a819a73386402e74562015dc58d46d4b4b9f3f6b58ffe6cd"} Oct 05 21:04:33 crc kubenswrapper[4754]: I1005 21:04:33.483204 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s2jfr" event={"ID":"e3e264b2-489c-4b52-9915-22da8bbe4653","Type":"ContainerStarted","Data":"f291c93b8b1af6463a8a644ae781bfa578acb918036b30ef9afceefdc5310f45"} Oct 05 21:04:33 crc kubenswrapper[4754]: I1005 21:04:33.483222 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s2jfr" event={"ID":"e3e264b2-489c-4b52-9915-22da8bbe4653","Type":"ContainerStarted","Data":"2b6cbdc6eeb1ba530f7a78eccabd82f15f016e971e21167af5b33174185b45d6"} Oct 05 21:04:33 crc kubenswrapper[4754]: I1005 21:04:33.483236 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s2jfr" event={"ID":"e3e264b2-489c-4b52-9915-22da8bbe4653","Type":"ContainerStarted","Data":"db602f027a42828e904624fa4fdf67d95ce8f3261c7d5e2b06c74dd1f4d4d912"} Oct 05 21:04:33 crc kubenswrapper[4754]: I1005 21:04:33.483249 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s2jfr" event={"ID":"e3e264b2-489c-4b52-9915-22da8bbe4653","Type":"ContainerStarted","Data":"6a15c9bf7a3899e3354fcbef0deeb73a162ba10217e4a4ab0a74575651bdbf5a"} Oct 05 21:04:33 crc kubenswrapper[4754]: I1005 21:04:33.483264 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s2jfr" event={"ID":"e3e264b2-489c-4b52-9915-22da8bbe4653","Type":"ContainerStarted","Data":"c097f379f3283f054916ec95107fbdb25947752db5f485b6518f4942542afdd6"} Oct 05 21:04:35 crc kubenswrapper[4754]: I1005 21:04:35.245105 4754 patch_prober.go:28] interesting pod/machine-config-daemon-b2h9k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 21:04:35 crc kubenswrapper[4754]: I1005 21:04:35.245176 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 21:04:35 crc kubenswrapper[4754]: I1005 21:04:35.245235 4754 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" Oct 05 21:04:35 crc kubenswrapper[4754]: I1005 21:04:35.246204 4754 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0b09d5fa150c23a262d03541134e26faa0bf59c28b4299e50ad59b83bfa2351d"} pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 05 21:04:35 crc kubenswrapper[4754]: I1005 21:04:35.246267 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" containerName="machine-config-daemon" containerID="cri-o://0b09d5fa150c23a262d03541134e26faa0bf59c28b4299e50ad59b83bfa2351d" gracePeriod=600 Oct 05 21:04:35 crc kubenswrapper[4754]: I1005 21:04:35.498100 4754 generic.go:334] "Generic (PLEG): container finished" podID="be95a413-6a59-45b4-84b7-b43cae694a26" containerID="0b09d5fa150c23a262d03541134e26faa0bf59c28b4299e50ad59b83bfa2351d" exitCode=0 Oct 05 21:04:35 crc kubenswrapper[4754]: I1005 21:04:35.498192 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" event={"ID":"be95a413-6a59-45b4-84b7-b43cae694a26","Type":"ContainerDied","Data":"0b09d5fa150c23a262d03541134e26faa0bf59c28b4299e50ad59b83bfa2351d"} Oct 05 21:04:35 crc kubenswrapper[4754]: I1005 21:04:35.498629 4754 scope.go:117] "RemoveContainer" containerID="1748c85d7a12794d4f4515c3318cf93b10e679b7d52cfbe7c2fbc83011ee324c" Oct 05 21:04:36 crc kubenswrapper[4754]: I1005 21:04:36.512939 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s2jfr" event={"ID":"e3e264b2-489c-4b52-9915-22da8bbe4653","Type":"ContainerStarted","Data":"9376497b8836bdbfcb510166e01c4d306bcedfc556171c64cf19ecd53e665258"} Oct 05 21:04:36 crc kubenswrapper[4754]: I1005 21:04:36.518682 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" event={"ID":"be95a413-6a59-45b4-84b7-b43cae694a26","Type":"ContainerStarted","Data":"42efe442464282aff7d7a4c1a4006bea20a662c700019a0224e31e9c8b2dfd60"} Oct 05 21:04:38 crc kubenswrapper[4754]: I1005 21:04:38.545684 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s2jfr" event={"ID":"e3e264b2-489c-4b52-9915-22da8bbe4653","Type":"ContainerStarted","Data":"1b4bf7f12af310f8e6488d0d42ac9f626a3662358bf8b209410cd60674968938"} Oct 05 21:04:38 crc kubenswrapper[4754]: I1005 21:04:38.546113 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-s2jfr" Oct 05 21:04:38 crc kubenswrapper[4754]: I1005 21:04:38.576288 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-s2jfr" podStartSLOduration=7.576268478 podStartE2EDuration="7.576268478s" podCreationTimestamp="2025-10-05 21:04:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 21:04:38.572000512 +0000 UTC m=+602.476119262" watchObservedRunningTime="2025-10-05 21:04:38.576268478 +0000 UTC m=+602.480387188" Oct 05 21:04:38 crc kubenswrapper[4754]: I1005 21:04:38.625002 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-s2jfr" Oct 05 21:04:39 crc kubenswrapper[4754]: I1005 21:04:39.554368 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-s2jfr" Oct 05 21:04:39 crc kubenswrapper[4754]: I1005 21:04:39.554894 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-s2jfr" Oct 05 21:04:39 crc kubenswrapper[4754]: I1005 21:04:39.604916 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-s2jfr" Oct 05 21:04:44 crc kubenswrapper[4754]: I1005 21:04:44.838308 4754 scope.go:117] "RemoveContainer" containerID="b31d3403864365b964b38f903edb9d1f9d1567917a71eafd5a7120faf14feb99" Oct 05 21:04:44 crc kubenswrapper[4754]: E1005 21:04:44.839219 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-cn76s_openshift-multus(02d5e3f9-73c2-4496-9aca-0787184aef19)\"" pod="openshift-multus/multus-cn76s" podUID="02d5e3f9-73c2-4496-9aca-0787184aef19" Oct 05 21:04:56 crc kubenswrapper[4754]: I1005 21:04:56.842749 4754 scope.go:117] "RemoveContainer" containerID="b31d3403864365b964b38f903edb9d1f9d1567917a71eafd5a7120faf14feb99" Oct 05 21:04:57 crc kubenswrapper[4754]: I1005 21:04:57.699565 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-cn76s_02d5e3f9-73c2-4496-9aca-0787184aef19/kube-multus/2.log" Oct 05 21:04:57 crc kubenswrapper[4754]: I1005 21:04:57.700112 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-cn76s" event={"ID":"02d5e3f9-73c2-4496-9aca-0787184aef19","Type":"ContainerStarted","Data":"e03c91b66f45238cd37dd3e046d54413ca6f071732210f82d602272885222a62"} Oct 05 21:05:01 crc kubenswrapper[4754]: I1005 21:05:01.665100 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-s2jfr" Oct 05 21:05:12 crc kubenswrapper[4754]: I1005 21:05:12.300883 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cpbjks"] Oct 05 21:05:12 crc kubenswrapper[4754]: I1005 21:05:12.304967 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cpbjks" Oct 05 21:05:12 crc kubenswrapper[4754]: I1005 21:05:12.306295 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-znl6n\" (UniqueName: \"kubernetes.io/projected/f7edcdad-0804-4ae8-bb6d-1118b6d0f620-kube-api-access-znl6n\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cpbjks\" (UID: \"f7edcdad-0804-4ae8-bb6d-1118b6d0f620\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cpbjks" Oct 05 21:05:12 crc kubenswrapper[4754]: I1005 21:05:12.306354 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f7edcdad-0804-4ae8-bb6d-1118b6d0f620-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cpbjks\" (UID: \"f7edcdad-0804-4ae8-bb6d-1118b6d0f620\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cpbjks" Oct 05 21:05:12 crc kubenswrapper[4754]: I1005 21:05:12.306380 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f7edcdad-0804-4ae8-bb6d-1118b6d0f620-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cpbjks\" (UID: \"f7edcdad-0804-4ae8-bb6d-1118b6d0f620\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cpbjks" Oct 05 21:05:12 crc kubenswrapper[4754]: I1005 21:05:12.312145 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 05 21:05:12 crc kubenswrapper[4754]: I1005 21:05:12.359418 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cpbjks"] Oct 05 21:05:12 crc kubenswrapper[4754]: I1005 21:05:12.407701 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-znl6n\" (UniqueName: \"kubernetes.io/projected/f7edcdad-0804-4ae8-bb6d-1118b6d0f620-kube-api-access-znl6n\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cpbjks\" (UID: \"f7edcdad-0804-4ae8-bb6d-1118b6d0f620\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cpbjks" Oct 05 21:05:12 crc kubenswrapper[4754]: I1005 21:05:12.408041 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f7edcdad-0804-4ae8-bb6d-1118b6d0f620-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cpbjks\" (UID: \"f7edcdad-0804-4ae8-bb6d-1118b6d0f620\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cpbjks" Oct 05 21:05:12 crc kubenswrapper[4754]: I1005 21:05:12.408094 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f7edcdad-0804-4ae8-bb6d-1118b6d0f620-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cpbjks\" (UID: \"f7edcdad-0804-4ae8-bb6d-1118b6d0f620\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cpbjks" Oct 05 21:05:12 crc kubenswrapper[4754]: I1005 21:05:12.408621 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f7edcdad-0804-4ae8-bb6d-1118b6d0f620-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cpbjks\" (UID: \"f7edcdad-0804-4ae8-bb6d-1118b6d0f620\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cpbjks" Oct 05 21:05:12 crc kubenswrapper[4754]: I1005 21:05:12.408636 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f7edcdad-0804-4ae8-bb6d-1118b6d0f620-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cpbjks\" (UID: \"f7edcdad-0804-4ae8-bb6d-1118b6d0f620\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cpbjks" Oct 05 21:05:12 crc kubenswrapper[4754]: I1005 21:05:12.434948 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-znl6n\" (UniqueName: \"kubernetes.io/projected/f7edcdad-0804-4ae8-bb6d-1118b6d0f620-kube-api-access-znl6n\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cpbjks\" (UID: \"f7edcdad-0804-4ae8-bb6d-1118b6d0f620\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cpbjks" Oct 05 21:05:12 crc kubenswrapper[4754]: I1005 21:05:12.677487 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cpbjks" Oct 05 21:05:12 crc kubenswrapper[4754]: I1005 21:05:12.930823 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cpbjks"] Oct 05 21:05:13 crc kubenswrapper[4754]: I1005 21:05:13.812132 4754 generic.go:334] "Generic (PLEG): container finished" podID="f7edcdad-0804-4ae8-bb6d-1118b6d0f620" containerID="1c36cbf99d85bb6783250538a007f4b01e4f6acbe2434dd60f3635fffe200576" exitCode=0 Oct 05 21:05:13 crc kubenswrapper[4754]: I1005 21:05:13.812176 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cpbjks" event={"ID":"f7edcdad-0804-4ae8-bb6d-1118b6d0f620","Type":"ContainerDied","Data":"1c36cbf99d85bb6783250538a007f4b01e4f6acbe2434dd60f3635fffe200576"} Oct 05 21:05:13 crc kubenswrapper[4754]: I1005 21:05:13.812202 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cpbjks" event={"ID":"f7edcdad-0804-4ae8-bb6d-1118b6d0f620","Type":"ContainerStarted","Data":"de4db7a2e06f3c5a280293aa32c366f54edc5c3726c591b42ac7c812527eea8c"} Oct 05 21:05:15 crc kubenswrapper[4754]: I1005 21:05:15.832640 4754 generic.go:334] "Generic (PLEG): container finished" podID="f7edcdad-0804-4ae8-bb6d-1118b6d0f620" containerID="6598d17bfa6a518ca55630cb475621987d3e4e6bd381ca36a39bb39da32137c9" exitCode=0 Oct 05 21:05:15 crc kubenswrapper[4754]: I1005 21:05:15.833059 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cpbjks" event={"ID":"f7edcdad-0804-4ae8-bb6d-1118b6d0f620","Type":"ContainerDied","Data":"6598d17bfa6a518ca55630cb475621987d3e4e6bd381ca36a39bb39da32137c9"} Oct 05 21:05:16 crc kubenswrapper[4754]: I1005 21:05:16.843851 4754 generic.go:334] "Generic (PLEG): container finished" podID="f7edcdad-0804-4ae8-bb6d-1118b6d0f620" containerID="a71a498283bebdd6c4a1ceef0de2da16bd7de011b3281ffc8d87d7673b63cfee" exitCode=0 Oct 05 21:05:16 crc kubenswrapper[4754]: I1005 21:05:16.846415 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cpbjks" event={"ID":"f7edcdad-0804-4ae8-bb6d-1118b6d0f620","Type":"ContainerDied","Data":"a71a498283bebdd6c4a1ceef0de2da16bd7de011b3281ffc8d87d7673b63cfee"} Oct 05 21:05:18 crc kubenswrapper[4754]: I1005 21:05:18.200254 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cpbjks" Oct 05 21:05:18 crc kubenswrapper[4754]: I1005 21:05:18.291734 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f7edcdad-0804-4ae8-bb6d-1118b6d0f620-bundle\") pod \"f7edcdad-0804-4ae8-bb6d-1118b6d0f620\" (UID: \"f7edcdad-0804-4ae8-bb6d-1118b6d0f620\") " Oct 05 21:05:18 crc kubenswrapper[4754]: I1005 21:05:18.291854 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f7edcdad-0804-4ae8-bb6d-1118b6d0f620-util\") pod \"f7edcdad-0804-4ae8-bb6d-1118b6d0f620\" (UID: \"f7edcdad-0804-4ae8-bb6d-1118b6d0f620\") " Oct 05 21:05:18 crc kubenswrapper[4754]: I1005 21:05:18.292069 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-znl6n\" (UniqueName: \"kubernetes.io/projected/f7edcdad-0804-4ae8-bb6d-1118b6d0f620-kube-api-access-znl6n\") pod \"f7edcdad-0804-4ae8-bb6d-1118b6d0f620\" (UID: \"f7edcdad-0804-4ae8-bb6d-1118b6d0f620\") " Oct 05 21:05:18 crc kubenswrapper[4754]: I1005 21:05:18.297094 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f7edcdad-0804-4ae8-bb6d-1118b6d0f620-bundle" (OuterVolumeSpecName: "bundle") pod "f7edcdad-0804-4ae8-bb6d-1118b6d0f620" (UID: "f7edcdad-0804-4ae8-bb6d-1118b6d0f620"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 21:05:18 crc kubenswrapper[4754]: I1005 21:05:18.301406 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f7edcdad-0804-4ae8-bb6d-1118b6d0f620-kube-api-access-znl6n" (OuterVolumeSpecName: "kube-api-access-znl6n") pod "f7edcdad-0804-4ae8-bb6d-1118b6d0f620" (UID: "f7edcdad-0804-4ae8-bb6d-1118b6d0f620"). InnerVolumeSpecName "kube-api-access-znl6n". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:05:18 crc kubenswrapper[4754]: I1005 21:05:18.316854 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f7edcdad-0804-4ae8-bb6d-1118b6d0f620-util" (OuterVolumeSpecName: "util") pod "f7edcdad-0804-4ae8-bb6d-1118b6d0f620" (UID: "f7edcdad-0804-4ae8-bb6d-1118b6d0f620"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 21:05:18 crc kubenswrapper[4754]: I1005 21:05:18.394484 4754 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f7edcdad-0804-4ae8-bb6d-1118b6d0f620-util\") on node \"crc\" DevicePath \"\"" Oct 05 21:05:18 crc kubenswrapper[4754]: I1005 21:05:18.394813 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-znl6n\" (UniqueName: \"kubernetes.io/projected/f7edcdad-0804-4ae8-bb6d-1118b6d0f620-kube-api-access-znl6n\") on node \"crc\" DevicePath \"\"" Oct 05 21:05:18 crc kubenswrapper[4754]: I1005 21:05:18.394847 4754 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f7edcdad-0804-4ae8-bb6d-1118b6d0f620-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 21:05:18 crc kubenswrapper[4754]: I1005 21:05:18.865592 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cpbjks" event={"ID":"f7edcdad-0804-4ae8-bb6d-1118b6d0f620","Type":"ContainerDied","Data":"de4db7a2e06f3c5a280293aa32c366f54edc5c3726c591b42ac7c812527eea8c"} Oct 05 21:05:18 crc kubenswrapper[4754]: I1005 21:05:18.865650 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="de4db7a2e06f3c5a280293aa32c366f54edc5c3726c591b42ac7c812527eea8c" Oct 05 21:05:18 crc kubenswrapper[4754]: I1005 21:05:18.865732 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cpbjks" Oct 05 21:05:19 crc kubenswrapper[4754]: I1005 21:05:19.972146 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-z2c5l"] Oct 05 21:05:19 crc kubenswrapper[4754]: E1005 21:05:19.973384 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7edcdad-0804-4ae8-bb6d-1118b6d0f620" containerName="pull" Oct 05 21:05:19 crc kubenswrapper[4754]: I1005 21:05:19.973470 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7edcdad-0804-4ae8-bb6d-1118b6d0f620" containerName="pull" Oct 05 21:05:19 crc kubenswrapper[4754]: E1005 21:05:19.973576 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7edcdad-0804-4ae8-bb6d-1118b6d0f620" containerName="extract" Oct 05 21:05:19 crc kubenswrapper[4754]: I1005 21:05:19.973642 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7edcdad-0804-4ae8-bb6d-1118b6d0f620" containerName="extract" Oct 05 21:05:19 crc kubenswrapper[4754]: E1005 21:05:19.973732 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7edcdad-0804-4ae8-bb6d-1118b6d0f620" containerName="util" Oct 05 21:05:19 crc kubenswrapper[4754]: I1005 21:05:19.973796 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7edcdad-0804-4ae8-bb6d-1118b6d0f620" containerName="util" Oct 05 21:05:19 crc kubenswrapper[4754]: I1005 21:05:19.973997 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7edcdad-0804-4ae8-bb6d-1118b6d0f620" containerName="extract" Oct 05 21:05:19 crc kubenswrapper[4754]: I1005 21:05:19.974536 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-858ddd8f98-z2c5l" Oct 05 21:05:19 crc kubenswrapper[4754]: I1005 21:05:19.977318 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Oct 05 21:05:19 crc kubenswrapper[4754]: I1005 21:05:19.978259 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-5d55t" Oct 05 21:05:19 crc kubenswrapper[4754]: I1005 21:05:19.978819 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Oct 05 21:05:19 crc kubenswrapper[4754]: I1005 21:05:19.990046 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-z2c5l"] Oct 05 21:05:20 crc kubenswrapper[4754]: I1005 21:05:20.015950 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mnkwr\" (UniqueName: \"kubernetes.io/projected/262b4fe5-27b8-41fe-8d8c-aed9ca555a6e-kube-api-access-mnkwr\") pod \"nmstate-operator-858ddd8f98-z2c5l\" (UID: \"262b4fe5-27b8-41fe-8d8c-aed9ca555a6e\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-z2c5l" Oct 05 21:05:20 crc kubenswrapper[4754]: I1005 21:05:20.117873 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mnkwr\" (UniqueName: \"kubernetes.io/projected/262b4fe5-27b8-41fe-8d8c-aed9ca555a6e-kube-api-access-mnkwr\") pod \"nmstate-operator-858ddd8f98-z2c5l\" (UID: \"262b4fe5-27b8-41fe-8d8c-aed9ca555a6e\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-z2c5l" Oct 05 21:05:20 crc kubenswrapper[4754]: I1005 21:05:20.144818 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mnkwr\" (UniqueName: \"kubernetes.io/projected/262b4fe5-27b8-41fe-8d8c-aed9ca555a6e-kube-api-access-mnkwr\") pod \"nmstate-operator-858ddd8f98-z2c5l\" (UID: \"262b4fe5-27b8-41fe-8d8c-aed9ca555a6e\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-z2c5l" Oct 05 21:05:20 crc kubenswrapper[4754]: I1005 21:05:20.292944 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-858ddd8f98-z2c5l" Oct 05 21:05:20 crc kubenswrapper[4754]: I1005 21:05:20.567247 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-z2c5l"] Oct 05 21:05:20 crc kubenswrapper[4754]: I1005 21:05:20.875882 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-858ddd8f98-z2c5l" event={"ID":"262b4fe5-27b8-41fe-8d8c-aed9ca555a6e","Type":"ContainerStarted","Data":"d97d716cf11af9663fd7255663f813fbd9b224acc675f862c544e66552ecb2ea"} Oct 05 21:05:23 crc kubenswrapper[4754]: I1005 21:05:23.899789 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-858ddd8f98-z2c5l" event={"ID":"262b4fe5-27b8-41fe-8d8c-aed9ca555a6e","Type":"ContainerStarted","Data":"3523e54e649b6ad69a5d9f76a49224a8c10299ecbf1d2ef9380ece720b25d7ac"} Oct 05 21:05:23 crc kubenswrapper[4754]: I1005 21:05:23.927179 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-858ddd8f98-z2c5l" podStartSLOduration=2.245349992 podStartE2EDuration="4.927151644s" podCreationTimestamp="2025-10-05 21:05:19 +0000 UTC" firstStartedPulling="2025-10-05 21:05:20.595327542 +0000 UTC m=+644.499446252" lastFinishedPulling="2025-10-05 21:05:23.277129154 +0000 UTC m=+647.181247904" observedRunningTime="2025-10-05 21:05:23.924025357 +0000 UTC m=+647.828144107" watchObservedRunningTime="2025-10-05 21:05:23.927151644 +0000 UTC m=+647.831270394" Oct 05 21:05:24 crc kubenswrapper[4754]: I1005 21:05:24.935840 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-7469b"] Oct 05 21:05:24 crc kubenswrapper[4754]: I1005 21:05:24.936960 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-7469b" Oct 05 21:05:24 crc kubenswrapper[4754]: I1005 21:05:24.938964 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-72js5" Oct 05 21:05:24 crc kubenswrapper[4754]: I1005 21:05:24.971358 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-7469b"] Oct 05 21:05:24 crc kubenswrapper[4754]: I1005 21:05:24.980710 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-nds52"] Oct 05 21:05:24 crc kubenswrapper[4754]: I1005 21:05:24.981559 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-nds52" Oct 05 21:05:24 crc kubenswrapper[4754]: I1005 21:05:24.982947 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Oct 05 21:05:24 crc kubenswrapper[4754]: I1005 21:05:24.990373 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-nds52"] Oct 05 21:05:25 crc kubenswrapper[4754]: I1005 21:05:25.000485 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pg6qs\" (UniqueName: \"kubernetes.io/projected/b381c1ba-7339-4ab8-bc4a-d3887197d1ea-kube-api-access-pg6qs\") pod \"nmstate-metrics-fdff9cb8d-7469b\" (UID: \"b381c1ba-7339-4ab8-bc4a-d3887197d1ea\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-7469b" Oct 05 21:05:25 crc kubenswrapper[4754]: I1005 21:05:25.000558 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/30f05ad2-a252-4b9c-890d-892b90aaa568-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-nds52\" (UID: \"30f05ad2-a252-4b9c-890d-892b90aaa568\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-nds52" Oct 05 21:05:25 crc kubenswrapper[4754]: I1005 21:05:25.000682 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lmlt2\" (UniqueName: \"kubernetes.io/projected/30f05ad2-a252-4b9c-890d-892b90aaa568-kube-api-access-lmlt2\") pod \"nmstate-webhook-6cdbc54649-nds52\" (UID: \"30f05ad2-a252-4b9c-890d-892b90aaa568\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-nds52" Oct 05 21:05:25 crc kubenswrapper[4754]: I1005 21:05:25.009847 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-d4qnm"] Oct 05 21:05:25 crc kubenswrapper[4754]: I1005 21:05:25.010735 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-d4qnm" Oct 05 21:05:25 crc kubenswrapper[4754]: I1005 21:05:25.102392 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m4cbn\" (UniqueName: \"kubernetes.io/projected/76cf164c-fcc4-49e8-866f-e4ec435a8044-kube-api-access-m4cbn\") pod \"nmstate-handler-d4qnm\" (UID: \"76cf164c-fcc4-49e8-866f-e4ec435a8044\") " pod="openshift-nmstate/nmstate-handler-d4qnm" Oct 05 21:05:25 crc kubenswrapper[4754]: I1005 21:05:25.102471 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/76cf164c-fcc4-49e8-866f-e4ec435a8044-dbus-socket\") pod \"nmstate-handler-d4qnm\" (UID: \"76cf164c-fcc4-49e8-866f-e4ec435a8044\") " pod="openshift-nmstate/nmstate-handler-d4qnm" Oct 05 21:05:25 crc kubenswrapper[4754]: I1005 21:05:25.102597 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/76cf164c-fcc4-49e8-866f-e4ec435a8044-nmstate-lock\") pod \"nmstate-handler-d4qnm\" (UID: \"76cf164c-fcc4-49e8-866f-e4ec435a8044\") " pod="openshift-nmstate/nmstate-handler-d4qnm" Oct 05 21:05:25 crc kubenswrapper[4754]: I1005 21:05:25.102813 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pg6qs\" (UniqueName: \"kubernetes.io/projected/b381c1ba-7339-4ab8-bc4a-d3887197d1ea-kube-api-access-pg6qs\") pod \"nmstate-metrics-fdff9cb8d-7469b\" (UID: \"b381c1ba-7339-4ab8-bc4a-d3887197d1ea\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-7469b" Oct 05 21:05:25 crc kubenswrapper[4754]: I1005 21:05:25.102858 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/30f05ad2-a252-4b9c-890d-892b90aaa568-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-nds52\" (UID: \"30f05ad2-a252-4b9c-890d-892b90aaa568\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-nds52" Oct 05 21:05:25 crc kubenswrapper[4754]: E1005 21:05:25.103062 4754 secret.go:188] Couldn't get secret openshift-nmstate/openshift-nmstate-webhook: secret "openshift-nmstate-webhook" not found Oct 05 21:05:25 crc kubenswrapper[4754]: I1005 21:05:25.103080 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lmlt2\" (UniqueName: \"kubernetes.io/projected/30f05ad2-a252-4b9c-890d-892b90aaa568-kube-api-access-lmlt2\") pod \"nmstate-webhook-6cdbc54649-nds52\" (UID: \"30f05ad2-a252-4b9c-890d-892b90aaa568\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-nds52" Oct 05 21:05:25 crc kubenswrapper[4754]: I1005 21:05:25.103108 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/76cf164c-fcc4-49e8-866f-e4ec435a8044-ovs-socket\") pod \"nmstate-handler-d4qnm\" (UID: \"76cf164c-fcc4-49e8-866f-e4ec435a8044\") " pod="openshift-nmstate/nmstate-handler-d4qnm" Oct 05 21:05:25 crc kubenswrapper[4754]: E1005 21:05:25.103138 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/30f05ad2-a252-4b9c-890d-892b90aaa568-tls-key-pair podName:30f05ad2-a252-4b9c-890d-892b90aaa568 nodeName:}" failed. No retries permitted until 2025-10-05 21:05:25.603113044 +0000 UTC m=+649.507231754 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-key-pair" (UniqueName: "kubernetes.io/secret/30f05ad2-a252-4b9c-890d-892b90aaa568-tls-key-pair") pod "nmstate-webhook-6cdbc54649-nds52" (UID: "30f05ad2-a252-4b9c-890d-892b90aaa568") : secret "openshift-nmstate-webhook" not found Oct 05 21:05:25 crc kubenswrapper[4754]: I1005 21:05:25.110461 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-pvfm8"] Oct 05 21:05:25 crc kubenswrapper[4754]: I1005 21:05:25.111521 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-pvfm8" Oct 05 21:05:25 crc kubenswrapper[4754]: I1005 21:05:25.113940 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Oct 05 21:05:25 crc kubenswrapper[4754]: I1005 21:05:25.114113 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-fz88v" Oct 05 21:05:25 crc kubenswrapper[4754]: I1005 21:05:25.114171 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Oct 05 21:05:25 crc kubenswrapper[4754]: I1005 21:05:25.129540 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-pvfm8"] Oct 05 21:05:25 crc kubenswrapper[4754]: I1005 21:05:25.139034 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lmlt2\" (UniqueName: \"kubernetes.io/projected/30f05ad2-a252-4b9c-890d-892b90aaa568-kube-api-access-lmlt2\") pod \"nmstate-webhook-6cdbc54649-nds52\" (UID: \"30f05ad2-a252-4b9c-890d-892b90aaa568\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-nds52" Oct 05 21:05:25 crc kubenswrapper[4754]: I1005 21:05:25.141514 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pg6qs\" (UniqueName: \"kubernetes.io/projected/b381c1ba-7339-4ab8-bc4a-d3887197d1ea-kube-api-access-pg6qs\") pod \"nmstate-metrics-fdff9cb8d-7469b\" (UID: \"b381c1ba-7339-4ab8-bc4a-d3887197d1ea\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-7469b" Oct 05 21:05:25 crc kubenswrapper[4754]: I1005 21:05:25.205277 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m4cbn\" (UniqueName: \"kubernetes.io/projected/76cf164c-fcc4-49e8-866f-e4ec435a8044-kube-api-access-m4cbn\") pod \"nmstate-handler-d4qnm\" (UID: \"76cf164c-fcc4-49e8-866f-e4ec435a8044\") " pod="openshift-nmstate/nmstate-handler-d4qnm" Oct 05 21:05:25 crc kubenswrapper[4754]: I1005 21:05:25.205328 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/76cf164c-fcc4-49e8-866f-e4ec435a8044-dbus-socket\") pod \"nmstate-handler-d4qnm\" (UID: \"76cf164c-fcc4-49e8-866f-e4ec435a8044\") " pod="openshift-nmstate/nmstate-handler-d4qnm" Oct 05 21:05:25 crc kubenswrapper[4754]: I1005 21:05:25.205347 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/76cf164c-fcc4-49e8-866f-e4ec435a8044-nmstate-lock\") pod \"nmstate-handler-d4qnm\" (UID: \"76cf164c-fcc4-49e8-866f-e4ec435a8044\") " pod="openshift-nmstate/nmstate-handler-d4qnm" Oct 05 21:05:25 crc kubenswrapper[4754]: I1005 21:05:25.205381 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/785a344e-fb90-43ec-a951-962783b9ccbf-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-pvfm8\" (UID: \"785a344e-fb90-43ec-a951-962783b9ccbf\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-pvfm8" Oct 05 21:05:25 crc kubenswrapper[4754]: I1005 21:05:25.205403 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wj7xv\" (UniqueName: \"kubernetes.io/projected/785a344e-fb90-43ec-a951-962783b9ccbf-kube-api-access-wj7xv\") pod \"nmstate-console-plugin-6b874cbd85-pvfm8\" (UID: \"785a344e-fb90-43ec-a951-962783b9ccbf\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-pvfm8" Oct 05 21:05:25 crc kubenswrapper[4754]: I1005 21:05:25.205478 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/76cf164c-fcc4-49e8-866f-e4ec435a8044-ovs-socket\") pod \"nmstate-handler-d4qnm\" (UID: \"76cf164c-fcc4-49e8-866f-e4ec435a8044\") " pod="openshift-nmstate/nmstate-handler-d4qnm" Oct 05 21:05:25 crc kubenswrapper[4754]: I1005 21:05:25.206235 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/785a344e-fb90-43ec-a951-962783b9ccbf-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-pvfm8\" (UID: \"785a344e-fb90-43ec-a951-962783b9ccbf\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-pvfm8" Oct 05 21:05:25 crc kubenswrapper[4754]: I1005 21:05:25.206282 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/76cf164c-fcc4-49e8-866f-e4ec435a8044-nmstate-lock\") pod \"nmstate-handler-d4qnm\" (UID: \"76cf164c-fcc4-49e8-866f-e4ec435a8044\") " pod="openshift-nmstate/nmstate-handler-d4qnm" Oct 05 21:05:25 crc kubenswrapper[4754]: I1005 21:05:25.206413 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/76cf164c-fcc4-49e8-866f-e4ec435a8044-ovs-socket\") pod \"nmstate-handler-d4qnm\" (UID: \"76cf164c-fcc4-49e8-866f-e4ec435a8044\") " pod="openshift-nmstate/nmstate-handler-d4qnm" Oct 05 21:05:25 crc kubenswrapper[4754]: I1005 21:05:25.206624 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/76cf164c-fcc4-49e8-866f-e4ec435a8044-dbus-socket\") pod \"nmstate-handler-d4qnm\" (UID: \"76cf164c-fcc4-49e8-866f-e4ec435a8044\") " pod="openshift-nmstate/nmstate-handler-d4qnm" Oct 05 21:05:25 crc kubenswrapper[4754]: I1005 21:05:25.237196 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m4cbn\" (UniqueName: \"kubernetes.io/projected/76cf164c-fcc4-49e8-866f-e4ec435a8044-kube-api-access-m4cbn\") pod \"nmstate-handler-d4qnm\" (UID: \"76cf164c-fcc4-49e8-866f-e4ec435a8044\") " pod="openshift-nmstate/nmstate-handler-d4qnm" Oct 05 21:05:25 crc kubenswrapper[4754]: I1005 21:05:25.251364 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-7469b" Oct 05 21:05:25 crc kubenswrapper[4754]: I1005 21:05:25.307648 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/785a344e-fb90-43ec-a951-962783b9ccbf-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-pvfm8\" (UID: \"785a344e-fb90-43ec-a951-962783b9ccbf\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-pvfm8" Oct 05 21:05:25 crc kubenswrapper[4754]: I1005 21:05:25.307746 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/785a344e-fb90-43ec-a951-962783b9ccbf-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-pvfm8\" (UID: \"785a344e-fb90-43ec-a951-962783b9ccbf\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-pvfm8" Oct 05 21:05:25 crc kubenswrapper[4754]: I1005 21:05:25.307788 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wj7xv\" (UniqueName: \"kubernetes.io/projected/785a344e-fb90-43ec-a951-962783b9ccbf-kube-api-access-wj7xv\") pod \"nmstate-console-plugin-6b874cbd85-pvfm8\" (UID: \"785a344e-fb90-43ec-a951-962783b9ccbf\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-pvfm8" Oct 05 21:05:25 crc kubenswrapper[4754]: E1005 21:05:25.308643 4754 secret.go:188] Couldn't get secret openshift-nmstate/plugin-serving-cert: secret "plugin-serving-cert" not found Oct 05 21:05:25 crc kubenswrapper[4754]: E1005 21:05:25.308735 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/785a344e-fb90-43ec-a951-962783b9ccbf-plugin-serving-cert podName:785a344e-fb90-43ec-a951-962783b9ccbf nodeName:}" failed. No retries permitted until 2025-10-05 21:05:25.808713324 +0000 UTC m=+649.712832034 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "plugin-serving-cert" (UniqueName: "kubernetes.io/secret/785a344e-fb90-43ec-a951-962783b9ccbf-plugin-serving-cert") pod "nmstate-console-plugin-6b874cbd85-pvfm8" (UID: "785a344e-fb90-43ec-a951-962783b9ccbf") : secret "plugin-serving-cert" not found Oct 05 21:05:25 crc kubenswrapper[4754]: I1005 21:05:25.309231 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/785a344e-fb90-43ec-a951-962783b9ccbf-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-pvfm8\" (UID: \"785a344e-fb90-43ec-a951-962783b9ccbf\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-pvfm8" Oct 05 21:05:25 crc kubenswrapper[4754]: I1005 21:05:25.326980 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-d4qnm" Oct 05 21:05:25 crc kubenswrapper[4754]: I1005 21:05:25.327303 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wj7xv\" (UniqueName: \"kubernetes.io/projected/785a344e-fb90-43ec-a951-962783b9ccbf-kube-api-access-wj7xv\") pod \"nmstate-console-plugin-6b874cbd85-pvfm8\" (UID: \"785a344e-fb90-43ec-a951-962783b9ccbf\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-pvfm8" Oct 05 21:05:25 crc kubenswrapper[4754]: W1005 21:05:25.394689 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod76cf164c_fcc4_49e8_866f_e4ec435a8044.slice/crio-c37f4739ae74cdbd6fc16c2afcaf906cc77acab6246943f8b181d49e9189c495 WatchSource:0}: Error finding container c37f4739ae74cdbd6fc16c2afcaf906cc77acab6246943f8b181d49e9189c495: Status 404 returned error can't find the container with id c37f4739ae74cdbd6fc16c2afcaf906cc77acab6246943f8b181d49e9189c495 Oct 05 21:05:25 crc kubenswrapper[4754]: I1005 21:05:25.395961 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-5bc466db5b-mnktt"] Oct 05 21:05:25 crc kubenswrapper[4754]: I1005 21:05:25.397247 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-5bc466db5b-mnktt" Oct 05 21:05:25 crc kubenswrapper[4754]: I1005 21:05:25.408680 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/0725ce57-f602-43b5-a156-e408a99f6b91-oauth-serving-cert\") pod \"console-5bc466db5b-mnktt\" (UID: \"0725ce57-f602-43b5-a156-e408a99f6b91\") " pod="openshift-console/console-5bc466db5b-mnktt" Oct 05 21:05:25 crc kubenswrapper[4754]: I1005 21:05:25.408733 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0725ce57-f602-43b5-a156-e408a99f6b91-service-ca\") pod \"console-5bc466db5b-mnktt\" (UID: \"0725ce57-f602-43b5-a156-e408a99f6b91\") " pod="openshift-console/console-5bc466db5b-mnktt" Oct 05 21:05:25 crc kubenswrapper[4754]: I1005 21:05:25.408758 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0725ce57-f602-43b5-a156-e408a99f6b91-trusted-ca-bundle\") pod \"console-5bc466db5b-mnktt\" (UID: \"0725ce57-f602-43b5-a156-e408a99f6b91\") " pod="openshift-console/console-5bc466db5b-mnktt" Oct 05 21:05:25 crc kubenswrapper[4754]: I1005 21:05:25.408794 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/0725ce57-f602-43b5-a156-e408a99f6b91-console-oauth-config\") pod \"console-5bc466db5b-mnktt\" (UID: \"0725ce57-f602-43b5-a156-e408a99f6b91\") " pod="openshift-console/console-5bc466db5b-mnktt" Oct 05 21:05:25 crc kubenswrapper[4754]: I1005 21:05:25.408820 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/0725ce57-f602-43b5-a156-e408a99f6b91-console-config\") pod \"console-5bc466db5b-mnktt\" (UID: \"0725ce57-f602-43b5-a156-e408a99f6b91\") " pod="openshift-console/console-5bc466db5b-mnktt" Oct 05 21:05:25 crc kubenswrapper[4754]: I1005 21:05:25.408866 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-twjsf\" (UniqueName: \"kubernetes.io/projected/0725ce57-f602-43b5-a156-e408a99f6b91-kube-api-access-twjsf\") pod \"console-5bc466db5b-mnktt\" (UID: \"0725ce57-f602-43b5-a156-e408a99f6b91\") " pod="openshift-console/console-5bc466db5b-mnktt" Oct 05 21:05:25 crc kubenswrapper[4754]: I1005 21:05:25.408894 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/0725ce57-f602-43b5-a156-e408a99f6b91-console-serving-cert\") pod \"console-5bc466db5b-mnktt\" (UID: \"0725ce57-f602-43b5-a156-e408a99f6b91\") " pod="openshift-console/console-5bc466db5b-mnktt" Oct 05 21:05:25 crc kubenswrapper[4754]: I1005 21:05:25.428916 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-5bc466db5b-mnktt"] Oct 05 21:05:25 crc kubenswrapper[4754]: I1005 21:05:25.517463 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-twjsf\" (UniqueName: \"kubernetes.io/projected/0725ce57-f602-43b5-a156-e408a99f6b91-kube-api-access-twjsf\") pod \"console-5bc466db5b-mnktt\" (UID: \"0725ce57-f602-43b5-a156-e408a99f6b91\") " pod="openshift-console/console-5bc466db5b-mnktt" Oct 05 21:05:25 crc kubenswrapper[4754]: I1005 21:05:25.519763 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/0725ce57-f602-43b5-a156-e408a99f6b91-console-serving-cert\") pod \"console-5bc466db5b-mnktt\" (UID: \"0725ce57-f602-43b5-a156-e408a99f6b91\") " pod="openshift-console/console-5bc466db5b-mnktt" Oct 05 21:05:25 crc kubenswrapper[4754]: I1005 21:05:25.519900 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/0725ce57-f602-43b5-a156-e408a99f6b91-oauth-serving-cert\") pod \"console-5bc466db5b-mnktt\" (UID: \"0725ce57-f602-43b5-a156-e408a99f6b91\") " pod="openshift-console/console-5bc466db5b-mnktt" Oct 05 21:05:25 crc kubenswrapper[4754]: I1005 21:05:25.519990 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0725ce57-f602-43b5-a156-e408a99f6b91-service-ca\") pod \"console-5bc466db5b-mnktt\" (UID: \"0725ce57-f602-43b5-a156-e408a99f6b91\") " pod="openshift-console/console-5bc466db5b-mnktt" Oct 05 21:05:25 crc kubenswrapper[4754]: I1005 21:05:25.520025 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0725ce57-f602-43b5-a156-e408a99f6b91-trusted-ca-bundle\") pod \"console-5bc466db5b-mnktt\" (UID: \"0725ce57-f602-43b5-a156-e408a99f6b91\") " pod="openshift-console/console-5bc466db5b-mnktt" Oct 05 21:05:25 crc kubenswrapper[4754]: I1005 21:05:25.520095 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/0725ce57-f602-43b5-a156-e408a99f6b91-console-oauth-config\") pod \"console-5bc466db5b-mnktt\" (UID: \"0725ce57-f602-43b5-a156-e408a99f6b91\") " pod="openshift-console/console-5bc466db5b-mnktt" Oct 05 21:05:25 crc kubenswrapper[4754]: I1005 21:05:25.520165 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/0725ce57-f602-43b5-a156-e408a99f6b91-console-config\") pod \"console-5bc466db5b-mnktt\" (UID: \"0725ce57-f602-43b5-a156-e408a99f6b91\") " pod="openshift-console/console-5bc466db5b-mnktt" Oct 05 21:05:25 crc kubenswrapper[4754]: I1005 21:05:25.521849 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/0725ce57-f602-43b5-a156-e408a99f6b91-console-config\") pod \"console-5bc466db5b-mnktt\" (UID: \"0725ce57-f602-43b5-a156-e408a99f6b91\") " pod="openshift-console/console-5bc466db5b-mnktt" Oct 05 21:05:25 crc kubenswrapper[4754]: I1005 21:05:25.523114 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0725ce57-f602-43b5-a156-e408a99f6b91-trusted-ca-bundle\") pod \"console-5bc466db5b-mnktt\" (UID: \"0725ce57-f602-43b5-a156-e408a99f6b91\") " pod="openshift-console/console-5bc466db5b-mnktt" Oct 05 21:05:25 crc kubenswrapper[4754]: I1005 21:05:25.523390 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/0725ce57-f602-43b5-a156-e408a99f6b91-oauth-serving-cert\") pod \"console-5bc466db5b-mnktt\" (UID: \"0725ce57-f602-43b5-a156-e408a99f6b91\") " pod="openshift-console/console-5bc466db5b-mnktt" Oct 05 21:05:25 crc kubenswrapper[4754]: I1005 21:05:25.524062 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0725ce57-f602-43b5-a156-e408a99f6b91-service-ca\") pod \"console-5bc466db5b-mnktt\" (UID: \"0725ce57-f602-43b5-a156-e408a99f6b91\") " pod="openshift-console/console-5bc466db5b-mnktt" Oct 05 21:05:25 crc kubenswrapper[4754]: I1005 21:05:25.533442 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/0725ce57-f602-43b5-a156-e408a99f6b91-console-oauth-config\") pod \"console-5bc466db5b-mnktt\" (UID: \"0725ce57-f602-43b5-a156-e408a99f6b91\") " pod="openshift-console/console-5bc466db5b-mnktt" Oct 05 21:05:25 crc kubenswrapper[4754]: I1005 21:05:25.538264 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/0725ce57-f602-43b5-a156-e408a99f6b91-console-serving-cert\") pod \"console-5bc466db5b-mnktt\" (UID: \"0725ce57-f602-43b5-a156-e408a99f6b91\") " pod="openshift-console/console-5bc466db5b-mnktt" Oct 05 21:05:25 crc kubenswrapper[4754]: I1005 21:05:25.542714 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-twjsf\" (UniqueName: \"kubernetes.io/projected/0725ce57-f602-43b5-a156-e408a99f6b91-kube-api-access-twjsf\") pod \"console-5bc466db5b-mnktt\" (UID: \"0725ce57-f602-43b5-a156-e408a99f6b91\") " pod="openshift-console/console-5bc466db5b-mnktt" Oct 05 21:05:25 crc kubenswrapper[4754]: I1005 21:05:25.621627 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/30f05ad2-a252-4b9c-890d-892b90aaa568-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-nds52\" (UID: \"30f05ad2-a252-4b9c-890d-892b90aaa568\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-nds52" Oct 05 21:05:25 crc kubenswrapper[4754]: I1005 21:05:25.625364 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/30f05ad2-a252-4b9c-890d-892b90aaa568-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-nds52\" (UID: \"30f05ad2-a252-4b9c-890d-892b90aaa568\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-nds52" Oct 05 21:05:25 crc kubenswrapper[4754]: I1005 21:05:25.694292 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-7469b"] Oct 05 21:05:25 crc kubenswrapper[4754]: W1005 21:05:25.700072 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb381c1ba_7339_4ab8_bc4a_d3887197d1ea.slice/crio-178bb7ed0a86411296d944517c103f6647514ed70fe8e57088136564ca6eadf8 WatchSource:0}: Error finding container 178bb7ed0a86411296d944517c103f6647514ed70fe8e57088136564ca6eadf8: Status 404 returned error can't find the container with id 178bb7ed0a86411296d944517c103f6647514ed70fe8e57088136564ca6eadf8 Oct 05 21:05:25 crc kubenswrapper[4754]: I1005 21:05:25.721625 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-5bc466db5b-mnktt" Oct 05 21:05:25 crc kubenswrapper[4754]: I1005 21:05:25.824615 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/785a344e-fb90-43ec-a951-962783b9ccbf-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-pvfm8\" (UID: \"785a344e-fb90-43ec-a951-962783b9ccbf\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-pvfm8" Oct 05 21:05:25 crc kubenswrapper[4754]: I1005 21:05:25.830122 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/785a344e-fb90-43ec-a951-962783b9ccbf-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-pvfm8\" (UID: \"785a344e-fb90-43ec-a951-962783b9ccbf\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-pvfm8" Oct 05 21:05:25 crc kubenswrapper[4754]: I1005 21:05:25.893749 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-nds52" Oct 05 21:05:25 crc kubenswrapper[4754]: I1005 21:05:25.910269 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-7469b" event={"ID":"b381c1ba-7339-4ab8-bc4a-d3887197d1ea","Type":"ContainerStarted","Data":"178bb7ed0a86411296d944517c103f6647514ed70fe8e57088136564ca6eadf8"} Oct 05 21:05:25 crc kubenswrapper[4754]: I1005 21:05:25.911292 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-d4qnm" event={"ID":"76cf164c-fcc4-49e8-866f-e4ec435a8044","Type":"ContainerStarted","Data":"c37f4739ae74cdbd6fc16c2afcaf906cc77acab6246943f8b181d49e9189c495"} Oct 05 21:05:25 crc kubenswrapper[4754]: I1005 21:05:25.941732 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-5bc466db5b-mnktt"] Oct 05 21:05:26 crc kubenswrapper[4754]: I1005 21:05:26.030013 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-pvfm8" Oct 05 21:05:26 crc kubenswrapper[4754]: I1005 21:05:26.110724 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-nds52"] Oct 05 21:05:26 crc kubenswrapper[4754]: I1005 21:05:26.334180 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-pvfm8"] Oct 05 21:05:26 crc kubenswrapper[4754]: I1005 21:05:26.922046 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-5bc466db5b-mnktt" event={"ID":"0725ce57-f602-43b5-a156-e408a99f6b91","Type":"ContainerStarted","Data":"9a38342a045bd5e9eeffcd44f059a06a7337a30ff10bf2fa6af31e97bb8f5954"} Oct 05 21:05:26 crc kubenswrapper[4754]: I1005 21:05:26.922415 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-5bc466db5b-mnktt" event={"ID":"0725ce57-f602-43b5-a156-e408a99f6b91","Type":"ContainerStarted","Data":"461c7a3e4135458804e481a8da02baf63aa9974172b5bb947c5522de3021f917"} Oct 05 21:05:26 crc kubenswrapper[4754]: I1005 21:05:26.924001 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-pvfm8" event={"ID":"785a344e-fb90-43ec-a951-962783b9ccbf","Type":"ContainerStarted","Data":"051cef9196921b19c344bb992f32968eccc0a2fce2a50665297fe901effb9f36"} Oct 05 21:05:26 crc kubenswrapper[4754]: I1005 21:05:26.925545 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-nds52" event={"ID":"30f05ad2-a252-4b9c-890d-892b90aaa568","Type":"ContainerStarted","Data":"e3647863f43f4240b44b2a576adb16cf1567ac8ab755bb199f204101fa1dfaba"} Oct 05 21:05:26 crc kubenswrapper[4754]: I1005 21:05:26.955950 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-5bc466db5b-mnktt" podStartSLOduration=1.955928788 podStartE2EDuration="1.955928788s" podCreationTimestamp="2025-10-05 21:05:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 21:05:26.953382103 +0000 UTC m=+650.857500823" watchObservedRunningTime="2025-10-05 21:05:26.955928788 +0000 UTC m=+650.860047508" Oct 05 21:05:28 crc kubenswrapper[4754]: I1005 21:05:28.944575 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-7469b" event={"ID":"b381c1ba-7339-4ab8-bc4a-d3887197d1ea","Type":"ContainerStarted","Data":"e7dc03527f6f31736221f4cf6f5e0c2888252e3f41a9f74c7fd65c76758d6191"} Oct 05 21:05:29 crc kubenswrapper[4754]: I1005 21:05:29.955006 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-nds52" event={"ID":"30f05ad2-a252-4b9c-890d-892b90aaa568","Type":"ContainerStarted","Data":"d05796c7dcad56bc9688ea04fbc66a9a3c97dc6df0f80c8b07b583f3a0ed84ba"} Oct 05 21:05:29 crc kubenswrapper[4754]: I1005 21:05:29.955400 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-nds52" Oct 05 21:05:29 crc kubenswrapper[4754]: I1005 21:05:29.957673 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-d4qnm" event={"ID":"76cf164c-fcc4-49e8-866f-e4ec435a8044","Type":"ContainerStarted","Data":"7d997e18bf30ba3ca35bb95cbf66cc27b9d5c5f8bb0aededbb56e55bde5fb8c1"} Oct 05 21:05:29 crc kubenswrapper[4754]: I1005 21:05:29.957945 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-d4qnm" Oct 05 21:05:29 crc kubenswrapper[4754]: I1005 21:05:29.978955 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-nds52" podStartSLOduration=3.312052941 podStartE2EDuration="5.978934197s" podCreationTimestamp="2025-10-05 21:05:24 +0000 UTC" firstStartedPulling="2025-10-05 21:05:26.135880843 +0000 UTC m=+650.039999553" lastFinishedPulling="2025-10-05 21:05:28.802762079 +0000 UTC m=+652.706880809" observedRunningTime="2025-10-05 21:05:29.978464225 +0000 UTC m=+653.882582945" watchObservedRunningTime="2025-10-05 21:05:29.978934197 +0000 UTC m=+653.883052917" Oct 05 21:05:29 crc kubenswrapper[4754]: I1005 21:05:29.999768 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-d4qnm" podStartSLOduration=2.702202795 podStartE2EDuration="5.999752946s" podCreationTimestamp="2025-10-05 21:05:24 +0000 UTC" firstStartedPulling="2025-10-05 21:05:25.401787392 +0000 UTC m=+649.305906102" lastFinishedPulling="2025-10-05 21:05:28.699337533 +0000 UTC m=+652.603456253" observedRunningTime="2025-10-05 21:05:29.996739198 +0000 UTC m=+653.900857908" watchObservedRunningTime="2025-10-05 21:05:29.999752946 +0000 UTC m=+653.903871656" Oct 05 21:05:30 crc kubenswrapper[4754]: I1005 21:05:30.965606 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-pvfm8" event={"ID":"785a344e-fb90-43ec-a951-962783b9ccbf","Type":"ContainerStarted","Data":"47d4aef496553ec7c36116f9acc76044825f5ba9107a371f8a112dfcb8fe15fd"} Oct 05 21:05:30 crc kubenswrapper[4754]: I1005 21:05:30.981387 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-pvfm8" podStartSLOduration=2.334041081 podStartE2EDuration="5.981367352s" podCreationTimestamp="2025-10-05 21:05:25 +0000 UTC" firstStartedPulling="2025-10-05 21:05:26.348943235 +0000 UTC m=+650.253061945" lastFinishedPulling="2025-10-05 21:05:29.996269496 +0000 UTC m=+653.900388216" observedRunningTime="2025-10-05 21:05:30.979979766 +0000 UTC m=+654.884098476" watchObservedRunningTime="2025-10-05 21:05:30.981367352 +0000 UTC m=+654.885486062" Oct 05 21:05:31 crc kubenswrapper[4754]: I1005 21:05:31.976314 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-7469b" event={"ID":"b381c1ba-7339-4ab8-bc4a-d3887197d1ea","Type":"ContainerStarted","Data":"cc0968c12710ef590dd47ea77f27bea7d69c966b39261b289b71f3adf35f12f0"} Oct 05 21:05:32 crc kubenswrapper[4754]: I1005 21:05:32.021647 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-7469b" podStartSLOduration=2.36963637 podStartE2EDuration="8.021605344s" podCreationTimestamp="2025-10-05 21:05:24 +0000 UTC" firstStartedPulling="2025-10-05 21:05:25.702698746 +0000 UTC m=+649.606817446" lastFinishedPulling="2025-10-05 21:05:31.35466771 +0000 UTC m=+655.258786420" observedRunningTime="2025-10-05 21:05:32.015290821 +0000 UTC m=+655.919409581" watchObservedRunningTime="2025-10-05 21:05:32.021605344 +0000 UTC m=+655.925724104" Oct 05 21:05:35 crc kubenswrapper[4754]: I1005 21:05:35.358356 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-d4qnm" Oct 05 21:05:35 crc kubenswrapper[4754]: I1005 21:05:35.721977 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-5bc466db5b-mnktt" Oct 05 21:05:35 crc kubenswrapper[4754]: I1005 21:05:35.722413 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-5bc466db5b-mnktt" Oct 05 21:05:35 crc kubenswrapper[4754]: I1005 21:05:35.728145 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-5bc466db5b-mnktt" Oct 05 21:05:36 crc kubenswrapper[4754]: I1005 21:05:36.015333 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-5bc466db5b-mnktt" Oct 05 21:05:36 crc kubenswrapper[4754]: I1005 21:05:36.103903 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-5sffm"] Oct 05 21:05:45 crc kubenswrapper[4754]: I1005 21:05:45.903016 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-nds52" Oct 05 21:06:01 crc kubenswrapper[4754]: I1005 21:06:01.149303 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-5sffm" podUID="c56f5eb8-3a12-428c-b347-78369dc19008" containerName="console" containerID="cri-o://e89d93557492b507b25bbe982c756e8821d7f1efc3a6df2fb3a35171cd924f24" gracePeriod=15 Oct 05 21:06:01 crc kubenswrapper[4754]: I1005 21:06:01.539564 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-5sffm_c56f5eb8-3a12-428c-b347-78369dc19008/console/0.log" Oct 05 21:06:01 crc kubenswrapper[4754]: I1005 21:06:01.539655 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-5sffm" Oct 05 21:06:01 crc kubenswrapper[4754]: I1005 21:06:01.686955 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c56f5eb8-3a12-428c-b347-78369dc19008-trusted-ca-bundle\") pod \"c56f5eb8-3a12-428c-b347-78369dc19008\" (UID: \"c56f5eb8-3a12-428c-b347-78369dc19008\") " Oct 05 21:06:01 crc kubenswrapper[4754]: I1005 21:06:01.687446 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/c56f5eb8-3a12-428c-b347-78369dc19008-oauth-serving-cert\") pod \"c56f5eb8-3a12-428c-b347-78369dc19008\" (UID: \"c56f5eb8-3a12-428c-b347-78369dc19008\") " Oct 05 21:06:01 crc kubenswrapper[4754]: I1005 21:06:01.687531 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/c56f5eb8-3a12-428c-b347-78369dc19008-console-serving-cert\") pod \"c56f5eb8-3a12-428c-b347-78369dc19008\" (UID: \"c56f5eb8-3a12-428c-b347-78369dc19008\") " Oct 05 21:06:01 crc kubenswrapper[4754]: I1005 21:06:01.687601 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/c56f5eb8-3a12-428c-b347-78369dc19008-console-config\") pod \"c56f5eb8-3a12-428c-b347-78369dc19008\" (UID: \"c56f5eb8-3a12-428c-b347-78369dc19008\") " Oct 05 21:06:01 crc kubenswrapper[4754]: I1005 21:06:01.687632 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m66bm\" (UniqueName: \"kubernetes.io/projected/c56f5eb8-3a12-428c-b347-78369dc19008-kube-api-access-m66bm\") pod \"c56f5eb8-3a12-428c-b347-78369dc19008\" (UID: \"c56f5eb8-3a12-428c-b347-78369dc19008\") " Oct 05 21:06:01 crc kubenswrapper[4754]: I1005 21:06:01.687685 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/c56f5eb8-3a12-428c-b347-78369dc19008-console-oauth-config\") pod \"c56f5eb8-3a12-428c-b347-78369dc19008\" (UID: \"c56f5eb8-3a12-428c-b347-78369dc19008\") " Oct 05 21:06:01 crc kubenswrapper[4754]: I1005 21:06:01.687707 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/c56f5eb8-3a12-428c-b347-78369dc19008-service-ca\") pod \"c56f5eb8-3a12-428c-b347-78369dc19008\" (UID: \"c56f5eb8-3a12-428c-b347-78369dc19008\") " Oct 05 21:06:01 crc kubenswrapper[4754]: I1005 21:06:01.688685 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c56f5eb8-3a12-428c-b347-78369dc19008-console-config" (OuterVolumeSpecName: "console-config") pod "c56f5eb8-3a12-428c-b347-78369dc19008" (UID: "c56f5eb8-3a12-428c-b347-78369dc19008"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 21:06:01 crc kubenswrapper[4754]: I1005 21:06:01.688768 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c56f5eb8-3a12-428c-b347-78369dc19008-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "c56f5eb8-3a12-428c-b347-78369dc19008" (UID: "c56f5eb8-3a12-428c-b347-78369dc19008"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 21:06:01 crc kubenswrapper[4754]: I1005 21:06:01.689302 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c56f5eb8-3a12-428c-b347-78369dc19008-service-ca" (OuterVolumeSpecName: "service-ca") pod "c56f5eb8-3a12-428c-b347-78369dc19008" (UID: "c56f5eb8-3a12-428c-b347-78369dc19008"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 21:06:01 crc kubenswrapper[4754]: I1005 21:06:01.690015 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c56f5eb8-3a12-428c-b347-78369dc19008-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "c56f5eb8-3a12-428c-b347-78369dc19008" (UID: "c56f5eb8-3a12-428c-b347-78369dc19008"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 21:06:01 crc kubenswrapper[4754]: I1005 21:06:01.697167 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c56f5eb8-3a12-428c-b347-78369dc19008-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "c56f5eb8-3a12-428c-b347-78369dc19008" (UID: "c56f5eb8-3a12-428c-b347-78369dc19008"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:06:01 crc kubenswrapper[4754]: I1005 21:06:01.698311 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c56f5eb8-3a12-428c-b347-78369dc19008-kube-api-access-m66bm" (OuterVolumeSpecName: "kube-api-access-m66bm") pod "c56f5eb8-3a12-428c-b347-78369dc19008" (UID: "c56f5eb8-3a12-428c-b347-78369dc19008"). InnerVolumeSpecName "kube-api-access-m66bm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:06:01 crc kubenswrapper[4754]: I1005 21:06:01.705956 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c56f5eb8-3a12-428c-b347-78369dc19008-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "c56f5eb8-3a12-428c-b347-78369dc19008" (UID: "c56f5eb8-3a12-428c-b347-78369dc19008"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:06:01 crc kubenswrapper[4754]: I1005 21:06:01.789234 4754 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/c56f5eb8-3a12-428c-b347-78369dc19008-console-config\") on node \"crc\" DevicePath \"\"" Oct 05 21:06:01 crc kubenswrapper[4754]: I1005 21:06:01.789265 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m66bm\" (UniqueName: \"kubernetes.io/projected/c56f5eb8-3a12-428c-b347-78369dc19008-kube-api-access-m66bm\") on node \"crc\" DevicePath \"\"" Oct 05 21:06:01 crc kubenswrapper[4754]: I1005 21:06:01.789278 4754 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/c56f5eb8-3a12-428c-b347-78369dc19008-console-oauth-config\") on node \"crc\" DevicePath \"\"" Oct 05 21:06:01 crc kubenswrapper[4754]: I1005 21:06:01.789287 4754 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/c56f5eb8-3a12-428c-b347-78369dc19008-service-ca\") on node \"crc\" DevicePath \"\"" Oct 05 21:06:01 crc kubenswrapper[4754]: I1005 21:06:01.789296 4754 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c56f5eb8-3a12-428c-b347-78369dc19008-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 21:06:01 crc kubenswrapper[4754]: I1005 21:06:01.789307 4754 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/c56f5eb8-3a12-428c-b347-78369dc19008-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 05 21:06:01 crc kubenswrapper[4754]: I1005 21:06:01.789317 4754 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/c56f5eb8-3a12-428c-b347-78369dc19008-console-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 05 21:06:02 crc kubenswrapper[4754]: I1005 21:06:02.189917 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-5sffm_c56f5eb8-3a12-428c-b347-78369dc19008/console/0.log" Oct 05 21:06:02 crc kubenswrapper[4754]: I1005 21:06:02.189965 4754 generic.go:334] "Generic (PLEG): container finished" podID="c56f5eb8-3a12-428c-b347-78369dc19008" containerID="e89d93557492b507b25bbe982c756e8821d7f1efc3a6df2fb3a35171cd924f24" exitCode=2 Oct 05 21:06:02 crc kubenswrapper[4754]: I1005 21:06:02.189993 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-5sffm" event={"ID":"c56f5eb8-3a12-428c-b347-78369dc19008","Type":"ContainerDied","Data":"e89d93557492b507b25bbe982c756e8821d7f1efc3a6df2fb3a35171cd924f24"} Oct 05 21:06:02 crc kubenswrapper[4754]: I1005 21:06:02.190030 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-5sffm" event={"ID":"c56f5eb8-3a12-428c-b347-78369dc19008","Type":"ContainerDied","Data":"1b5f2b6c650425bb321b1679f0ac14848e2a2b5706e1443dbdeda08a6b35615f"} Oct 05 21:06:02 crc kubenswrapper[4754]: I1005 21:06:02.190047 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-5sffm" Oct 05 21:06:02 crc kubenswrapper[4754]: I1005 21:06:02.190052 4754 scope.go:117] "RemoveContainer" containerID="e89d93557492b507b25bbe982c756e8821d7f1efc3a6df2fb3a35171cd924f24" Oct 05 21:06:02 crc kubenswrapper[4754]: I1005 21:06:02.208343 4754 scope.go:117] "RemoveContainer" containerID="e89d93557492b507b25bbe982c756e8821d7f1efc3a6df2fb3a35171cd924f24" Oct 05 21:06:02 crc kubenswrapper[4754]: E1005 21:06:02.208897 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e89d93557492b507b25bbe982c756e8821d7f1efc3a6df2fb3a35171cd924f24\": container with ID starting with e89d93557492b507b25bbe982c756e8821d7f1efc3a6df2fb3a35171cd924f24 not found: ID does not exist" containerID="e89d93557492b507b25bbe982c756e8821d7f1efc3a6df2fb3a35171cd924f24" Oct 05 21:06:02 crc kubenswrapper[4754]: I1005 21:06:02.208980 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e89d93557492b507b25bbe982c756e8821d7f1efc3a6df2fb3a35171cd924f24"} err="failed to get container status \"e89d93557492b507b25bbe982c756e8821d7f1efc3a6df2fb3a35171cd924f24\": rpc error: code = NotFound desc = could not find container \"e89d93557492b507b25bbe982c756e8821d7f1efc3a6df2fb3a35171cd924f24\": container with ID starting with e89d93557492b507b25bbe982c756e8821d7f1efc3a6df2fb3a35171cd924f24 not found: ID does not exist" Oct 05 21:06:02 crc kubenswrapper[4754]: I1005 21:06:02.231550 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-5sffm"] Oct 05 21:06:02 crc kubenswrapper[4754]: I1005 21:06:02.241300 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-5sffm"] Oct 05 21:06:02 crc kubenswrapper[4754]: I1005 21:06:02.849997 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c56f5eb8-3a12-428c-b347-78369dc19008" path="/var/lib/kubelet/pods/c56f5eb8-3a12-428c-b347-78369dc19008/volumes" Oct 05 21:06:02 crc kubenswrapper[4754]: I1005 21:06:02.918276 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2kzp4h"] Oct 05 21:06:02 crc kubenswrapper[4754]: E1005 21:06:02.918697 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c56f5eb8-3a12-428c-b347-78369dc19008" containerName="console" Oct 05 21:06:02 crc kubenswrapper[4754]: I1005 21:06:02.918720 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="c56f5eb8-3a12-428c-b347-78369dc19008" containerName="console" Oct 05 21:06:02 crc kubenswrapper[4754]: I1005 21:06:02.918909 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="c56f5eb8-3a12-428c-b347-78369dc19008" containerName="console" Oct 05 21:06:02 crc kubenswrapper[4754]: I1005 21:06:02.920275 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2kzp4h" Oct 05 21:06:02 crc kubenswrapper[4754]: I1005 21:06:02.925055 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 05 21:06:02 crc kubenswrapper[4754]: I1005 21:06:02.933569 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2kzp4h"] Oct 05 21:06:03 crc kubenswrapper[4754]: I1005 21:06:03.005898 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/3da3d4d5-ef4b-4dde-8b69-0d8e566703c1-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2kzp4h\" (UID: \"3da3d4d5-ef4b-4dde-8b69-0d8e566703c1\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2kzp4h" Oct 05 21:06:03 crc kubenswrapper[4754]: I1005 21:06:03.006299 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/3da3d4d5-ef4b-4dde-8b69-0d8e566703c1-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2kzp4h\" (UID: \"3da3d4d5-ef4b-4dde-8b69-0d8e566703c1\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2kzp4h" Oct 05 21:06:03 crc kubenswrapper[4754]: I1005 21:06:03.006528 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6zfkt\" (UniqueName: \"kubernetes.io/projected/3da3d4d5-ef4b-4dde-8b69-0d8e566703c1-kube-api-access-6zfkt\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2kzp4h\" (UID: \"3da3d4d5-ef4b-4dde-8b69-0d8e566703c1\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2kzp4h" Oct 05 21:06:03 crc kubenswrapper[4754]: I1005 21:06:03.109556 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6zfkt\" (UniqueName: \"kubernetes.io/projected/3da3d4d5-ef4b-4dde-8b69-0d8e566703c1-kube-api-access-6zfkt\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2kzp4h\" (UID: \"3da3d4d5-ef4b-4dde-8b69-0d8e566703c1\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2kzp4h" Oct 05 21:06:03 crc kubenswrapper[4754]: I1005 21:06:03.109711 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/3da3d4d5-ef4b-4dde-8b69-0d8e566703c1-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2kzp4h\" (UID: \"3da3d4d5-ef4b-4dde-8b69-0d8e566703c1\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2kzp4h" Oct 05 21:06:03 crc kubenswrapper[4754]: I1005 21:06:03.109922 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/3da3d4d5-ef4b-4dde-8b69-0d8e566703c1-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2kzp4h\" (UID: \"3da3d4d5-ef4b-4dde-8b69-0d8e566703c1\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2kzp4h" Oct 05 21:06:03 crc kubenswrapper[4754]: I1005 21:06:03.110609 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/3da3d4d5-ef4b-4dde-8b69-0d8e566703c1-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2kzp4h\" (UID: \"3da3d4d5-ef4b-4dde-8b69-0d8e566703c1\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2kzp4h" Oct 05 21:06:03 crc kubenswrapper[4754]: I1005 21:06:03.110844 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/3da3d4d5-ef4b-4dde-8b69-0d8e566703c1-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2kzp4h\" (UID: \"3da3d4d5-ef4b-4dde-8b69-0d8e566703c1\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2kzp4h" Oct 05 21:06:03 crc kubenswrapper[4754]: I1005 21:06:03.145398 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6zfkt\" (UniqueName: \"kubernetes.io/projected/3da3d4d5-ef4b-4dde-8b69-0d8e566703c1-kube-api-access-6zfkt\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2kzp4h\" (UID: \"3da3d4d5-ef4b-4dde-8b69-0d8e566703c1\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2kzp4h" Oct 05 21:06:03 crc kubenswrapper[4754]: I1005 21:06:03.244915 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2kzp4h" Oct 05 21:06:03 crc kubenswrapper[4754]: I1005 21:06:03.704869 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2kzp4h"] Oct 05 21:06:04 crc kubenswrapper[4754]: I1005 21:06:04.210015 4754 generic.go:334] "Generic (PLEG): container finished" podID="3da3d4d5-ef4b-4dde-8b69-0d8e566703c1" containerID="5ef244dee8ccacc4fb097cb71db7239d4a314725149bdb1c43e68fbcba4deeed" exitCode=0 Oct 05 21:06:04 crc kubenswrapper[4754]: I1005 21:06:04.210079 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2kzp4h" event={"ID":"3da3d4d5-ef4b-4dde-8b69-0d8e566703c1","Type":"ContainerDied","Data":"5ef244dee8ccacc4fb097cb71db7239d4a314725149bdb1c43e68fbcba4deeed"} Oct 05 21:06:04 crc kubenswrapper[4754]: I1005 21:06:04.210147 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2kzp4h" event={"ID":"3da3d4d5-ef4b-4dde-8b69-0d8e566703c1","Type":"ContainerStarted","Data":"d47279d9130f9815a1aac96c5b40ed51a97f0b85074d5999c36de1bac33b0bc7"} Oct 05 21:06:06 crc kubenswrapper[4754]: I1005 21:06:06.228269 4754 generic.go:334] "Generic (PLEG): container finished" podID="3da3d4d5-ef4b-4dde-8b69-0d8e566703c1" containerID="10006cf65db271ab82cd18f82a45ce4a66f45e3cf78bb61bb9aa3cbb7de09278" exitCode=0 Oct 05 21:06:06 crc kubenswrapper[4754]: I1005 21:06:06.228373 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2kzp4h" event={"ID":"3da3d4d5-ef4b-4dde-8b69-0d8e566703c1","Type":"ContainerDied","Data":"10006cf65db271ab82cd18f82a45ce4a66f45e3cf78bb61bb9aa3cbb7de09278"} Oct 05 21:06:07 crc kubenswrapper[4754]: I1005 21:06:07.235922 4754 generic.go:334] "Generic (PLEG): container finished" podID="3da3d4d5-ef4b-4dde-8b69-0d8e566703c1" containerID="685732f77c2ae441d0736fa274c99d8da6be4bd768cd1935af57bfc952b83c14" exitCode=0 Oct 05 21:06:07 crc kubenswrapper[4754]: I1005 21:06:07.235972 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2kzp4h" event={"ID":"3da3d4d5-ef4b-4dde-8b69-0d8e566703c1","Type":"ContainerDied","Data":"685732f77c2ae441d0736fa274c99d8da6be4bd768cd1935af57bfc952b83c14"} Oct 05 21:06:08 crc kubenswrapper[4754]: I1005 21:06:08.534013 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2kzp4h" Oct 05 21:06:08 crc kubenswrapper[4754]: I1005 21:06:08.701018 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/3da3d4d5-ef4b-4dde-8b69-0d8e566703c1-util\") pod \"3da3d4d5-ef4b-4dde-8b69-0d8e566703c1\" (UID: \"3da3d4d5-ef4b-4dde-8b69-0d8e566703c1\") " Oct 05 21:06:08 crc kubenswrapper[4754]: I1005 21:06:08.701074 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/3da3d4d5-ef4b-4dde-8b69-0d8e566703c1-bundle\") pod \"3da3d4d5-ef4b-4dde-8b69-0d8e566703c1\" (UID: \"3da3d4d5-ef4b-4dde-8b69-0d8e566703c1\") " Oct 05 21:06:08 crc kubenswrapper[4754]: I1005 21:06:08.701188 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6zfkt\" (UniqueName: \"kubernetes.io/projected/3da3d4d5-ef4b-4dde-8b69-0d8e566703c1-kube-api-access-6zfkt\") pod \"3da3d4d5-ef4b-4dde-8b69-0d8e566703c1\" (UID: \"3da3d4d5-ef4b-4dde-8b69-0d8e566703c1\") " Oct 05 21:06:08 crc kubenswrapper[4754]: I1005 21:06:08.702587 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3da3d4d5-ef4b-4dde-8b69-0d8e566703c1-bundle" (OuterVolumeSpecName: "bundle") pod "3da3d4d5-ef4b-4dde-8b69-0d8e566703c1" (UID: "3da3d4d5-ef4b-4dde-8b69-0d8e566703c1"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 21:06:08 crc kubenswrapper[4754]: I1005 21:06:08.702836 4754 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/3da3d4d5-ef4b-4dde-8b69-0d8e566703c1-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 21:06:08 crc kubenswrapper[4754]: I1005 21:06:08.715177 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3da3d4d5-ef4b-4dde-8b69-0d8e566703c1-kube-api-access-6zfkt" (OuterVolumeSpecName: "kube-api-access-6zfkt") pod "3da3d4d5-ef4b-4dde-8b69-0d8e566703c1" (UID: "3da3d4d5-ef4b-4dde-8b69-0d8e566703c1"). InnerVolumeSpecName "kube-api-access-6zfkt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:06:08 crc kubenswrapper[4754]: I1005 21:06:08.730404 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3da3d4d5-ef4b-4dde-8b69-0d8e566703c1-util" (OuterVolumeSpecName: "util") pod "3da3d4d5-ef4b-4dde-8b69-0d8e566703c1" (UID: "3da3d4d5-ef4b-4dde-8b69-0d8e566703c1"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 21:06:08 crc kubenswrapper[4754]: I1005 21:06:08.804346 4754 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/3da3d4d5-ef4b-4dde-8b69-0d8e566703c1-util\") on node \"crc\" DevicePath \"\"" Oct 05 21:06:08 crc kubenswrapper[4754]: I1005 21:06:08.804401 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6zfkt\" (UniqueName: \"kubernetes.io/projected/3da3d4d5-ef4b-4dde-8b69-0d8e566703c1-kube-api-access-6zfkt\") on node \"crc\" DevicePath \"\"" Oct 05 21:06:09 crc kubenswrapper[4754]: I1005 21:06:09.254003 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2kzp4h" event={"ID":"3da3d4d5-ef4b-4dde-8b69-0d8e566703c1","Type":"ContainerDied","Data":"d47279d9130f9815a1aac96c5b40ed51a97f0b85074d5999c36de1bac33b0bc7"} Oct 05 21:06:09 crc kubenswrapper[4754]: I1005 21:06:09.254060 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d47279d9130f9815a1aac96c5b40ed51a97f0b85074d5999c36de1bac33b0bc7" Oct 05 21:06:09 crc kubenswrapper[4754]: I1005 21:06:09.254114 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2kzp4h" Oct 05 21:06:17 crc kubenswrapper[4754]: I1005 21:06:17.895761 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-645fb4cd96-7qplt"] Oct 05 21:06:17 crc kubenswrapper[4754]: E1005 21:06:17.896407 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3da3d4d5-ef4b-4dde-8b69-0d8e566703c1" containerName="pull" Oct 05 21:06:17 crc kubenswrapper[4754]: I1005 21:06:17.896419 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="3da3d4d5-ef4b-4dde-8b69-0d8e566703c1" containerName="pull" Oct 05 21:06:17 crc kubenswrapper[4754]: E1005 21:06:17.896434 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3da3d4d5-ef4b-4dde-8b69-0d8e566703c1" containerName="util" Oct 05 21:06:17 crc kubenswrapper[4754]: I1005 21:06:17.896440 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="3da3d4d5-ef4b-4dde-8b69-0d8e566703c1" containerName="util" Oct 05 21:06:17 crc kubenswrapper[4754]: E1005 21:06:17.896449 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3da3d4d5-ef4b-4dde-8b69-0d8e566703c1" containerName="extract" Oct 05 21:06:17 crc kubenswrapper[4754]: I1005 21:06:17.896457 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="3da3d4d5-ef4b-4dde-8b69-0d8e566703c1" containerName="extract" Oct 05 21:06:17 crc kubenswrapper[4754]: I1005 21:06:17.896573 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="3da3d4d5-ef4b-4dde-8b69-0d8e566703c1" containerName="extract" Oct 05 21:06:17 crc kubenswrapper[4754]: I1005 21:06:17.896913 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-645fb4cd96-7qplt" Oct 05 21:06:17 crc kubenswrapper[4754]: I1005 21:06:17.901280 4754 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Oct 05 21:06:17 crc kubenswrapper[4754]: I1005 21:06:17.901364 4754 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Oct 05 21:06:17 crc kubenswrapper[4754]: I1005 21:06:17.901486 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Oct 05 21:06:17 crc kubenswrapper[4754]: I1005 21:06:17.901935 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Oct 05 21:06:17 crc kubenswrapper[4754]: I1005 21:06:17.903226 4754 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-9f98c" Oct 05 21:06:17 crc kubenswrapper[4754]: I1005 21:06:17.962786 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-645fb4cd96-7qplt"] Oct 05 21:06:18 crc kubenswrapper[4754]: I1005 21:06:18.030253 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-chfgj\" (UniqueName: \"kubernetes.io/projected/697d9fce-0466-4ba2-a532-214dfd58a20e-kube-api-access-chfgj\") pod \"metallb-operator-controller-manager-645fb4cd96-7qplt\" (UID: \"697d9fce-0466-4ba2-a532-214dfd58a20e\") " pod="metallb-system/metallb-operator-controller-manager-645fb4cd96-7qplt" Oct 05 21:06:18 crc kubenswrapper[4754]: I1005 21:06:18.030343 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/697d9fce-0466-4ba2-a532-214dfd58a20e-apiservice-cert\") pod \"metallb-operator-controller-manager-645fb4cd96-7qplt\" (UID: \"697d9fce-0466-4ba2-a532-214dfd58a20e\") " pod="metallb-system/metallb-operator-controller-manager-645fb4cd96-7qplt" Oct 05 21:06:18 crc kubenswrapper[4754]: I1005 21:06:18.030379 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/697d9fce-0466-4ba2-a532-214dfd58a20e-webhook-cert\") pod \"metallb-operator-controller-manager-645fb4cd96-7qplt\" (UID: \"697d9fce-0466-4ba2-a532-214dfd58a20e\") " pod="metallb-system/metallb-operator-controller-manager-645fb4cd96-7qplt" Oct 05 21:06:18 crc kubenswrapper[4754]: I1005 21:06:18.131844 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-chfgj\" (UniqueName: \"kubernetes.io/projected/697d9fce-0466-4ba2-a532-214dfd58a20e-kube-api-access-chfgj\") pod \"metallb-operator-controller-manager-645fb4cd96-7qplt\" (UID: \"697d9fce-0466-4ba2-a532-214dfd58a20e\") " pod="metallb-system/metallb-operator-controller-manager-645fb4cd96-7qplt" Oct 05 21:06:18 crc kubenswrapper[4754]: I1005 21:06:18.132257 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/697d9fce-0466-4ba2-a532-214dfd58a20e-apiservice-cert\") pod \"metallb-operator-controller-manager-645fb4cd96-7qplt\" (UID: \"697d9fce-0466-4ba2-a532-214dfd58a20e\") " pod="metallb-system/metallb-operator-controller-manager-645fb4cd96-7qplt" Oct 05 21:06:18 crc kubenswrapper[4754]: I1005 21:06:18.132355 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/697d9fce-0466-4ba2-a532-214dfd58a20e-webhook-cert\") pod \"metallb-operator-controller-manager-645fb4cd96-7qplt\" (UID: \"697d9fce-0466-4ba2-a532-214dfd58a20e\") " pod="metallb-system/metallb-operator-controller-manager-645fb4cd96-7qplt" Oct 05 21:06:18 crc kubenswrapper[4754]: I1005 21:06:18.143016 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/697d9fce-0466-4ba2-a532-214dfd58a20e-webhook-cert\") pod \"metallb-operator-controller-manager-645fb4cd96-7qplt\" (UID: \"697d9fce-0466-4ba2-a532-214dfd58a20e\") " pod="metallb-system/metallb-operator-controller-manager-645fb4cd96-7qplt" Oct 05 21:06:18 crc kubenswrapper[4754]: I1005 21:06:18.146124 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/697d9fce-0466-4ba2-a532-214dfd58a20e-apiservice-cert\") pod \"metallb-operator-controller-manager-645fb4cd96-7qplt\" (UID: \"697d9fce-0466-4ba2-a532-214dfd58a20e\") " pod="metallb-system/metallb-operator-controller-manager-645fb4cd96-7qplt" Oct 05 21:06:18 crc kubenswrapper[4754]: I1005 21:06:18.154048 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-chfgj\" (UniqueName: \"kubernetes.io/projected/697d9fce-0466-4ba2-a532-214dfd58a20e-kube-api-access-chfgj\") pod \"metallb-operator-controller-manager-645fb4cd96-7qplt\" (UID: \"697d9fce-0466-4ba2-a532-214dfd58a20e\") " pod="metallb-system/metallb-operator-controller-manager-645fb4cd96-7qplt" Oct 05 21:06:18 crc kubenswrapper[4754]: I1005 21:06:18.213057 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-645fb4cd96-7qplt" Oct 05 21:06:18 crc kubenswrapper[4754]: I1005 21:06:18.265732 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-5ff795db68-8xphz"] Oct 05 21:06:18 crc kubenswrapper[4754]: I1005 21:06:18.267184 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-5ff795db68-8xphz" Oct 05 21:06:18 crc kubenswrapper[4754]: I1005 21:06:18.282368 4754 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Oct 05 21:06:18 crc kubenswrapper[4754]: I1005 21:06:18.282814 4754 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-rhhbf" Oct 05 21:06:18 crc kubenswrapper[4754]: I1005 21:06:18.283612 4754 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Oct 05 21:06:18 crc kubenswrapper[4754]: I1005 21:06:18.318391 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-5ff795db68-8xphz"] Oct 05 21:06:18 crc kubenswrapper[4754]: I1005 21:06:18.444096 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/56375d90-5d62-4d6f-897f-28ca18bf3682-apiservice-cert\") pod \"metallb-operator-webhook-server-5ff795db68-8xphz\" (UID: \"56375d90-5d62-4d6f-897f-28ca18bf3682\") " pod="metallb-system/metallb-operator-webhook-server-5ff795db68-8xphz" Oct 05 21:06:18 crc kubenswrapper[4754]: I1005 21:06:18.444148 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/56375d90-5d62-4d6f-897f-28ca18bf3682-webhook-cert\") pod \"metallb-operator-webhook-server-5ff795db68-8xphz\" (UID: \"56375d90-5d62-4d6f-897f-28ca18bf3682\") " pod="metallb-system/metallb-operator-webhook-server-5ff795db68-8xphz" Oct 05 21:06:18 crc kubenswrapper[4754]: I1005 21:06:18.444196 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b2hqg\" (UniqueName: \"kubernetes.io/projected/56375d90-5d62-4d6f-897f-28ca18bf3682-kube-api-access-b2hqg\") pod \"metallb-operator-webhook-server-5ff795db68-8xphz\" (UID: \"56375d90-5d62-4d6f-897f-28ca18bf3682\") " pod="metallb-system/metallb-operator-webhook-server-5ff795db68-8xphz" Oct 05 21:06:18 crc kubenswrapper[4754]: I1005 21:06:18.545661 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/56375d90-5d62-4d6f-897f-28ca18bf3682-apiservice-cert\") pod \"metallb-operator-webhook-server-5ff795db68-8xphz\" (UID: \"56375d90-5d62-4d6f-897f-28ca18bf3682\") " pod="metallb-system/metallb-operator-webhook-server-5ff795db68-8xphz" Oct 05 21:06:18 crc kubenswrapper[4754]: I1005 21:06:18.545716 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/56375d90-5d62-4d6f-897f-28ca18bf3682-webhook-cert\") pod \"metallb-operator-webhook-server-5ff795db68-8xphz\" (UID: \"56375d90-5d62-4d6f-897f-28ca18bf3682\") " pod="metallb-system/metallb-operator-webhook-server-5ff795db68-8xphz" Oct 05 21:06:18 crc kubenswrapper[4754]: I1005 21:06:18.545754 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b2hqg\" (UniqueName: \"kubernetes.io/projected/56375d90-5d62-4d6f-897f-28ca18bf3682-kube-api-access-b2hqg\") pod \"metallb-operator-webhook-server-5ff795db68-8xphz\" (UID: \"56375d90-5d62-4d6f-897f-28ca18bf3682\") " pod="metallb-system/metallb-operator-webhook-server-5ff795db68-8xphz" Oct 05 21:06:18 crc kubenswrapper[4754]: I1005 21:06:18.554554 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/56375d90-5d62-4d6f-897f-28ca18bf3682-apiservice-cert\") pod \"metallb-operator-webhook-server-5ff795db68-8xphz\" (UID: \"56375d90-5d62-4d6f-897f-28ca18bf3682\") " pod="metallb-system/metallb-operator-webhook-server-5ff795db68-8xphz" Oct 05 21:06:18 crc kubenswrapper[4754]: I1005 21:06:18.555051 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/56375d90-5d62-4d6f-897f-28ca18bf3682-webhook-cert\") pod \"metallb-operator-webhook-server-5ff795db68-8xphz\" (UID: \"56375d90-5d62-4d6f-897f-28ca18bf3682\") " pod="metallb-system/metallb-operator-webhook-server-5ff795db68-8xphz" Oct 05 21:06:18 crc kubenswrapper[4754]: I1005 21:06:18.587389 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b2hqg\" (UniqueName: \"kubernetes.io/projected/56375d90-5d62-4d6f-897f-28ca18bf3682-kube-api-access-b2hqg\") pod \"metallb-operator-webhook-server-5ff795db68-8xphz\" (UID: \"56375d90-5d62-4d6f-897f-28ca18bf3682\") " pod="metallb-system/metallb-operator-webhook-server-5ff795db68-8xphz" Oct 05 21:06:18 crc kubenswrapper[4754]: I1005 21:06:18.621748 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-5ff795db68-8xphz" Oct 05 21:06:18 crc kubenswrapper[4754]: I1005 21:06:18.670949 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-645fb4cd96-7qplt"] Oct 05 21:06:19 crc kubenswrapper[4754]: I1005 21:06:19.050964 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-5ff795db68-8xphz"] Oct 05 21:06:19 crc kubenswrapper[4754]: W1005 21:06:19.059822 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod56375d90_5d62_4d6f_897f_28ca18bf3682.slice/crio-86f86a0fb19088c75eb78e9e63bcb0bf14901dd79869b8a825bc6c02c30c649c WatchSource:0}: Error finding container 86f86a0fb19088c75eb78e9e63bcb0bf14901dd79869b8a825bc6c02c30c649c: Status 404 returned error can't find the container with id 86f86a0fb19088c75eb78e9e63bcb0bf14901dd79869b8a825bc6c02c30c649c Oct 05 21:06:19 crc kubenswrapper[4754]: I1005 21:06:19.322897 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-5ff795db68-8xphz" event={"ID":"56375d90-5d62-4d6f-897f-28ca18bf3682","Type":"ContainerStarted","Data":"86f86a0fb19088c75eb78e9e63bcb0bf14901dd79869b8a825bc6c02c30c649c"} Oct 05 21:06:19 crc kubenswrapper[4754]: I1005 21:06:19.324594 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-645fb4cd96-7qplt" event={"ID":"697d9fce-0466-4ba2-a532-214dfd58a20e","Type":"ContainerStarted","Data":"80fa7b339361e026ac5bf644fb4475c2ef016721b29dcac02b825c208d65257f"} Oct 05 21:06:25 crc kubenswrapper[4754]: I1005 21:06:25.366403 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-5ff795db68-8xphz" event={"ID":"56375d90-5d62-4d6f-897f-28ca18bf3682","Type":"ContainerStarted","Data":"ad748107f905197ffb41c229bd44036b3f8232e6854e8af8b97cac55f8d5cafb"} Oct 05 21:06:25 crc kubenswrapper[4754]: I1005 21:06:25.367183 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-5ff795db68-8xphz" Oct 05 21:06:25 crc kubenswrapper[4754]: I1005 21:06:25.368933 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-645fb4cd96-7qplt" event={"ID":"697d9fce-0466-4ba2-a532-214dfd58a20e","Type":"ContainerStarted","Data":"6c39b94da66d4e4813f3f7efdf76c9a72bd648a27c2d8023a99e09343c956c8e"} Oct 05 21:06:25 crc kubenswrapper[4754]: I1005 21:06:25.369166 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-645fb4cd96-7qplt" Oct 05 21:06:25 crc kubenswrapper[4754]: I1005 21:06:25.386292 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-5ff795db68-8xphz" podStartSLOduration=1.8977754409999998 podStartE2EDuration="7.386268784s" podCreationTimestamp="2025-10-05 21:06:18 +0000 UTC" firstStartedPulling="2025-10-05 21:06:19.06903597 +0000 UTC m=+702.973154670" lastFinishedPulling="2025-10-05 21:06:24.557529293 +0000 UTC m=+708.461648013" observedRunningTime="2025-10-05 21:06:25.384707044 +0000 UTC m=+709.288825744" watchObservedRunningTime="2025-10-05 21:06:25.386268784 +0000 UTC m=+709.290387494" Oct 05 21:06:25 crc kubenswrapper[4754]: I1005 21:06:25.418817 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-645fb4cd96-7qplt" podStartSLOduration=2.5726400590000003 podStartE2EDuration="8.418794636s" podCreationTimestamp="2025-10-05 21:06:17 +0000 UTC" firstStartedPulling="2025-10-05 21:06:18.701180853 +0000 UTC m=+702.605299563" lastFinishedPulling="2025-10-05 21:06:24.54733543 +0000 UTC m=+708.451454140" observedRunningTime="2025-10-05 21:06:25.413647502 +0000 UTC m=+709.317766222" watchObservedRunningTime="2025-10-05 21:06:25.418794636 +0000 UTC m=+709.322913356" Oct 05 21:06:35 crc kubenswrapper[4754]: I1005 21:06:35.245390 4754 patch_prober.go:28] interesting pod/machine-config-daemon-b2h9k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 21:06:35 crc kubenswrapper[4754]: I1005 21:06:35.245730 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 21:06:38 crc kubenswrapper[4754]: I1005 21:06:38.630477 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-5ff795db68-8xphz" Oct 05 21:06:58 crc kubenswrapper[4754]: I1005 21:06:58.217994 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-645fb4cd96-7qplt" Oct 05 21:06:59 crc kubenswrapper[4754]: I1005 21:06:59.007382 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-zt7sz"] Oct 05 21:06:59 crc kubenswrapper[4754]: I1005 21:06:59.008310 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-zt7sz" Oct 05 21:06:59 crc kubenswrapper[4754]: I1005 21:06:59.010238 4754 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Oct 05 21:06:59 crc kubenswrapper[4754]: I1005 21:06:59.010545 4754 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-kjbpn" Oct 05 21:06:59 crc kubenswrapper[4754]: I1005 21:06:59.027111 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-tfh49"] Oct 05 21:06:59 crc kubenswrapper[4754]: I1005 21:06:59.029929 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-tfh49" Oct 05 21:06:59 crc kubenswrapper[4754]: I1005 21:06:59.033484 4754 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Oct 05 21:06:59 crc kubenswrapper[4754]: I1005 21:06:59.033517 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Oct 05 21:06:59 crc kubenswrapper[4754]: I1005 21:06:59.040100 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-zt7sz"] Oct 05 21:06:59 crc kubenswrapper[4754]: I1005 21:06:59.127301 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-hkvd8"] Oct 05 21:06:59 crc kubenswrapper[4754]: I1005 21:06:59.128385 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-hkvd8" Oct 05 21:06:59 crc kubenswrapper[4754]: I1005 21:06:59.130852 4754 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Oct 05 21:06:59 crc kubenswrapper[4754]: I1005 21:06:59.131214 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Oct 05 21:06:59 crc kubenswrapper[4754]: I1005 21:06:59.131259 4754 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-5kd85" Oct 05 21:06:59 crc kubenswrapper[4754]: I1005 21:06:59.133049 4754 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Oct 05 21:06:59 crc kubenswrapper[4754]: I1005 21:06:59.148918 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-68d546b9d8-78nmb"] Oct 05 21:06:59 crc kubenswrapper[4754]: I1005 21:06:59.150066 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-68d546b9d8-78nmb" Oct 05 21:06:59 crc kubenswrapper[4754]: I1005 21:06:59.151790 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/dbf73baa-e798-43fa-95de-daec87f5809a-metrics\") pod \"frr-k8s-tfh49\" (UID: \"dbf73baa-e798-43fa-95de-daec87f5809a\") " pod="metallb-system/frr-k8s-tfh49" Oct 05 21:06:59 crc kubenswrapper[4754]: I1005 21:06:59.151831 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p9xds\" (UniqueName: \"kubernetes.io/projected/dbf73baa-e798-43fa-95de-daec87f5809a-kube-api-access-p9xds\") pod \"frr-k8s-tfh49\" (UID: \"dbf73baa-e798-43fa-95de-daec87f5809a\") " pod="metallb-system/frr-k8s-tfh49" Oct 05 21:06:59 crc kubenswrapper[4754]: I1005 21:06:59.151878 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/69687995-5b0d-4aae-9551-1b1b9cb2bec3-cert\") pod \"frr-k8s-webhook-server-64bf5d555-zt7sz\" (UID: \"69687995-5b0d-4aae-9551-1b1b9cb2bec3\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-zt7sz" Oct 05 21:06:59 crc kubenswrapper[4754]: I1005 21:06:59.151907 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/dbf73baa-e798-43fa-95de-daec87f5809a-frr-sockets\") pod \"frr-k8s-tfh49\" (UID: \"dbf73baa-e798-43fa-95de-daec87f5809a\") " pod="metallb-system/frr-k8s-tfh49" Oct 05 21:06:59 crc kubenswrapper[4754]: I1005 21:06:59.151951 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/dbf73baa-e798-43fa-95de-daec87f5809a-frr-startup\") pod \"frr-k8s-tfh49\" (UID: \"dbf73baa-e798-43fa-95de-daec87f5809a\") " pod="metallb-system/frr-k8s-tfh49" Oct 05 21:06:59 crc kubenswrapper[4754]: I1005 21:06:59.151987 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/dbf73baa-e798-43fa-95de-daec87f5809a-reloader\") pod \"frr-k8s-tfh49\" (UID: \"dbf73baa-e798-43fa-95de-daec87f5809a\") " pod="metallb-system/frr-k8s-tfh49" Oct 05 21:06:59 crc kubenswrapper[4754]: I1005 21:06:59.152010 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/dbf73baa-e798-43fa-95de-daec87f5809a-frr-conf\") pod \"frr-k8s-tfh49\" (UID: \"dbf73baa-e798-43fa-95de-daec87f5809a\") " pod="metallb-system/frr-k8s-tfh49" Oct 05 21:06:59 crc kubenswrapper[4754]: I1005 21:06:59.152039 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/dbf73baa-e798-43fa-95de-daec87f5809a-metrics-certs\") pod \"frr-k8s-tfh49\" (UID: \"dbf73baa-e798-43fa-95de-daec87f5809a\") " pod="metallb-system/frr-k8s-tfh49" Oct 05 21:06:59 crc kubenswrapper[4754]: I1005 21:06:59.152070 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4jj84\" (UniqueName: \"kubernetes.io/projected/69687995-5b0d-4aae-9551-1b1b9cb2bec3-kube-api-access-4jj84\") pod \"frr-k8s-webhook-server-64bf5d555-zt7sz\" (UID: \"69687995-5b0d-4aae-9551-1b1b9cb2bec3\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-zt7sz" Oct 05 21:06:59 crc kubenswrapper[4754]: I1005 21:06:59.152727 4754 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Oct 05 21:06:59 crc kubenswrapper[4754]: I1005 21:06:59.176696 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-68d546b9d8-78nmb"] Oct 05 21:06:59 crc kubenswrapper[4754]: I1005 21:06:59.252672 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/dbf73baa-e798-43fa-95de-daec87f5809a-frr-startup\") pod \"frr-k8s-tfh49\" (UID: \"dbf73baa-e798-43fa-95de-daec87f5809a\") " pod="metallb-system/frr-k8s-tfh49" Oct 05 21:06:59 crc kubenswrapper[4754]: I1005 21:06:59.252733 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/34ab3bd8-96ff-4f52-ae2e-e8cbd748827d-metrics-certs\") pod \"speaker-hkvd8\" (UID: \"34ab3bd8-96ff-4f52-ae2e-e8cbd748827d\") " pod="metallb-system/speaker-hkvd8" Oct 05 21:06:59 crc kubenswrapper[4754]: I1005 21:06:59.252753 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d8zr7\" (UniqueName: \"kubernetes.io/projected/34ab3bd8-96ff-4f52-ae2e-e8cbd748827d-kube-api-access-d8zr7\") pod \"speaker-hkvd8\" (UID: \"34ab3bd8-96ff-4f52-ae2e-e8cbd748827d\") " pod="metallb-system/speaker-hkvd8" Oct 05 21:06:59 crc kubenswrapper[4754]: I1005 21:06:59.252774 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/dbf73baa-e798-43fa-95de-daec87f5809a-reloader\") pod \"frr-k8s-tfh49\" (UID: \"dbf73baa-e798-43fa-95de-daec87f5809a\") " pod="metallb-system/frr-k8s-tfh49" Oct 05 21:06:59 crc kubenswrapper[4754]: I1005 21:06:59.252793 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/dbf73baa-e798-43fa-95de-daec87f5809a-frr-conf\") pod \"frr-k8s-tfh49\" (UID: \"dbf73baa-e798-43fa-95de-daec87f5809a\") " pod="metallb-system/frr-k8s-tfh49" Oct 05 21:06:59 crc kubenswrapper[4754]: I1005 21:06:59.252810 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/34ab3bd8-96ff-4f52-ae2e-e8cbd748827d-memberlist\") pod \"speaker-hkvd8\" (UID: \"34ab3bd8-96ff-4f52-ae2e-e8cbd748827d\") " pod="metallb-system/speaker-hkvd8" Oct 05 21:06:59 crc kubenswrapper[4754]: I1005 21:06:59.252828 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/dbf73baa-e798-43fa-95de-daec87f5809a-metrics-certs\") pod \"frr-k8s-tfh49\" (UID: \"dbf73baa-e798-43fa-95de-daec87f5809a\") " pod="metallb-system/frr-k8s-tfh49" Oct 05 21:06:59 crc kubenswrapper[4754]: I1005 21:06:59.252854 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4jj84\" (UniqueName: \"kubernetes.io/projected/69687995-5b0d-4aae-9551-1b1b9cb2bec3-kube-api-access-4jj84\") pod \"frr-k8s-webhook-server-64bf5d555-zt7sz\" (UID: \"69687995-5b0d-4aae-9551-1b1b9cb2bec3\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-zt7sz" Oct 05 21:06:59 crc kubenswrapper[4754]: I1005 21:06:59.252878 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/22f6b8d5-a7f7-4db9-bd76-eb9e5e23ea79-metrics-certs\") pod \"controller-68d546b9d8-78nmb\" (UID: \"22f6b8d5-a7f7-4db9-bd76-eb9e5e23ea79\") " pod="metallb-system/controller-68d546b9d8-78nmb" Oct 05 21:06:59 crc kubenswrapper[4754]: I1005 21:06:59.252900 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/dbf73baa-e798-43fa-95de-daec87f5809a-metrics\") pod \"frr-k8s-tfh49\" (UID: \"dbf73baa-e798-43fa-95de-daec87f5809a\") " pod="metallb-system/frr-k8s-tfh49" Oct 05 21:06:59 crc kubenswrapper[4754]: I1005 21:06:59.252918 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p9xds\" (UniqueName: \"kubernetes.io/projected/dbf73baa-e798-43fa-95de-daec87f5809a-kube-api-access-p9xds\") pod \"frr-k8s-tfh49\" (UID: \"dbf73baa-e798-43fa-95de-daec87f5809a\") " pod="metallb-system/frr-k8s-tfh49" Oct 05 21:06:59 crc kubenswrapper[4754]: I1005 21:06:59.252933 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/34ab3bd8-96ff-4f52-ae2e-e8cbd748827d-metallb-excludel2\") pod \"speaker-hkvd8\" (UID: \"34ab3bd8-96ff-4f52-ae2e-e8cbd748827d\") " pod="metallb-system/speaker-hkvd8" Oct 05 21:06:59 crc kubenswrapper[4754]: I1005 21:06:59.252966 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/69687995-5b0d-4aae-9551-1b1b9cb2bec3-cert\") pod \"frr-k8s-webhook-server-64bf5d555-zt7sz\" (UID: \"69687995-5b0d-4aae-9551-1b1b9cb2bec3\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-zt7sz" Oct 05 21:06:59 crc kubenswrapper[4754]: I1005 21:06:59.252983 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-79xnc\" (UniqueName: \"kubernetes.io/projected/22f6b8d5-a7f7-4db9-bd76-eb9e5e23ea79-kube-api-access-79xnc\") pod \"controller-68d546b9d8-78nmb\" (UID: \"22f6b8d5-a7f7-4db9-bd76-eb9e5e23ea79\") " pod="metallb-system/controller-68d546b9d8-78nmb" Oct 05 21:06:59 crc kubenswrapper[4754]: I1005 21:06:59.252999 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/22f6b8d5-a7f7-4db9-bd76-eb9e5e23ea79-cert\") pod \"controller-68d546b9d8-78nmb\" (UID: \"22f6b8d5-a7f7-4db9-bd76-eb9e5e23ea79\") " pod="metallb-system/controller-68d546b9d8-78nmb" Oct 05 21:06:59 crc kubenswrapper[4754]: I1005 21:06:59.253017 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/dbf73baa-e798-43fa-95de-daec87f5809a-frr-sockets\") pod \"frr-k8s-tfh49\" (UID: \"dbf73baa-e798-43fa-95de-daec87f5809a\") " pod="metallb-system/frr-k8s-tfh49" Oct 05 21:06:59 crc kubenswrapper[4754]: I1005 21:06:59.253426 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/dbf73baa-e798-43fa-95de-daec87f5809a-frr-sockets\") pod \"frr-k8s-tfh49\" (UID: \"dbf73baa-e798-43fa-95de-daec87f5809a\") " pod="metallb-system/frr-k8s-tfh49" Oct 05 21:06:59 crc kubenswrapper[4754]: I1005 21:06:59.254104 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/dbf73baa-e798-43fa-95de-daec87f5809a-frr-startup\") pod \"frr-k8s-tfh49\" (UID: \"dbf73baa-e798-43fa-95de-daec87f5809a\") " pod="metallb-system/frr-k8s-tfh49" Oct 05 21:06:59 crc kubenswrapper[4754]: I1005 21:06:59.254295 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/dbf73baa-e798-43fa-95de-daec87f5809a-reloader\") pod \"frr-k8s-tfh49\" (UID: \"dbf73baa-e798-43fa-95de-daec87f5809a\") " pod="metallb-system/frr-k8s-tfh49" Oct 05 21:06:59 crc kubenswrapper[4754]: I1005 21:06:59.254455 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/dbf73baa-e798-43fa-95de-daec87f5809a-frr-conf\") pod \"frr-k8s-tfh49\" (UID: \"dbf73baa-e798-43fa-95de-daec87f5809a\") " pod="metallb-system/frr-k8s-tfh49" Oct 05 21:06:59 crc kubenswrapper[4754]: E1005 21:06:59.254541 4754 secret.go:188] Couldn't get secret metallb-system/frr-k8s-certs-secret: secret "frr-k8s-certs-secret" not found Oct 05 21:06:59 crc kubenswrapper[4754]: E1005 21:06:59.254582 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/dbf73baa-e798-43fa-95de-daec87f5809a-metrics-certs podName:dbf73baa-e798-43fa-95de-daec87f5809a nodeName:}" failed. No retries permitted until 2025-10-05 21:06:59.754568002 +0000 UTC m=+743.658686712 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/dbf73baa-e798-43fa-95de-daec87f5809a-metrics-certs") pod "frr-k8s-tfh49" (UID: "dbf73baa-e798-43fa-95de-daec87f5809a") : secret "frr-k8s-certs-secret" not found Oct 05 21:06:59 crc kubenswrapper[4754]: I1005 21:06:59.255103 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/dbf73baa-e798-43fa-95de-daec87f5809a-metrics\") pod \"frr-k8s-tfh49\" (UID: \"dbf73baa-e798-43fa-95de-daec87f5809a\") " pod="metallb-system/frr-k8s-tfh49" Oct 05 21:06:59 crc kubenswrapper[4754]: E1005 21:06:59.255276 4754 secret.go:188] Couldn't get secret metallb-system/frr-k8s-webhook-server-cert: secret "frr-k8s-webhook-server-cert" not found Oct 05 21:06:59 crc kubenswrapper[4754]: E1005 21:06:59.255312 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/69687995-5b0d-4aae-9551-1b1b9cb2bec3-cert podName:69687995-5b0d-4aae-9551-1b1b9cb2bec3 nodeName:}" failed. No retries permitted until 2025-10-05 21:06:59.755304011 +0000 UTC m=+743.659422721 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/69687995-5b0d-4aae-9551-1b1b9cb2bec3-cert") pod "frr-k8s-webhook-server-64bf5d555-zt7sz" (UID: "69687995-5b0d-4aae-9551-1b1b9cb2bec3") : secret "frr-k8s-webhook-server-cert" not found Oct 05 21:06:59 crc kubenswrapper[4754]: I1005 21:06:59.277819 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4jj84\" (UniqueName: \"kubernetes.io/projected/69687995-5b0d-4aae-9551-1b1b9cb2bec3-kube-api-access-4jj84\") pod \"frr-k8s-webhook-server-64bf5d555-zt7sz\" (UID: \"69687995-5b0d-4aae-9551-1b1b9cb2bec3\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-zt7sz" Oct 05 21:06:59 crc kubenswrapper[4754]: I1005 21:06:59.283982 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p9xds\" (UniqueName: \"kubernetes.io/projected/dbf73baa-e798-43fa-95de-daec87f5809a-kube-api-access-p9xds\") pod \"frr-k8s-tfh49\" (UID: \"dbf73baa-e798-43fa-95de-daec87f5809a\") " pod="metallb-system/frr-k8s-tfh49" Oct 05 21:06:59 crc kubenswrapper[4754]: I1005 21:06:59.354056 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/22f6b8d5-a7f7-4db9-bd76-eb9e5e23ea79-metrics-certs\") pod \"controller-68d546b9d8-78nmb\" (UID: \"22f6b8d5-a7f7-4db9-bd76-eb9e5e23ea79\") " pod="metallb-system/controller-68d546b9d8-78nmb" Oct 05 21:06:59 crc kubenswrapper[4754]: I1005 21:06:59.354107 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/34ab3bd8-96ff-4f52-ae2e-e8cbd748827d-metallb-excludel2\") pod \"speaker-hkvd8\" (UID: \"34ab3bd8-96ff-4f52-ae2e-e8cbd748827d\") " pod="metallb-system/speaker-hkvd8" Oct 05 21:06:59 crc kubenswrapper[4754]: I1005 21:06:59.354145 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-79xnc\" (UniqueName: \"kubernetes.io/projected/22f6b8d5-a7f7-4db9-bd76-eb9e5e23ea79-kube-api-access-79xnc\") pod \"controller-68d546b9d8-78nmb\" (UID: \"22f6b8d5-a7f7-4db9-bd76-eb9e5e23ea79\") " pod="metallb-system/controller-68d546b9d8-78nmb" Oct 05 21:06:59 crc kubenswrapper[4754]: I1005 21:06:59.354162 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/22f6b8d5-a7f7-4db9-bd76-eb9e5e23ea79-cert\") pod \"controller-68d546b9d8-78nmb\" (UID: \"22f6b8d5-a7f7-4db9-bd76-eb9e5e23ea79\") " pod="metallb-system/controller-68d546b9d8-78nmb" Oct 05 21:06:59 crc kubenswrapper[4754]: I1005 21:06:59.354202 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/34ab3bd8-96ff-4f52-ae2e-e8cbd748827d-metrics-certs\") pod \"speaker-hkvd8\" (UID: \"34ab3bd8-96ff-4f52-ae2e-e8cbd748827d\") " pod="metallb-system/speaker-hkvd8" Oct 05 21:06:59 crc kubenswrapper[4754]: I1005 21:06:59.354218 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d8zr7\" (UniqueName: \"kubernetes.io/projected/34ab3bd8-96ff-4f52-ae2e-e8cbd748827d-kube-api-access-d8zr7\") pod \"speaker-hkvd8\" (UID: \"34ab3bd8-96ff-4f52-ae2e-e8cbd748827d\") " pod="metallb-system/speaker-hkvd8" Oct 05 21:06:59 crc kubenswrapper[4754]: I1005 21:06:59.354248 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/34ab3bd8-96ff-4f52-ae2e-e8cbd748827d-memberlist\") pod \"speaker-hkvd8\" (UID: \"34ab3bd8-96ff-4f52-ae2e-e8cbd748827d\") " pod="metallb-system/speaker-hkvd8" Oct 05 21:06:59 crc kubenswrapper[4754]: E1005 21:06:59.354347 4754 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Oct 05 21:06:59 crc kubenswrapper[4754]: E1005 21:06:59.354392 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/34ab3bd8-96ff-4f52-ae2e-e8cbd748827d-memberlist podName:34ab3bd8-96ff-4f52-ae2e-e8cbd748827d nodeName:}" failed. No retries permitted until 2025-10-05 21:06:59.854379703 +0000 UTC m=+743.758498413 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/34ab3bd8-96ff-4f52-ae2e-e8cbd748827d-memberlist") pod "speaker-hkvd8" (UID: "34ab3bd8-96ff-4f52-ae2e-e8cbd748827d") : secret "metallb-memberlist" not found Oct 05 21:06:59 crc kubenswrapper[4754]: I1005 21:06:59.355648 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/34ab3bd8-96ff-4f52-ae2e-e8cbd748827d-metallb-excludel2\") pod \"speaker-hkvd8\" (UID: \"34ab3bd8-96ff-4f52-ae2e-e8cbd748827d\") " pod="metallb-system/speaker-hkvd8" Oct 05 21:06:59 crc kubenswrapper[4754]: I1005 21:06:59.358956 4754 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Oct 05 21:06:59 crc kubenswrapper[4754]: I1005 21:06:59.359168 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/22f6b8d5-a7f7-4db9-bd76-eb9e5e23ea79-metrics-certs\") pod \"controller-68d546b9d8-78nmb\" (UID: \"22f6b8d5-a7f7-4db9-bd76-eb9e5e23ea79\") " pod="metallb-system/controller-68d546b9d8-78nmb" Oct 05 21:06:59 crc kubenswrapper[4754]: I1005 21:06:59.363334 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/34ab3bd8-96ff-4f52-ae2e-e8cbd748827d-metrics-certs\") pod \"speaker-hkvd8\" (UID: \"34ab3bd8-96ff-4f52-ae2e-e8cbd748827d\") " pod="metallb-system/speaker-hkvd8" Oct 05 21:06:59 crc kubenswrapper[4754]: I1005 21:06:59.370868 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/22f6b8d5-a7f7-4db9-bd76-eb9e5e23ea79-cert\") pod \"controller-68d546b9d8-78nmb\" (UID: \"22f6b8d5-a7f7-4db9-bd76-eb9e5e23ea79\") " pod="metallb-system/controller-68d546b9d8-78nmb" Oct 05 21:06:59 crc kubenswrapper[4754]: I1005 21:06:59.387896 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d8zr7\" (UniqueName: \"kubernetes.io/projected/34ab3bd8-96ff-4f52-ae2e-e8cbd748827d-kube-api-access-d8zr7\") pod \"speaker-hkvd8\" (UID: \"34ab3bd8-96ff-4f52-ae2e-e8cbd748827d\") " pod="metallb-system/speaker-hkvd8" Oct 05 21:06:59 crc kubenswrapper[4754]: I1005 21:06:59.388520 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-79xnc\" (UniqueName: \"kubernetes.io/projected/22f6b8d5-a7f7-4db9-bd76-eb9e5e23ea79-kube-api-access-79xnc\") pod \"controller-68d546b9d8-78nmb\" (UID: \"22f6b8d5-a7f7-4db9-bd76-eb9e5e23ea79\") " pod="metallb-system/controller-68d546b9d8-78nmb" Oct 05 21:06:59 crc kubenswrapper[4754]: I1005 21:06:59.463345 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-68d546b9d8-78nmb" Oct 05 21:06:59 crc kubenswrapper[4754]: I1005 21:06:59.658646 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-68d546b9d8-78nmb"] Oct 05 21:06:59 crc kubenswrapper[4754]: W1005 21:06:59.665601 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod22f6b8d5_a7f7_4db9_bd76_eb9e5e23ea79.slice/crio-b87a1586c80d28ba5dc7dc02f6f9a732ba24f41f917e3abb53ac824c0cb97368 WatchSource:0}: Error finding container b87a1586c80d28ba5dc7dc02f6f9a732ba24f41f917e3abb53ac824c0cb97368: Status 404 returned error can't find the container with id b87a1586c80d28ba5dc7dc02f6f9a732ba24f41f917e3abb53ac824c0cb97368 Oct 05 21:06:59 crc kubenswrapper[4754]: I1005 21:06:59.758094 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/dbf73baa-e798-43fa-95de-daec87f5809a-metrics-certs\") pod \"frr-k8s-tfh49\" (UID: \"dbf73baa-e798-43fa-95de-daec87f5809a\") " pod="metallb-system/frr-k8s-tfh49" Oct 05 21:06:59 crc kubenswrapper[4754]: I1005 21:06:59.758386 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/69687995-5b0d-4aae-9551-1b1b9cb2bec3-cert\") pod \"frr-k8s-webhook-server-64bf5d555-zt7sz\" (UID: \"69687995-5b0d-4aae-9551-1b1b9cb2bec3\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-zt7sz" Oct 05 21:06:59 crc kubenswrapper[4754]: I1005 21:06:59.761781 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/dbf73baa-e798-43fa-95de-daec87f5809a-metrics-certs\") pod \"frr-k8s-tfh49\" (UID: \"dbf73baa-e798-43fa-95de-daec87f5809a\") " pod="metallb-system/frr-k8s-tfh49" Oct 05 21:06:59 crc kubenswrapper[4754]: I1005 21:06:59.762034 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/69687995-5b0d-4aae-9551-1b1b9cb2bec3-cert\") pod \"frr-k8s-webhook-server-64bf5d555-zt7sz\" (UID: \"69687995-5b0d-4aae-9551-1b1b9cb2bec3\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-zt7sz" Oct 05 21:06:59 crc kubenswrapper[4754]: I1005 21:06:59.859693 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/34ab3bd8-96ff-4f52-ae2e-e8cbd748827d-memberlist\") pod \"speaker-hkvd8\" (UID: \"34ab3bd8-96ff-4f52-ae2e-e8cbd748827d\") " pod="metallb-system/speaker-hkvd8" Oct 05 21:06:59 crc kubenswrapper[4754]: E1005 21:06:59.859877 4754 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Oct 05 21:06:59 crc kubenswrapper[4754]: E1005 21:06:59.859941 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/34ab3bd8-96ff-4f52-ae2e-e8cbd748827d-memberlist podName:34ab3bd8-96ff-4f52-ae2e-e8cbd748827d nodeName:}" failed. No retries permitted until 2025-10-05 21:07:00.85992513 +0000 UTC m=+744.764043850 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/34ab3bd8-96ff-4f52-ae2e-e8cbd748827d-memberlist") pod "speaker-hkvd8" (UID: "34ab3bd8-96ff-4f52-ae2e-e8cbd748827d") : secret "metallb-memberlist" not found Oct 05 21:06:59 crc kubenswrapper[4754]: I1005 21:06:59.929571 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-zt7sz" Oct 05 21:06:59 crc kubenswrapper[4754]: I1005 21:06:59.947741 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-tfh49" Oct 05 21:07:00 crc kubenswrapper[4754]: I1005 21:07:00.144258 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-zt7sz"] Oct 05 21:07:00 crc kubenswrapper[4754]: I1005 21:07:00.602374 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-78nmb" event={"ID":"22f6b8d5-a7f7-4db9-bd76-eb9e5e23ea79","Type":"ContainerStarted","Data":"499bd26e4893d02768cb676fd3b64408001ad2bab495c0e69d5960c43bb8ee35"} Oct 05 21:07:00 crc kubenswrapper[4754]: I1005 21:07:00.602424 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-78nmb" event={"ID":"22f6b8d5-a7f7-4db9-bd76-eb9e5e23ea79","Type":"ContainerStarted","Data":"b87a1586c80d28ba5dc7dc02f6f9a732ba24f41f917e3abb53ac824c0cb97368"} Oct 05 21:07:00 crc kubenswrapper[4754]: I1005 21:07:00.603784 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-zt7sz" event={"ID":"69687995-5b0d-4aae-9551-1b1b9cb2bec3","Type":"ContainerStarted","Data":"3392998d4d43c767fb107a2a2ef2e89038bdcebac74590d003e38e6319f1f86b"} Oct 05 21:07:00 crc kubenswrapper[4754]: I1005 21:07:00.605183 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-tfh49" event={"ID":"dbf73baa-e798-43fa-95de-daec87f5809a","Type":"ContainerStarted","Data":"5b361e56ba7ff6ea1e3c8de99c2aa03bca03df24a573698ff994ae76968f4ccc"} Oct 05 21:07:00 crc kubenswrapper[4754]: I1005 21:07:00.873206 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/34ab3bd8-96ff-4f52-ae2e-e8cbd748827d-memberlist\") pod \"speaker-hkvd8\" (UID: \"34ab3bd8-96ff-4f52-ae2e-e8cbd748827d\") " pod="metallb-system/speaker-hkvd8" Oct 05 21:07:00 crc kubenswrapper[4754]: I1005 21:07:00.923997 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/34ab3bd8-96ff-4f52-ae2e-e8cbd748827d-memberlist\") pod \"speaker-hkvd8\" (UID: \"34ab3bd8-96ff-4f52-ae2e-e8cbd748827d\") " pod="metallb-system/speaker-hkvd8" Oct 05 21:07:00 crc kubenswrapper[4754]: I1005 21:07:00.941527 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-hkvd8" Oct 05 21:07:01 crc kubenswrapper[4754]: W1005 21:07:01.064564 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod34ab3bd8_96ff_4f52_ae2e_e8cbd748827d.slice/crio-fb59f09748cca7e0330495e1988d977899d99917823972f9efb72cc6589a7230 WatchSource:0}: Error finding container fb59f09748cca7e0330495e1988d977899d99917823972f9efb72cc6589a7230: Status 404 returned error can't find the container with id fb59f09748cca7e0330495e1988d977899d99917823972f9efb72cc6589a7230 Oct 05 21:07:01 crc kubenswrapper[4754]: I1005 21:07:01.618035 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-hkvd8" event={"ID":"34ab3bd8-96ff-4f52-ae2e-e8cbd748827d","Type":"ContainerStarted","Data":"dd168b32f43b008eda2c13dc83df58152d1bcaa60217ae053be4ec6c2fff1ac1"} Oct 05 21:07:01 crc kubenswrapper[4754]: I1005 21:07:01.618378 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-hkvd8" event={"ID":"34ab3bd8-96ff-4f52-ae2e-e8cbd748827d","Type":"ContainerStarted","Data":"fb59f09748cca7e0330495e1988d977899d99917823972f9efb72cc6589a7230"} Oct 05 21:07:01 crc kubenswrapper[4754]: I1005 21:07:01.622254 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-78nmb" event={"ID":"22f6b8d5-a7f7-4db9-bd76-eb9e5e23ea79","Type":"ContainerStarted","Data":"79f42f03d004800a46e19edf68619a57dfcabc4d73946a068a276b8a675aabf7"} Oct 05 21:07:01 crc kubenswrapper[4754]: I1005 21:07:01.623301 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-68d546b9d8-78nmb" Oct 05 21:07:02 crc kubenswrapper[4754]: I1005 21:07:02.645215 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-hkvd8" event={"ID":"34ab3bd8-96ff-4f52-ae2e-e8cbd748827d","Type":"ContainerStarted","Data":"effab35d32a7b1c8a04dc25ea277f4e753bb72c5c348350d3adb1944c7f87bb3"} Oct 05 21:07:02 crc kubenswrapper[4754]: I1005 21:07:02.645309 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-hkvd8" Oct 05 21:07:02 crc kubenswrapper[4754]: I1005 21:07:02.672865 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-hkvd8" podStartSLOduration=3.672848245 podStartE2EDuration="3.672848245s" podCreationTimestamp="2025-10-05 21:06:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 21:07:02.668229395 +0000 UTC m=+746.572348105" watchObservedRunningTime="2025-10-05 21:07:02.672848245 +0000 UTC m=+746.576966955" Oct 05 21:07:02 crc kubenswrapper[4754]: I1005 21:07:02.673614 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-68d546b9d8-78nmb" podStartSLOduration=3.673590274 podStartE2EDuration="3.673590274s" podCreationTimestamp="2025-10-05 21:06:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 21:07:01.647954385 +0000 UTC m=+745.552073095" watchObservedRunningTime="2025-10-05 21:07:02.673590274 +0000 UTC m=+746.577708984" Oct 05 21:07:05 crc kubenswrapper[4754]: I1005 21:07:05.245158 4754 patch_prober.go:28] interesting pod/machine-config-daemon-b2h9k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 21:07:05 crc kubenswrapper[4754]: I1005 21:07:05.245813 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 21:07:05 crc kubenswrapper[4754]: I1005 21:07:05.832458 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-rvsbj"] Oct 05 21:07:05 crc kubenswrapper[4754]: I1005 21:07:05.832675 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-rvsbj" podUID="b1286ed6-4e15-46e0-b4c3-3b7e3cad6057" containerName="controller-manager" containerID="cri-o://c237d2e1ac6335fad6f3391f43562c0780d719ecafeac3b0c81b33874c1d1da4" gracePeriod=30 Oct 05 21:07:05 crc kubenswrapper[4754]: I1005 21:07:05.942145 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-pnhf9"] Oct 05 21:07:05 crc kubenswrapper[4754]: I1005 21:07:05.942643 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-pnhf9" podUID="1e16d79e-eb1e-4b8b-8f0a-acb370163d17" containerName="route-controller-manager" containerID="cri-o://2a2f7d826d242de3ef14c3ee7cf71e919b4aa69848335790746b587439befeb6" gracePeriod=30 Oct 05 21:07:06 crc kubenswrapper[4754]: I1005 21:07:06.693899 4754 generic.go:334] "Generic (PLEG): container finished" podID="1e16d79e-eb1e-4b8b-8f0a-acb370163d17" containerID="2a2f7d826d242de3ef14c3ee7cf71e919b4aa69848335790746b587439befeb6" exitCode=0 Oct 05 21:07:06 crc kubenswrapper[4754]: I1005 21:07:06.693994 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-pnhf9" event={"ID":"1e16d79e-eb1e-4b8b-8f0a-acb370163d17","Type":"ContainerDied","Data":"2a2f7d826d242de3ef14c3ee7cf71e919b4aa69848335790746b587439befeb6"} Oct 05 21:07:06 crc kubenswrapper[4754]: I1005 21:07:06.697405 4754 generic.go:334] "Generic (PLEG): container finished" podID="b1286ed6-4e15-46e0-b4c3-3b7e3cad6057" containerID="c237d2e1ac6335fad6f3391f43562c0780d719ecafeac3b0c81b33874c1d1da4" exitCode=0 Oct 05 21:07:06 crc kubenswrapper[4754]: I1005 21:07:06.697462 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-rvsbj" event={"ID":"b1286ed6-4e15-46e0-b4c3-3b7e3cad6057","Type":"ContainerDied","Data":"c237d2e1ac6335fad6f3391f43562c0780d719ecafeac3b0c81b33874c1d1da4"} Oct 05 21:07:08 crc kubenswrapper[4754]: I1005 21:07:08.651296 4754 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-rvsbj container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" start-of-body= Oct 05 21:07:08 crc kubenswrapper[4754]: I1005 21:07:08.651760 4754 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-rvsbj" podUID="b1286ed6-4e15-46e0-b4c3-3b7e3cad6057" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" Oct 05 21:07:09 crc kubenswrapper[4754]: I1005 21:07:09.529742 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-rvsbj" Oct 05 21:07:09 crc kubenswrapper[4754]: I1005 21:07:09.585906 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-57b4dc6b9f-6kxmr"] Oct 05 21:07:09 crc kubenswrapper[4754]: E1005 21:07:09.586181 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1286ed6-4e15-46e0-b4c3-3b7e3cad6057" containerName="controller-manager" Oct 05 21:07:09 crc kubenswrapper[4754]: I1005 21:07:09.586199 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1286ed6-4e15-46e0-b4c3-3b7e3cad6057" containerName="controller-manager" Oct 05 21:07:09 crc kubenswrapper[4754]: I1005 21:07:09.586331 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="b1286ed6-4e15-46e0-b4c3-3b7e3cad6057" containerName="controller-manager" Oct 05 21:07:09 crc kubenswrapper[4754]: I1005 21:07:09.586739 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-57b4dc6b9f-6kxmr" Oct 05 21:07:09 crc kubenswrapper[4754]: I1005 21:07:09.631714 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-57b4dc6b9f-6kxmr"] Oct 05 21:07:09 crc kubenswrapper[4754]: I1005 21:07:09.639911 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-pnhf9" Oct 05 21:07:09 crc kubenswrapper[4754]: I1005 21:07:09.659079 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b1286ed6-4e15-46e0-b4c3-3b7e3cad6057-config\") pod \"b1286ed6-4e15-46e0-b4c3-3b7e3cad6057\" (UID: \"b1286ed6-4e15-46e0-b4c3-3b7e3cad6057\") " Oct 05 21:07:09 crc kubenswrapper[4754]: I1005 21:07:09.659210 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nnwxh\" (UniqueName: \"kubernetes.io/projected/b1286ed6-4e15-46e0-b4c3-3b7e3cad6057-kube-api-access-nnwxh\") pod \"b1286ed6-4e15-46e0-b4c3-3b7e3cad6057\" (UID: \"b1286ed6-4e15-46e0-b4c3-3b7e3cad6057\") " Oct 05 21:07:09 crc kubenswrapper[4754]: I1005 21:07:09.659254 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b1286ed6-4e15-46e0-b4c3-3b7e3cad6057-proxy-ca-bundles\") pod \"b1286ed6-4e15-46e0-b4c3-3b7e3cad6057\" (UID: \"b1286ed6-4e15-46e0-b4c3-3b7e3cad6057\") " Oct 05 21:07:09 crc kubenswrapper[4754]: I1005 21:07:09.659280 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b1286ed6-4e15-46e0-b4c3-3b7e3cad6057-client-ca\") pod \"b1286ed6-4e15-46e0-b4c3-3b7e3cad6057\" (UID: \"b1286ed6-4e15-46e0-b4c3-3b7e3cad6057\") " Oct 05 21:07:09 crc kubenswrapper[4754]: I1005 21:07:09.659316 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b1286ed6-4e15-46e0-b4c3-3b7e3cad6057-serving-cert\") pod \"b1286ed6-4e15-46e0-b4c3-3b7e3cad6057\" (UID: \"b1286ed6-4e15-46e0-b4c3-3b7e3cad6057\") " Oct 05 21:07:09 crc kubenswrapper[4754]: I1005 21:07:09.665112 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b1286ed6-4e15-46e0-b4c3-3b7e3cad6057-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "b1286ed6-4e15-46e0-b4c3-3b7e3cad6057" (UID: "b1286ed6-4e15-46e0-b4c3-3b7e3cad6057"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 21:07:09 crc kubenswrapper[4754]: I1005 21:07:09.666014 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b1286ed6-4e15-46e0-b4c3-3b7e3cad6057-config" (OuterVolumeSpecName: "config") pod "b1286ed6-4e15-46e0-b4c3-3b7e3cad6057" (UID: "b1286ed6-4e15-46e0-b4c3-3b7e3cad6057"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 21:07:09 crc kubenswrapper[4754]: I1005 21:07:09.668144 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b1286ed6-4e15-46e0-b4c3-3b7e3cad6057-client-ca" (OuterVolumeSpecName: "client-ca") pod "b1286ed6-4e15-46e0-b4c3-3b7e3cad6057" (UID: "b1286ed6-4e15-46e0-b4c3-3b7e3cad6057"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 21:07:09 crc kubenswrapper[4754]: I1005 21:07:09.679290 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b1286ed6-4e15-46e0-b4c3-3b7e3cad6057-kube-api-access-nnwxh" (OuterVolumeSpecName: "kube-api-access-nnwxh") pod "b1286ed6-4e15-46e0-b4c3-3b7e3cad6057" (UID: "b1286ed6-4e15-46e0-b4c3-3b7e3cad6057"). InnerVolumeSpecName "kube-api-access-nnwxh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:07:09 crc kubenswrapper[4754]: I1005 21:07:09.682087 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b1286ed6-4e15-46e0-b4c3-3b7e3cad6057-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "b1286ed6-4e15-46e0-b4c3-3b7e3cad6057" (UID: "b1286ed6-4e15-46e0-b4c3-3b7e3cad6057"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:07:09 crc kubenswrapper[4754]: I1005 21:07:09.732917 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-rvsbj" event={"ID":"b1286ed6-4e15-46e0-b4c3-3b7e3cad6057","Type":"ContainerDied","Data":"2fd8d2336ece00bcebac4481b957ad7bb2adf173e4f8f96087c3c4d72ea26934"} Oct 05 21:07:09 crc kubenswrapper[4754]: I1005 21:07:09.733129 4754 scope.go:117] "RemoveContainer" containerID="c237d2e1ac6335fad6f3391f43562c0780d719ecafeac3b0c81b33874c1d1da4" Oct 05 21:07:09 crc kubenswrapper[4754]: I1005 21:07:09.733316 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-rvsbj" Oct 05 21:07:09 crc kubenswrapper[4754]: I1005 21:07:09.741181 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-pnhf9" Oct 05 21:07:09 crc kubenswrapper[4754]: I1005 21:07:09.741763 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-pnhf9" event={"ID":"1e16d79e-eb1e-4b8b-8f0a-acb370163d17","Type":"ContainerDied","Data":"8d09260517c2f421547a20e3a6a926190586ddec9333c715783763a4f76d958f"} Oct 05 21:07:09 crc kubenswrapper[4754]: I1005 21:07:09.752093 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-zt7sz" event={"ID":"69687995-5b0d-4aae-9551-1b1b9cb2bec3","Type":"ContainerStarted","Data":"7b1b7b6c8181f29730e50b95d9ba45140c7f794afaf05e156a4ca92c15fa861e"} Oct 05 21:07:09 crc kubenswrapper[4754]: I1005 21:07:09.753091 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-zt7sz" Oct 05 21:07:09 crc kubenswrapper[4754]: I1005 21:07:09.754854 4754 generic.go:334] "Generic (PLEG): container finished" podID="dbf73baa-e798-43fa-95de-daec87f5809a" containerID="07386a85ae653ffd494cd5042a93c380581dbdc92d862d26c9312e494e72102f" exitCode=0 Oct 05 21:07:09 crc kubenswrapper[4754]: I1005 21:07:09.754900 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-tfh49" event={"ID":"dbf73baa-e798-43fa-95de-daec87f5809a","Type":"ContainerDied","Data":"07386a85ae653ffd494cd5042a93c380581dbdc92d862d26c9312e494e72102f"} Oct 05 21:07:09 crc kubenswrapper[4754]: I1005 21:07:09.761109 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1e16d79e-eb1e-4b8b-8f0a-acb370163d17-config\") pod \"1e16d79e-eb1e-4b8b-8f0a-acb370163d17\" (UID: \"1e16d79e-eb1e-4b8b-8f0a-acb370163d17\") " Oct 05 21:07:09 crc kubenswrapper[4754]: I1005 21:07:09.761141 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1e16d79e-eb1e-4b8b-8f0a-acb370163d17-client-ca\") pod \"1e16d79e-eb1e-4b8b-8f0a-acb370163d17\" (UID: \"1e16d79e-eb1e-4b8b-8f0a-acb370163d17\") " Oct 05 21:07:09 crc kubenswrapper[4754]: I1005 21:07:09.761179 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9wb4b\" (UniqueName: \"kubernetes.io/projected/1e16d79e-eb1e-4b8b-8f0a-acb370163d17-kube-api-access-9wb4b\") pod \"1e16d79e-eb1e-4b8b-8f0a-acb370163d17\" (UID: \"1e16d79e-eb1e-4b8b-8f0a-acb370163d17\") " Oct 05 21:07:09 crc kubenswrapper[4754]: I1005 21:07:09.761422 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1e16d79e-eb1e-4b8b-8f0a-acb370163d17-serving-cert\") pod \"1e16d79e-eb1e-4b8b-8f0a-acb370163d17\" (UID: \"1e16d79e-eb1e-4b8b-8f0a-acb370163d17\") " Oct 05 21:07:09 crc kubenswrapper[4754]: I1005 21:07:09.761669 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d95c7b41-3b9e-41ec-a760-7f982c6ac65b-config\") pod \"controller-manager-57b4dc6b9f-6kxmr\" (UID: \"d95c7b41-3b9e-41ec-a760-7f982c6ac65b\") " pod="openshift-controller-manager/controller-manager-57b4dc6b9f-6kxmr" Oct 05 21:07:09 crc kubenswrapper[4754]: I1005 21:07:09.761694 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d95c7b41-3b9e-41ec-a760-7f982c6ac65b-serving-cert\") pod \"controller-manager-57b4dc6b9f-6kxmr\" (UID: \"d95c7b41-3b9e-41ec-a760-7f982c6ac65b\") " pod="openshift-controller-manager/controller-manager-57b4dc6b9f-6kxmr" Oct 05 21:07:09 crc kubenswrapper[4754]: I1005 21:07:09.761725 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d95c7b41-3b9e-41ec-a760-7f982c6ac65b-client-ca\") pod \"controller-manager-57b4dc6b9f-6kxmr\" (UID: \"d95c7b41-3b9e-41ec-a760-7f982c6ac65b\") " pod="openshift-controller-manager/controller-manager-57b4dc6b9f-6kxmr" Oct 05 21:07:09 crc kubenswrapper[4754]: I1005 21:07:09.761755 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d95c7b41-3b9e-41ec-a760-7f982c6ac65b-proxy-ca-bundles\") pod \"controller-manager-57b4dc6b9f-6kxmr\" (UID: \"d95c7b41-3b9e-41ec-a760-7f982c6ac65b\") " pod="openshift-controller-manager/controller-manager-57b4dc6b9f-6kxmr" Oct 05 21:07:09 crc kubenswrapper[4754]: I1005 21:07:09.761771 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t9bps\" (UniqueName: \"kubernetes.io/projected/d95c7b41-3b9e-41ec-a760-7f982c6ac65b-kube-api-access-t9bps\") pod \"controller-manager-57b4dc6b9f-6kxmr\" (UID: \"d95c7b41-3b9e-41ec-a760-7f982c6ac65b\") " pod="openshift-controller-manager/controller-manager-57b4dc6b9f-6kxmr" Oct 05 21:07:09 crc kubenswrapper[4754]: I1005 21:07:09.761813 4754 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b1286ed6-4e15-46e0-b4c3-3b7e3cad6057-config\") on node \"crc\" DevicePath \"\"" Oct 05 21:07:09 crc kubenswrapper[4754]: I1005 21:07:09.761824 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nnwxh\" (UniqueName: \"kubernetes.io/projected/b1286ed6-4e15-46e0-b4c3-3b7e3cad6057-kube-api-access-nnwxh\") on node \"crc\" DevicePath \"\"" Oct 05 21:07:09 crc kubenswrapper[4754]: I1005 21:07:09.761833 4754 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b1286ed6-4e15-46e0-b4c3-3b7e3cad6057-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 05 21:07:09 crc kubenswrapper[4754]: I1005 21:07:09.761844 4754 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b1286ed6-4e15-46e0-b4c3-3b7e3cad6057-client-ca\") on node \"crc\" DevicePath \"\"" Oct 05 21:07:09 crc kubenswrapper[4754]: I1005 21:07:09.761854 4754 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b1286ed6-4e15-46e0-b4c3-3b7e3cad6057-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 05 21:07:09 crc kubenswrapper[4754]: I1005 21:07:09.763543 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1e16d79e-eb1e-4b8b-8f0a-acb370163d17-config" (OuterVolumeSpecName: "config") pod "1e16d79e-eb1e-4b8b-8f0a-acb370163d17" (UID: "1e16d79e-eb1e-4b8b-8f0a-acb370163d17"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 21:07:09 crc kubenswrapper[4754]: I1005 21:07:09.764029 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1e16d79e-eb1e-4b8b-8f0a-acb370163d17-client-ca" (OuterVolumeSpecName: "client-ca") pod "1e16d79e-eb1e-4b8b-8f0a-acb370163d17" (UID: "1e16d79e-eb1e-4b8b-8f0a-acb370163d17"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 21:07:09 crc kubenswrapper[4754]: I1005 21:07:09.768613 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e16d79e-eb1e-4b8b-8f0a-acb370163d17-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1e16d79e-eb1e-4b8b-8f0a-acb370163d17" (UID: "1e16d79e-eb1e-4b8b-8f0a-acb370163d17"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:07:09 crc kubenswrapper[4754]: I1005 21:07:09.776694 4754 scope.go:117] "RemoveContainer" containerID="2a2f7d826d242de3ef14c3ee7cf71e919b4aa69848335790746b587439befeb6" Oct 05 21:07:09 crc kubenswrapper[4754]: I1005 21:07:09.780020 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1e16d79e-eb1e-4b8b-8f0a-acb370163d17-kube-api-access-9wb4b" (OuterVolumeSpecName: "kube-api-access-9wb4b") pod "1e16d79e-eb1e-4b8b-8f0a-acb370163d17" (UID: "1e16d79e-eb1e-4b8b-8f0a-acb370163d17"). InnerVolumeSpecName "kube-api-access-9wb4b". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:07:09 crc kubenswrapper[4754]: I1005 21:07:09.784338 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-zt7sz" podStartSLOduration=2.642262413 podStartE2EDuration="11.784314488s" podCreationTimestamp="2025-10-05 21:06:58 +0000 UTC" firstStartedPulling="2025-10-05 21:07:00.151687545 +0000 UTC m=+744.055806255" lastFinishedPulling="2025-10-05 21:07:09.29373962 +0000 UTC m=+753.197858330" observedRunningTime="2025-10-05 21:07:09.778768674 +0000 UTC m=+753.682887384" watchObservedRunningTime="2025-10-05 21:07:09.784314488 +0000 UTC m=+753.688433198" Oct 05 21:07:09 crc kubenswrapper[4754]: I1005 21:07:09.809605 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-rvsbj"] Oct 05 21:07:09 crc kubenswrapper[4754]: I1005 21:07:09.812819 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-rvsbj"] Oct 05 21:07:09 crc kubenswrapper[4754]: I1005 21:07:09.864201 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d95c7b41-3b9e-41ec-a760-7f982c6ac65b-serving-cert\") pod \"controller-manager-57b4dc6b9f-6kxmr\" (UID: \"d95c7b41-3b9e-41ec-a760-7f982c6ac65b\") " pod="openshift-controller-manager/controller-manager-57b4dc6b9f-6kxmr" Oct 05 21:07:09 crc kubenswrapper[4754]: I1005 21:07:09.864282 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d95c7b41-3b9e-41ec-a760-7f982c6ac65b-client-ca\") pod \"controller-manager-57b4dc6b9f-6kxmr\" (UID: \"d95c7b41-3b9e-41ec-a760-7f982c6ac65b\") " pod="openshift-controller-manager/controller-manager-57b4dc6b9f-6kxmr" Oct 05 21:07:09 crc kubenswrapper[4754]: I1005 21:07:09.864334 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d95c7b41-3b9e-41ec-a760-7f982c6ac65b-proxy-ca-bundles\") pod \"controller-manager-57b4dc6b9f-6kxmr\" (UID: \"d95c7b41-3b9e-41ec-a760-7f982c6ac65b\") " pod="openshift-controller-manager/controller-manager-57b4dc6b9f-6kxmr" Oct 05 21:07:09 crc kubenswrapper[4754]: I1005 21:07:09.864353 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t9bps\" (UniqueName: \"kubernetes.io/projected/d95c7b41-3b9e-41ec-a760-7f982c6ac65b-kube-api-access-t9bps\") pod \"controller-manager-57b4dc6b9f-6kxmr\" (UID: \"d95c7b41-3b9e-41ec-a760-7f982c6ac65b\") " pod="openshift-controller-manager/controller-manager-57b4dc6b9f-6kxmr" Oct 05 21:07:09 crc kubenswrapper[4754]: I1005 21:07:09.864426 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d95c7b41-3b9e-41ec-a760-7f982c6ac65b-config\") pod \"controller-manager-57b4dc6b9f-6kxmr\" (UID: \"d95c7b41-3b9e-41ec-a760-7f982c6ac65b\") " pod="openshift-controller-manager/controller-manager-57b4dc6b9f-6kxmr" Oct 05 21:07:09 crc kubenswrapper[4754]: I1005 21:07:09.864469 4754 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1e16d79e-eb1e-4b8b-8f0a-acb370163d17-config\") on node \"crc\" DevicePath \"\"" Oct 05 21:07:09 crc kubenswrapper[4754]: I1005 21:07:09.864481 4754 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1e16d79e-eb1e-4b8b-8f0a-acb370163d17-client-ca\") on node \"crc\" DevicePath \"\"" Oct 05 21:07:09 crc kubenswrapper[4754]: I1005 21:07:09.864646 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9wb4b\" (UniqueName: \"kubernetes.io/projected/1e16d79e-eb1e-4b8b-8f0a-acb370163d17-kube-api-access-9wb4b\") on node \"crc\" DevicePath \"\"" Oct 05 21:07:09 crc kubenswrapper[4754]: I1005 21:07:09.864662 4754 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1e16d79e-eb1e-4b8b-8f0a-acb370163d17-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 05 21:07:09 crc kubenswrapper[4754]: I1005 21:07:09.867258 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d95c7b41-3b9e-41ec-a760-7f982c6ac65b-config\") pod \"controller-manager-57b4dc6b9f-6kxmr\" (UID: \"d95c7b41-3b9e-41ec-a760-7f982c6ac65b\") " pod="openshift-controller-manager/controller-manager-57b4dc6b9f-6kxmr" Oct 05 21:07:09 crc kubenswrapper[4754]: I1005 21:07:09.868425 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d95c7b41-3b9e-41ec-a760-7f982c6ac65b-proxy-ca-bundles\") pod \"controller-manager-57b4dc6b9f-6kxmr\" (UID: \"d95c7b41-3b9e-41ec-a760-7f982c6ac65b\") " pod="openshift-controller-manager/controller-manager-57b4dc6b9f-6kxmr" Oct 05 21:07:09 crc kubenswrapper[4754]: I1005 21:07:09.869033 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d95c7b41-3b9e-41ec-a760-7f982c6ac65b-client-ca\") pod \"controller-manager-57b4dc6b9f-6kxmr\" (UID: \"d95c7b41-3b9e-41ec-a760-7f982c6ac65b\") " pod="openshift-controller-manager/controller-manager-57b4dc6b9f-6kxmr" Oct 05 21:07:09 crc kubenswrapper[4754]: I1005 21:07:09.870132 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d95c7b41-3b9e-41ec-a760-7f982c6ac65b-serving-cert\") pod \"controller-manager-57b4dc6b9f-6kxmr\" (UID: \"d95c7b41-3b9e-41ec-a760-7f982c6ac65b\") " pod="openshift-controller-manager/controller-manager-57b4dc6b9f-6kxmr" Oct 05 21:07:09 crc kubenswrapper[4754]: I1005 21:07:09.885895 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t9bps\" (UniqueName: \"kubernetes.io/projected/d95c7b41-3b9e-41ec-a760-7f982c6ac65b-kube-api-access-t9bps\") pod \"controller-manager-57b4dc6b9f-6kxmr\" (UID: \"d95c7b41-3b9e-41ec-a760-7f982c6ac65b\") " pod="openshift-controller-manager/controller-manager-57b4dc6b9f-6kxmr" Oct 05 21:07:09 crc kubenswrapper[4754]: I1005 21:07:09.922883 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-57b4dc6b9f-6kxmr" Oct 05 21:07:10 crc kubenswrapper[4754]: I1005 21:07:10.078191 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-pnhf9"] Oct 05 21:07:10 crc kubenswrapper[4754]: I1005 21:07:10.082145 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-pnhf9"] Oct 05 21:07:10 crc kubenswrapper[4754]: I1005 21:07:10.355747 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-57b4dc6b9f-6kxmr"] Oct 05 21:07:10 crc kubenswrapper[4754]: W1005 21:07:10.367232 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd95c7b41_3b9e_41ec_a760_7f982c6ac65b.slice/crio-095e96d19546d5af2f965fbe2f19b7172eb9554ce779fb6671a4d22790d51fcc WatchSource:0}: Error finding container 095e96d19546d5af2f965fbe2f19b7172eb9554ce779fb6671a4d22790d51fcc: Status 404 returned error can't find the container with id 095e96d19546d5af2f965fbe2f19b7172eb9554ce779fb6671a4d22790d51fcc Oct 05 21:07:10 crc kubenswrapper[4754]: I1005 21:07:10.465709 4754 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-pnhf9 container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.21:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 05 21:07:10 crc kubenswrapper[4754]: I1005 21:07:10.465768 4754 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-pnhf9" podUID="1e16d79e-eb1e-4b8b-8f0a-acb370163d17" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.21:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 05 21:07:10 crc kubenswrapper[4754]: I1005 21:07:10.766288 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-57b4dc6b9f-6kxmr" event={"ID":"d95c7b41-3b9e-41ec-a760-7f982c6ac65b","Type":"ContainerStarted","Data":"f67bb806de743dc0958d952382ecd4a08fa32a55f11222df30432ac506ddf23d"} Oct 05 21:07:10 crc kubenswrapper[4754]: I1005 21:07:10.766930 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-57b4dc6b9f-6kxmr" Oct 05 21:07:10 crc kubenswrapper[4754]: I1005 21:07:10.766965 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-57b4dc6b9f-6kxmr" event={"ID":"d95c7b41-3b9e-41ec-a760-7f982c6ac65b","Type":"ContainerStarted","Data":"095e96d19546d5af2f965fbe2f19b7172eb9554ce779fb6671a4d22790d51fcc"} Oct 05 21:07:10 crc kubenswrapper[4754]: I1005 21:07:10.771382 4754 generic.go:334] "Generic (PLEG): container finished" podID="dbf73baa-e798-43fa-95de-daec87f5809a" containerID="e3a2dc086d4fcda873bf43442e6f649500ec3f54f14c1c3b972dc185f15a001c" exitCode=0 Oct 05 21:07:10 crc kubenswrapper[4754]: I1005 21:07:10.772534 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-tfh49" event={"ID":"dbf73baa-e798-43fa-95de-daec87f5809a","Type":"ContainerDied","Data":"e3a2dc086d4fcda873bf43442e6f649500ec3f54f14c1c3b972dc185f15a001c"} Oct 05 21:07:10 crc kubenswrapper[4754]: I1005 21:07:10.773575 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-57b4dc6b9f-6kxmr" Oct 05 21:07:10 crc kubenswrapper[4754]: I1005 21:07:10.844238 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-57b4dc6b9f-6kxmr" podStartSLOduration=4.844220058 podStartE2EDuration="4.844220058s" podCreationTimestamp="2025-10-05 21:07:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 21:07:10.809122816 +0000 UTC m=+754.713241526" watchObservedRunningTime="2025-10-05 21:07:10.844220058 +0000 UTC m=+754.748338768" Oct 05 21:07:10 crc kubenswrapper[4754]: I1005 21:07:10.846136 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1e16d79e-eb1e-4b8b-8f0a-acb370163d17" path="/var/lib/kubelet/pods/1e16d79e-eb1e-4b8b-8f0a-acb370163d17/volumes" Oct 05 21:07:10 crc kubenswrapper[4754]: I1005 21:07:10.848152 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b1286ed6-4e15-46e0-b4c3-3b7e3cad6057" path="/var/lib/kubelet/pods/b1286ed6-4e15-46e0-b4c3-3b7e3cad6057/volumes" Oct 05 21:07:11 crc kubenswrapper[4754]: I1005 21:07:11.780259 4754 generic.go:334] "Generic (PLEG): container finished" podID="dbf73baa-e798-43fa-95de-daec87f5809a" containerID="2b20482431ffe8247ee34b5e502b05f88fd35ebdf4569998506f70cc20949d0b" exitCode=0 Oct 05 21:07:11 crc kubenswrapper[4754]: I1005 21:07:11.780346 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-tfh49" event={"ID":"dbf73baa-e798-43fa-95de-daec87f5809a","Type":"ContainerDied","Data":"2b20482431ffe8247ee34b5e502b05f88fd35ebdf4569998506f70cc20949d0b"} Oct 05 21:07:12 crc kubenswrapper[4754]: I1005 21:07:12.528678 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-56f9db776c-7hxmn"] Oct 05 21:07:12 crc kubenswrapper[4754]: E1005 21:07:12.529599 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e16d79e-eb1e-4b8b-8f0a-acb370163d17" containerName="route-controller-manager" Oct 05 21:07:12 crc kubenswrapper[4754]: I1005 21:07:12.529619 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e16d79e-eb1e-4b8b-8f0a-acb370163d17" containerName="route-controller-manager" Oct 05 21:07:12 crc kubenswrapper[4754]: I1005 21:07:12.529755 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e16d79e-eb1e-4b8b-8f0a-acb370163d17" containerName="route-controller-manager" Oct 05 21:07:12 crc kubenswrapper[4754]: I1005 21:07:12.530159 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-56f9db776c-7hxmn" Oct 05 21:07:12 crc kubenswrapper[4754]: I1005 21:07:12.533719 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Oct 05 21:07:12 crc kubenswrapper[4754]: I1005 21:07:12.533941 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Oct 05 21:07:12 crc kubenswrapper[4754]: I1005 21:07:12.534080 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Oct 05 21:07:12 crc kubenswrapper[4754]: I1005 21:07:12.534616 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Oct 05 21:07:12 crc kubenswrapper[4754]: I1005 21:07:12.534785 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Oct 05 21:07:12 crc kubenswrapper[4754]: I1005 21:07:12.535059 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Oct 05 21:07:12 crc kubenswrapper[4754]: I1005 21:07:12.552719 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-56f9db776c-7hxmn"] Oct 05 21:07:12 crc kubenswrapper[4754]: I1005 21:07:12.609005 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/62322590-38cc-46a0-9896-4eb2c2da7159-client-ca\") pod \"route-controller-manager-56f9db776c-7hxmn\" (UID: \"62322590-38cc-46a0-9896-4eb2c2da7159\") " pod="openshift-route-controller-manager/route-controller-manager-56f9db776c-7hxmn" Oct 05 21:07:12 crc kubenswrapper[4754]: I1005 21:07:12.609066 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6v68p\" (UniqueName: \"kubernetes.io/projected/62322590-38cc-46a0-9896-4eb2c2da7159-kube-api-access-6v68p\") pod \"route-controller-manager-56f9db776c-7hxmn\" (UID: \"62322590-38cc-46a0-9896-4eb2c2da7159\") " pod="openshift-route-controller-manager/route-controller-manager-56f9db776c-7hxmn" Oct 05 21:07:12 crc kubenswrapper[4754]: I1005 21:07:12.609126 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/62322590-38cc-46a0-9896-4eb2c2da7159-config\") pod \"route-controller-manager-56f9db776c-7hxmn\" (UID: \"62322590-38cc-46a0-9896-4eb2c2da7159\") " pod="openshift-route-controller-manager/route-controller-manager-56f9db776c-7hxmn" Oct 05 21:07:12 crc kubenswrapper[4754]: I1005 21:07:12.609167 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/62322590-38cc-46a0-9896-4eb2c2da7159-serving-cert\") pod \"route-controller-manager-56f9db776c-7hxmn\" (UID: \"62322590-38cc-46a0-9896-4eb2c2da7159\") " pod="openshift-route-controller-manager/route-controller-manager-56f9db776c-7hxmn" Oct 05 21:07:12 crc kubenswrapper[4754]: I1005 21:07:12.710201 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/62322590-38cc-46a0-9896-4eb2c2da7159-client-ca\") pod \"route-controller-manager-56f9db776c-7hxmn\" (UID: \"62322590-38cc-46a0-9896-4eb2c2da7159\") " pod="openshift-route-controller-manager/route-controller-manager-56f9db776c-7hxmn" Oct 05 21:07:12 crc kubenswrapper[4754]: I1005 21:07:12.710254 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6v68p\" (UniqueName: \"kubernetes.io/projected/62322590-38cc-46a0-9896-4eb2c2da7159-kube-api-access-6v68p\") pod \"route-controller-manager-56f9db776c-7hxmn\" (UID: \"62322590-38cc-46a0-9896-4eb2c2da7159\") " pod="openshift-route-controller-manager/route-controller-manager-56f9db776c-7hxmn" Oct 05 21:07:12 crc kubenswrapper[4754]: I1005 21:07:12.710307 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/62322590-38cc-46a0-9896-4eb2c2da7159-config\") pod \"route-controller-manager-56f9db776c-7hxmn\" (UID: \"62322590-38cc-46a0-9896-4eb2c2da7159\") " pod="openshift-route-controller-manager/route-controller-manager-56f9db776c-7hxmn" Oct 05 21:07:12 crc kubenswrapper[4754]: I1005 21:07:12.710354 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/62322590-38cc-46a0-9896-4eb2c2da7159-serving-cert\") pod \"route-controller-manager-56f9db776c-7hxmn\" (UID: \"62322590-38cc-46a0-9896-4eb2c2da7159\") " pod="openshift-route-controller-manager/route-controller-manager-56f9db776c-7hxmn" Oct 05 21:07:12 crc kubenswrapper[4754]: I1005 21:07:12.711385 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/62322590-38cc-46a0-9896-4eb2c2da7159-client-ca\") pod \"route-controller-manager-56f9db776c-7hxmn\" (UID: \"62322590-38cc-46a0-9896-4eb2c2da7159\") " pod="openshift-route-controller-manager/route-controller-manager-56f9db776c-7hxmn" Oct 05 21:07:12 crc kubenswrapper[4754]: I1005 21:07:12.711447 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/62322590-38cc-46a0-9896-4eb2c2da7159-config\") pod \"route-controller-manager-56f9db776c-7hxmn\" (UID: \"62322590-38cc-46a0-9896-4eb2c2da7159\") " pod="openshift-route-controller-manager/route-controller-manager-56f9db776c-7hxmn" Oct 05 21:07:12 crc kubenswrapper[4754]: I1005 21:07:12.717103 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/62322590-38cc-46a0-9896-4eb2c2da7159-serving-cert\") pod \"route-controller-manager-56f9db776c-7hxmn\" (UID: \"62322590-38cc-46a0-9896-4eb2c2da7159\") " pod="openshift-route-controller-manager/route-controller-manager-56f9db776c-7hxmn" Oct 05 21:07:12 crc kubenswrapper[4754]: I1005 21:07:12.731476 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6v68p\" (UniqueName: \"kubernetes.io/projected/62322590-38cc-46a0-9896-4eb2c2da7159-kube-api-access-6v68p\") pod \"route-controller-manager-56f9db776c-7hxmn\" (UID: \"62322590-38cc-46a0-9896-4eb2c2da7159\") " pod="openshift-route-controller-manager/route-controller-manager-56f9db776c-7hxmn" Oct 05 21:07:12 crc kubenswrapper[4754]: I1005 21:07:12.805567 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-tfh49" event={"ID":"dbf73baa-e798-43fa-95de-daec87f5809a","Type":"ContainerStarted","Data":"f4e869cdd2ca3779ee4d6631dcdbb797e88398ec1138055ce7c637c49baee8d7"} Oct 05 21:07:12 crc kubenswrapper[4754]: I1005 21:07:12.805640 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-tfh49" event={"ID":"dbf73baa-e798-43fa-95de-daec87f5809a","Type":"ContainerStarted","Data":"19181dac8385f0d5fa862a503c3deddc167b1a127b842fb41b59c2580e99f680"} Oct 05 21:07:12 crc kubenswrapper[4754]: I1005 21:07:12.805690 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-tfh49" event={"ID":"dbf73baa-e798-43fa-95de-daec87f5809a","Type":"ContainerStarted","Data":"508ed6068d40a523dfa24a9e299ee902e47e4d11dbe90feb9b99db5e688f95cd"} Oct 05 21:07:12 crc kubenswrapper[4754]: I1005 21:07:12.805707 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-tfh49" event={"ID":"dbf73baa-e798-43fa-95de-daec87f5809a","Type":"ContainerStarted","Data":"e157959dd617592bd56fea10cdcb58b984e7114829ebc9e649f18a08883d5f9c"} Oct 05 21:07:12 crc kubenswrapper[4754]: I1005 21:07:12.852438 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-56f9db776c-7hxmn" Oct 05 21:07:13 crc kubenswrapper[4754]: I1005 21:07:13.370970 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-56f9db776c-7hxmn"] Oct 05 21:07:13 crc kubenswrapper[4754]: W1005 21:07:13.403273 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod62322590_38cc_46a0_9896_4eb2c2da7159.slice/crio-cb22265c98769fc3ab17fdabd8e4aed1eae9234e533b50119adc5efcb907d333 WatchSource:0}: Error finding container cb22265c98769fc3ab17fdabd8e4aed1eae9234e533b50119adc5efcb907d333: Status 404 returned error can't find the container with id cb22265c98769fc3ab17fdabd8e4aed1eae9234e533b50119adc5efcb907d333 Oct 05 21:07:13 crc kubenswrapper[4754]: I1005 21:07:13.814633 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-56f9db776c-7hxmn" event={"ID":"62322590-38cc-46a0-9896-4eb2c2da7159","Type":"ContainerStarted","Data":"55b19e512d9b95c8a50f97cb3add05a43cdaab18d45687f6c1b08de62c7eeb5b"} Oct 05 21:07:13 crc kubenswrapper[4754]: I1005 21:07:13.814715 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-56f9db776c-7hxmn" event={"ID":"62322590-38cc-46a0-9896-4eb2c2da7159","Type":"ContainerStarted","Data":"cb22265c98769fc3ab17fdabd8e4aed1eae9234e533b50119adc5efcb907d333"} Oct 05 21:07:13 crc kubenswrapper[4754]: I1005 21:07:13.821926 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-tfh49" event={"ID":"dbf73baa-e798-43fa-95de-daec87f5809a","Type":"ContainerStarted","Data":"af2e438a693a22f863658c5f547723d0533a0111cf5847caf801bf0529e76f86"} Oct 05 21:07:13 crc kubenswrapper[4754]: I1005 21:07:13.821977 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-tfh49" event={"ID":"dbf73baa-e798-43fa-95de-daec87f5809a","Type":"ContainerStarted","Data":"7ab25d21fa59ff553b5f0d44409232c63be29c49bc7881c211a9fe37a39d9e4c"} Oct 05 21:07:13 crc kubenswrapper[4754]: I1005 21:07:13.822222 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-tfh49" Oct 05 21:07:13 crc kubenswrapper[4754]: I1005 21:07:13.856285 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-tfh49" podStartSLOduration=6.699759842 podStartE2EDuration="15.856259393s" podCreationTimestamp="2025-10-05 21:06:58 +0000 UTC" firstStartedPulling="2025-10-05 21:07:00.219985808 +0000 UTC m=+744.124104518" lastFinishedPulling="2025-10-05 21:07:09.376485359 +0000 UTC m=+753.280604069" observedRunningTime="2025-10-05 21:07:13.849722213 +0000 UTC m=+757.753840923" watchObservedRunningTime="2025-10-05 21:07:13.856259393 +0000 UTC m=+757.760378123" Oct 05 21:07:14 crc kubenswrapper[4754]: I1005 21:07:14.858289 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-56f9db776c-7hxmn" podStartSLOduration=8.858263579 podStartE2EDuration="8.858263579s" podCreationTimestamp="2025-10-05 21:07:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 21:07:14.857295314 +0000 UTC m=+758.761414054" watchObservedRunningTime="2025-10-05 21:07:14.858263579 +0000 UTC m=+758.762382289" Oct 05 21:07:14 crc kubenswrapper[4754]: I1005 21:07:14.948440 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-tfh49" Oct 05 21:07:14 crc kubenswrapper[4754]: I1005 21:07:14.995971 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-tfh49" Oct 05 21:07:17 crc kubenswrapper[4754]: I1005 21:07:17.102611 4754 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 05 21:07:19 crc kubenswrapper[4754]: I1005 21:07:19.469831 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-68d546b9d8-78nmb" Oct 05 21:07:19 crc kubenswrapper[4754]: I1005 21:07:19.937215 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-zt7sz" Oct 05 21:07:20 crc kubenswrapper[4754]: I1005 21:07:20.959451 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-hkvd8" Oct 05 21:07:22 crc kubenswrapper[4754]: I1005 21:07:22.853908 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-56f9db776c-7hxmn" Oct 05 21:07:22 crc kubenswrapper[4754]: I1005 21:07:22.864862 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-56f9db776c-7hxmn" Oct 05 21:07:23 crc kubenswrapper[4754]: I1005 21:07:23.964461 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-wdfds"] Oct 05 21:07:23 crc kubenswrapper[4754]: I1005 21:07:23.966310 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-wdfds" Oct 05 21:07:23 crc kubenswrapper[4754]: I1005 21:07:23.968826 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Oct 05 21:07:23 crc kubenswrapper[4754]: I1005 21:07:23.971608 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Oct 05 21:07:24 crc kubenswrapper[4754]: I1005 21:07:24.005768 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-wdfds"] Oct 05 21:07:24 crc kubenswrapper[4754]: I1005 21:07:24.092035 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mkpwz\" (UniqueName: \"kubernetes.io/projected/62280402-3bee-4592-b313-26cd43649a80-kube-api-access-mkpwz\") pod \"openstack-operator-index-wdfds\" (UID: \"62280402-3bee-4592-b313-26cd43649a80\") " pod="openstack-operators/openstack-operator-index-wdfds" Oct 05 21:07:24 crc kubenswrapper[4754]: I1005 21:07:24.193066 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mkpwz\" (UniqueName: \"kubernetes.io/projected/62280402-3bee-4592-b313-26cd43649a80-kube-api-access-mkpwz\") pod \"openstack-operator-index-wdfds\" (UID: \"62280402-3bee-4592-b313-26cd43649a80\") " pod="openstack-operators/openstack-operator-index-wdfds" Oct 05 21:07:24 crc kubenswrapper[4754]: I1005 21:07:24.222354 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mkpwz\" (UniqueName: \"kubernetes.io/projected/62280402-3bee-4592-b313-26cd43649a80-kube-api-access-mkpwz\") pod \"openstack-operator-index-wdfds\" (UID: \"62280402-3bee-4592-b313-26cd43649a80\") " pod="openstack-operators/openstack-operator-index-wdfds" Oct 05 21:07:24 crc kubenswrapper[4754]: I1005 21:07:24.299058 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-wdfds" Oct 05 21:07:24 crc kubenswrapper[4754]: I1005 21:07:24.799703 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-wdfds"] Oct 05 21:07:24 crc kubenswrapper[4754]: W1005 21:07:24.805339 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod62280402_3bee_4592_b313_26cd43649a80.slice/crio-553edd142a75614ea0ee631ca4117d7c471157e7b252b2771a4e5ae74684ccef WatchSource:0}: Error finding container 553edd142a75614ea0ee631ca4117d7c471157e7b252b2771a4e5ae74684ccef: Status 404 returned error can't find the container with id 553edd142a75614ea0ee631ca4117d7c471157e7b252b2771a4e5ae74684ccef Oct 05 21:07:24 crc kubenswrapper[4754]: I1005 21:07:24.903336 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-wdfds" event={"ID":"62280402-3bee-4592-b313-26cd43649a80","Type":"ContainerStarted","Data":"553edd142a75614ea0ee631ca4117d7c471157e7b252b2771a4e5ae74684ccef"} Oct 05 21:07:26 crc kubenswrapper[4754]: I1005 21:07:26.323387 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-wdfds"] Oct 05 21:07:26 crc kubenswrapper[4754]: I1005 21:07:26.738731 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-pgjjb"] Oct 05 21:07:26 crc kubenswrapper[4754]: I1005 21:07:26.752338 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-pgjjb"] Oct 05 21:07:26 crc kubenswrapper[4754]: I1005 21:07:26.752518 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-pgjjb" Oct 05 21:07:26 crc kubenswrapper[4754]: I1005 21:07:26.756004 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-9kxkp" Oct 05 21:07:26 crc kubenswrapper[4754]: I1005 21:07:26.843313 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rjngt\" (UniqueName: \"kubernetes.io/projected/506033e6-79e8-4b4e-acb8-97e63316a15b-kube-api-access-rjngt\") pod \"openstack-operator-index-pgjjb\" (UID: \"506033e6-79e8-4b4e-acb8-97e63316a15b\") " pod="openstack-operators/openstack-operator-index-pgjjb" Oct 05 21:07:26 crc kubenswrapper[4754]: I1005 21:07:26.946203 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rjngt\" (UniqueName: \"kubernetes.io/projected/506033e6-79e8-4b4e-acb8-97e63316a15b-kube-api-access-rjngt\") pod \"openstack-operator-index-pgjjb\" (UID: \"506033e6-79e8-4b4e-acb8-97e63316a15b\") " pod="openstack-operators/openstack-operator-index-pgjjb" Oct 05 21:07:26 crc kubenswrapper[4754]: I1005 21:07:26.972874 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rjngt\" (UniqueName: \"kubernetes.io/projected/506033e6-79e8-4b4e-acb8-97e63316a15b-kube-api-access-rjngt\") pod \"openstack-operator-index-pgjjb\" (UID: \"506033e6-79e8-4b4e-acb8-97e63316a15b\") " pod="openstack-operators/openstack-operator-index-pgjjb" Oct 05 21:07:27 crc kubenswrapper[4754]: I1005 21:07:27.088270 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-pgjjb" Oct 05 21:07:27 crc kubenswrapper[4754]: I1005 21:07:27.554323 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-pgjjb"] Oct 05 21:07:27 crc kubenswrapper[4754]: I1005 21:07:27.930236 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-pgjjb" event={"ID":"506033e6-79e8-4b4e-acb8-97e63316a15b","Type":"ContainerStarted","Data":"cfde7deafc501d075d3a1b8528e70dcac51181dbbe85b93d7345c7300fe2eb24"} Oct 05 21:07:29 crc kubenswrapper[4754]: I1005 21:07:29.945798 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-pgjjb" event={"ID":"506033e6-79e8-4b4e-acb8-97e63316a15b","Type":"ContainerStarted","Data":"ac5e786448c135a022c3da5c2d4c9710d929d8630f0b00aa78487c14810e1a0a"} Oct 05 21:07:29 crc kubenswrapper[4754]: I1005 21:07:29.949032 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-wdfds" event={"ID":"62280402-3bee-4592-b313-26cd43649a80","Type":"ContainerStarted","Data":"c4e133f17f24e6d763a4ddacde076f6cde5537a054926027fb42106f1e72af1d"} Oct 05 21:07:29 crc kubenswrapper[4754]: I1005 21:07:29.949156 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-wdfds" podUID="62280402-3bee-4592-b313-26cd43649a80" containerName="registry-server" containerID="cri-o://c4e133f17f24e6d763a4ddacde076f6cde5537a054926027fb42106f1e72af1d" gracePeriod=2 Oct 05 21:07:29 crc kubenswrapper[4754]: I1005 21:07:29.954181 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-tfh49" Oct 05 21:07:29 crc kubenswrapper[4754]: I1005 21:07:29.973839 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-pgjjb" podStartSLOduration=2.494479211 podStartE2EDuration="3.97381228s" podCreationTimestamp="2025-10-05 21:07:26 +0000 UTC" firstStartedPulling="2025-10-05 21:07:27.883639681 +0000 UTC m=+771.787758421" lastFinishedPulling="2025-10-05 21:07:29.36297274 +0000 UTC m=+773.267091490" observedRunningTime="2025-10-05 21:07:29.969644492 +0000 UTC m=+773.873763232" watchObservedRunningTime="2025-10-05 21:07:29.97381228 +0000 UTC m=+773.877931000" Oct 05 21:07:30 crc kubenswrapper[4754]: I1005 21:07:30.023671 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-wdfds" podStartSLOduration=2.470462654 podStartE2EDuration="7.023650684s" podCreationTimestamp="2025-10-05 21:07:23 +0000 UTC" firstStartedPulling="2025-10-05 21:07:24.80860389 +0000 UTC m=+768.712722600" lastFinishedPulling="2025-10-05 21:07:29.36179188 +0000 UTC m=+773.265910630" observedRunningTime="2025-10-05 21:07:30.020521273 +0000 UTC m=+773.924639983" watchObservedRunningTime="2025-10-05 21:07:30.023650684 +0000 UTC m=+773.927769394" Oct 05 21:07:30 crc kubenswrapper[4754]: I1005 21:07:30.438800 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-wdfds" Oct 05 21:07:30 crc kubenswrapper[4754]: I1005 21:07:30.603234 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mkpwz\" (UniqueName: \"kubernetes.io/projected/62280402-3bee-4592-b313-26cd43649a80-kube-api-access-mkpwz\") pod \"62280402-3bee-4592-b313-26cd43649a80\" (UID: \"62280402-3bee-4592-b313-26cd43649a80\") " Oct 05 21:07:30 crc kubenswrapper[4754]: I1005 21:07:30.617590 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/62280402-3bee-4592-b313-26cd43649a80-kube-api-access-mkpwz" (OuterVolumeSpecName: "kube-api-access-mkpwz") pod "62280402-3bee-4592-b313-26cd43649a80" (UID: "62280402-3bee-4592-b313-26cd43649a80"). InnerVolumeSpecName "kube-api-access-mkpwz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:07:30 crc kubenswrapper[4754]: I1005 21:07:30.705440 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mkpwz\" (UniqueName: \"kubernetes.io/projected/62280402-3bee-4592-b313-26cd43649a80-kube-api-access-mkpwz\") on node \"crc\" DevicePath \"\"" Oct 05 21:07:30 crc kubenswrapper[4754]: I1005 21:07:30.960407 4754 generic.go:334] "Generic (PLEG): container finished" podID="62280402-3bee-4592-b313-26cd43649a80" containerID="c4e133f17f24e6d763a4ddacde076f6cde5537a054926027fb42106f1e72af1d" exitCode=0 Oct 05 21:07:30 crc kubenswrapper[4754]: I1005 21:07:30.960525 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-wdfds" Oct 05 21:07:30 crc kubenswrapper[4754]: I1005 21:07:30.960569 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-wdfds" event={"ID":"62280402-3bee-4592-b313-26cd43649a80","Type":"ContainerDied","Data":"c4e133f17f24e6d763a4ddacde076f6cde5537a054926027fb42106f1e72af1d"} Oct 05 21:07:30 crc kubenswrapper[4754]: I1005 21:07:30.960607 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-wdfds" event={"ID":"62280402-3bee-4592-b313-26cd43649a80","Type":"ContainerDied","Data":"553edd142a75614ea0ee631ca4117d7c471157e7b252b2771a4e5ae74684ccef"} Oct 05 21:07:30 crc kubenswrapper[4754]: I1005 21:07:30.960631 4754 scope.go:117] "RemoveContainer" containerID="c4e133f17f24e6d763a4ddacde076f6cde5537a054926027fb42106f1e72af1d" Oct 05 21:07:30 crc kubenswrapper[4754]: I1005 21:07:30.993400 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-wdfds"] Oct 05 21:07:30 crc kubenswrapper[4754]: I1005 21:07:30.999321 4754 scope.go:117] "RemoveContainer" containerID="c4e133f17f24e6d763a4ddacde076f6cde5537a054926027fb42106f1e72af1d" Oct 05 21:07:30 crc kubenswrapper[4754]: I1005 21:07:30.999608 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-wdfds"] Oct 05 21:07:30 crc kubenswrapper[4754]: E1005 21:07:30.999891 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c4e133f17f24e6d763a4ddacde076f6cde5537a054926027fb42106f1e72af1d\": container with ID starting with c4e133f17f24e6d763a4ddacde076f6cde5537a054926027fb42106f1e72af1d not found: ID does not exist" containerID="c4e133f17f24e6d763a4ddacde076f6cde5537a054926027fb42106f1e72af1d" Oct 05 21:07:31 crc kubenswrapper[4754]: I1005 21:07:30.999932 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c4e133f17f24e6d763a4ddacde076f6cde5537a054926027fb42106f1e72af1d"} err="failed to get container status \"c4e133f17f24e6d763a4ddacde076f6cde5537a054926027fb42106f1e72af1d\": rpc error: code = NotFound desc = could not find container \"c4e133f17f24e6d763a4ddacde076f6cde5537a054926027fb42106f1e72af1d\": container with ID starting with c4e133f17f24e6d763a4ddacde076f6cde5537a054926027fb42106f1e72af1d not found: ID does not exist" Oct 05 21:07:32 crc kubenswrapper[4754]: I1005 21:07:32.851352 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="62280402-3bee-4592-b313-26cd43649a80" path="/var/lib/kubelet/pods/62280402-3bee-4592-b313-26cd43649a80/volumes" Oct 05 21:07:35 crc kubenswrapper[4754]: I1005 21:07:35.246141 4754 patch_prober.go:28] interesting pod/machine-config-daemon-b2h9k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 21:07:35 crc kubenswrapper[4754]: I1005 21:07:35.246317 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 21:07:35 crc kubenswrapper[4754]: I1005 21:07:35.246413 4754 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" Oct 05 21:07:35 crc kubenswrapper[4754]: I1005 21:07:35.247268 4754 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"42efe442464282aff7d7a4c1a4006bea20a662c700019a0224e31e9c8b2dfd60"} pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 05 21:07:35 crc kubenswrapper[4754]: I1005 21:07:35.247373 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" containerName="machine-config-daemon" containerID="cri-o://42efe442464282aff7d7a4c1a4006bea20a662c700019a0224e31e9c8b2dfd60" gracePeriod=600 Oct 05 21:07:36 crc kubenswrapper[4754]: I1005 21:07:36.018809 4754 generic.go:334] "Generic (PLEG): container finished" podID="be95a413-6a59-45b4-84b7-b43cae694a26" containerID="42efe442464282aff7d7a4c1a4006bea20a662c700019a0224e31e9c8b2dfd60" exitCode=0 Oct 05 21:07:36 crc kubenswrapper[4754]: I1005 21:07:36.018868 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" event={"ID":"be95a413-6a59-45b4-84b7-b43cae694a26","Type":"ContainerDied","Data":"42efe442464282aff7d7a4c1a4006bea20a662c700019a0224e31e9c8b2dfd60"} Oct 05 21:07:36 crc kubenswrapper[4754]: I1005 21:07:36.018939 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" event={"ID":"be95a413-6a59-45b4-84b7-b43cae694a26","Type":"ContainerStarted","Data":"204a9d4b26fceb896f0714f752c6bee1970d2349381ac50fae1a7322b50bafde"} Oct 05 21:07:36 crc kubenswrapper[4754]: I1005 21:07:36.018982 4754 scope.go:117] "RemoveContainer" containerID="0b09d5fa150c23a262d03541134e26faa0bf59c28b4299e50ad59b83bfa2351d" Oct 05 21:07:37 crc kubenswrapper[4754]: I1005 21:07:37.089072 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-pgjjb" Oct 05 21:07:37 crc kubenswrapper[4754]: I1005 21:07:37.089656 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-pgjjb" Oct 05 21:07:37 crc kubenswrapper[4754]: I1005 21:07:37.126635 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-pgjjb" Oct 05 21:07:38 crc kubenswrapper[4754]: I1005 21:07:38.088049 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-pgjjb" Oct 05 21:07:43 crc kubenswrapper[4754]: I1005 21:07:43.982423 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/aded1cecd2f966ec868aa0a6ebe9c364f1aced79b6d6348d02479f11239rr6r"] Oct 05 21:07:43 crc kubenswrapper[4754]: E1005 21:07:43.983609 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62280402-3bee-4592-b313-26cd43649a80" containerName="registry-server" Oct 05 21:07:43 crc kubenswrapper[4754]: I1005 21:07:43.983625 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="62280402-3bee-4592-b313-26cd43649a80" containerName="registry-server" Oct 05 21:07:43 crc kubenswrapper[4754]: I1005 21:07:43.983775 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="62280402-3bee-4592-b313-26cd43649a80" containerName="registry-server" Oct 05 21:07:43 crc kubenswrapper[4754]: I1005 21:07:43.984756 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/aded1cecd2f966ec868aa0a6ebe9c364f1aced79b6d6348d02479f11239rr6r" Oct 05 21:07:43 crc kubenswrapper[4754]: I1005 21:07:43.994261 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-hpgfk" Oct 05 21:07:43 crc kubenswrapper[4754]: I1005 21:07:43.997069 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/aded1cecd2f966ec868aa0a6ebe9c364f1aced79b6d6348d02479f11239rr6r"] Oct 05 21:07:44 crc kubenswrapper[4754]: I1005 21:07:44.048059 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5f6867a7-6e2b-4f7a-bcdd-a0ff2035eee8-util\") pod \"aded1cecd2f966ec868aa0a6ebe9c364f1aced79b6d6348d02479f11239rr6r\" (UID: \"5f6867a7-6e2b-4f7a-bcdd-a0ff2035eee8\") " pod="openstack-operators/aded1cecd2f966ec868aa0a6ebe9c364f1aced79b6d6348d02479f11239rr6r" Oct 05 21:07:44 crc kubenswrapper[4754]: I1005 21:07:44.048444 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zdn7f\" (UniqueName: \"kubernetes.io/projected/5f6867a7-6e2b-4f7a-bcdd-a0ff2035eee8-kube-api-access-zdn7f\") pod \"aded1cecd2f966ec868aa0a6ebe9c364f1aced79b6d6348d02479f11239rr6r\" (UID: \"5f6867a7-6e2b-4f7a-bcdd-a0ff2035eee8\") " pod="openstack-operators/aded1cecd2f966ec868aa0a6ebe9c364f1aced79b6d6348d02479f11239rr6r" Oct 05 21:07:44 crc kubenswrapper[4754]: I1005 21:07:44.048569 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5f6867a7-6e2b-4f7a-bcdd-a0ff2035eee8-bundle\") pod \"aded1cecd2f966ec868aa0a6ebe9c364f1aced79b6d6348d02479f11239rr6r\" (UID: \"5f6867a7-6e2b-4f7a-bcdd-a0ff2035eee8\") " pod="openstack-operators/aded1cecd2f966ec868aa0a6ebe9c364f1aced79b6d6348d02479f11239rr6r" Oct 05 21:07:44 crc kubenswrapper[4754]: I1005 21:07:44.149844 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zdn7f\" (UniqueName: \"kubernetes.io/projected/5f6867a7-6e2b-4f7a-bcdd-a0ff2035eee8-kube-api-access-zdn7f\") pod \"aded1cecd2f966ec868aa0a6ebe9c364f1aced79b6d6348d02479f11239rr6r\" (UID: \"5f6867a7-6e2b-4f7a-bcdd-a0ff2035eee8\") " pod="openstack-operators/aded1cecd2f966ec868aa0a6ebe9c364f1aced79b6d6348d02479f11239rr6r" Oct 05 21:07:44 crc kubenswrapper[4754]: I1005 21:07:44.150182 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5f6867a7-6e2b-4f7a-bcdd-a0ff2035eee8-bundle\") pod \"aded1cecd2f966ec868aa0a6ebe9c364f1aced79b6d6348d02479f11239rr6r\" (UID: \"5f6867a7-6e2b-4f7a-bcdd-a0ff2035eee8\") " pod="openstack-operators/aded1cecd2f966ec868aa0a6ebe9c364f1aced79b6d6348d02479f11239rr6r" Oct 05 21:07:44 crc kubenswrapper[4754]: I1005 21:07:44.150867 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5f6867a7-6e2b-4f7a-bcdd-a0ff2035eee8-util\") pod \"aded1cecd2f966ec868aa0a6ebe9c364f1aced79b6d6348d02479f11239rr6r\" (UID: \"5f6867a7-6e2b-4f7a-bcdd-a0ff2035eee8\") " pod="openstack-operators/aded1cecd2f966ec868aa0a6ebe9c364f1aced79b6d6348d02479f11239rr6r" Oct 05 21:07:44 crc kubenswrapper[4754]: I1005 21:07:44.150786 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5f6867a7-6e2b-4f7a-bcdd-a0ff2035eee8-bundle\") pod \"aded1cecd2f966ec868aa0a6ebe9c364f1aced79b6d6348d02479f11239rr6r\" (UID: \"5f6867a7-6e2b-4f7a-bcdd-a0ff2035eee8\") " pod="openstack-operators/aded1cecd2f966ec868aa0a6ebe9c364f1aced79b6d6348d02479f11239rr6r" Oct 05 21:07:44 crc kubenswrapper[4754]: I1005 21:07:44.151174 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5f6867a7-6e2b-4f7a-bcdd-a0ff2035eee8-util\") pod \"aded1cecd2f966ec868aa0a6ebe9c364f1aced79b6d6348d02479f11239rr6r\" (UID: \"5f6867a7-6e2b-4f7a-bcdd-a0ff2035eee8\") " pod="openstack-operators/aded1cecd2f966ec868aa0a6ebe9c364f1aced79b6d6348d02479f11239rr6r" Oct 05 21:07:44 crc kubenswrapper[4754]: I1005 21:07:44.179083 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zdn7f\" (UniqueName: \"kubernetes.io/projected/5f6867a7-6e2b-4f7a-bcdd-a0ff2035eee8-kube-api-access-zdn7f\") pod \"aded1cecd2f966ec868aa0a6ebe9c364f1aced79b6d6348d02479f11239rr6r\" (UID: \"5f6867a7-6e2b-4f7a-bcdd-a0ff2035eee8\") " pod="openstack-operators/aded1cecd2f966ec868aa0a6ebe9c364f1aced79b6d6348d02479f11239rr6r" Oct 05 21:07:44 crc kubenswrapper[4754]: I1005 21:07:44.304518 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/aded1cecd2f966ec868aa0a6ebe9c364f1aced79b6d6348d02479f11239rr6r" Oct 05 21:07:44 crc kubenswrapper[4754]: I1005 21:07:44.761088 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/aded1cecd2f966ec868aa0a6ebe9c364f1aced79b6d6348d02479f11239rr6r"] Oct 05 21:07:44 crc kubenswrapper[4754]: W1005 21:07:44.777780 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5f6867a7_6e2b_4f7a_bcdd_a0ff2035eee8.slice/crio-7b2d10d2d64393e96db256939eb95bd8e4c5233fd52ba2323bf4d1f2632bdde5 WatchSource:0}: Error finding container 7b2d10d2d64393e96db256939eb95bd8e4c5233fd52ba2323bf4d1f2632bdde5: Status 404 returned error can't find the container with id 7b2d10d2d64393e96db256939eb95bd8e4c5233fd52ba2323bf4d1f2632bdde5 Oct 05 21:07:45 crc kubenswrapper[4754]: I1005 21:07:45.097687 4754 generic.go:334] "Generic (PLEG): container finished" podID="5f6867a7-6e2b-4f7a-bcdd-a0ff2035eee8" containerID="ef250cafb07cdf8b8943c3060c96e3415d7a4926716a495954dad5782b4032a4" exitCode=0 Oct 05 21:07:45 crc kubenswrapper[4754]: I1005 21:07:45.097761 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/aded1cecd2f966ec868aa0a6ebe9c364f1aced79b6d6348d02479f11239rr6r" event={"ID":"5f6867a7-6e2b-4f7a-bcdd-a0ff2035eee8","Type":"ContainerDied","Data":"ef250cafb07cdf8b8943c3060c96e3415d7a4926716a495954dad5782b4032a4"} Oct 05 21:07:45 crc kubenswrapper[4754]: I1005 21:07:45.097792 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/aded1cecd2f966ec868aa0a6ebe9c364f1aced79b6d6348d02479f11239rr6r" event={"ID":"5f6867a7-6e2b-4f7a-bcdd-a0ff2035eee8","Type":"ContainerStarted","Data":"7b2d10d2d64393e96db256939eb95bd8e4c5233fd52ba2323bf4d1f2632bdde5"} Oct 05 21:07:46 crc kubenswrapper[4754]: I1005 21:07:46.110547 4754 generic.go:334] "Generic (PLEG): container finished" podID="5f6867a7-6e2b-4f7a-bcdd-a0ff2035eee8" containerID="cf435f3e442d2ca43d0d90afea6dfd7433f5f0cf80aeb31b6736513981ac744b" exitCode=0 Oct 05 21:07:46 crc kubenswrapper[4754]: I1005 21:07:46.110794 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/aded1cecd2f966ec868aa0a6ebe9c364f1aced79b6d6348d02479f11239rr6r" event={"ID":"5f6867a7-6e2b-4f7a-bcdd-a0ff2035eee8","Type":"ContainerDied","Data":"cf435f3e442d2ca43d0d90afea6dfd7433f5f0cf80aeb31b6736513981ac744b"} Oct 05 21:07:47 crc kubenswrapper[4754]: I1005 21:07:47.139238 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-dvhsp"] Oct 05 21:07:47 crc kubenswrapper[4754]: I1005 21:07:47.142111 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dvhsp" Oct 05 21:07:47 crc kubenswrapper[4754]: I1005 21:07:47.150739 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-dvhsp"] Oct 05 21:07:47 crc kubenswrapper[4754]: I1005 21:07:47.151913 4754 generic.go:334] "Generic (PLEG): container finished" podID="5f6867a7-6e2b-4f7a-bcdd-a0ff2035eee8" containerID="5ccffbd0e6f877bc7c95082b359d9e23088d7c502a30ad8575d2ac0097966530" exitCode=0 Oct 05 21:07:47 crc kubenswrapper[4754]: I1005 21:07:47.151983 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/aded1cecd2f966ec868aa0a6ebe9c364f1aced79b6d6348d02479f11239rr6r" event={"ID":"5f6867a7-6e2b-4f7a-bcdd-a0ff2035eee8","Type":"ContainerDied","Data":"5ccffbd0e6f877bc7c95082b359d9e23088d7c502a30ad8575d2ac0097966530"} Oct 05 21:07:47 crc kubenswrapper[4754]: I1005 21:07:47.211417 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e1ad29ae-ca71-48f1-a2b2-28cab1053e14-utilities\") pod \"community-operators-dvhsp\" (UID: \"e1ad29ae-ca71-48f1-a2b2-28cab1053e14\") " pod="openshift-marketplace/community-operators-dvhsp" Oct 05 21:07:47 crc kubenswrapper[4754]: I1005 21:07:47.211546 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-psrsg\" (UniqueName: \"kubernetes.io/projected/e1ad29ae-ca71-48f1-a2b2-28cab1053e14-kube-api-access-psrsg\") pod \"community-operators-dvhsp\" (UID: \"e1ad29ae-ca71-48f1-a2b2-28cab1053e14\") " pod="openshift-marketplace/community-operators-dvhsp" Oct 05 21:07:47 crc kubenswrapper[4754]: I1005 21:07:47.211581 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e1ad29ae-ca71-48f1-a2b2-28cab1053e14-catalog-content\") pod \"community-operators-dvhsp\" (UID: \"e1ad29ae-ca71-48f1-a2b2-28cab1053e14\") " pod="openshift-marketplace/community-operators-dvhsp" Oct 05 21:07:47 crc kubenswrapper[4754]: I1005 21:07:47.312585 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-psrsg\" (UniqueName: \"kubernetes.io/projected/e1ad29ae-ca71-48f1-a2b2-28cab1053e14-kube-api-access-psrsg\") pod \"community-operators-dvhsp\" (UID: \"e1ad29ae-ca71-48f1-a2b2-28cab1053e14\") " pod="openshift-marketplace/community-operators-dvhsp" Oct 05 21:07:47 crc kubenswrapper[4754]: I1005 21:07:47.312629 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e1ad29ae-ca71-48f1-a2b2-28cab1053e14-catalog-content\") pod \"community-operators-dvhsp\" (UID: \"e1ad29ae-ca71-48f1-a2b2-28cab1053e14\") " pod="openshift-marketplace/community-operators-dvhsp" Oct 05 21:07:47 crc kubenswrapper[4754]: I1005 21:07:47.312700 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e1ad29ae-ca71-48f1-a2b2-28cab1053e14-utilities\") pod \"community-operators-dvhsp\" (UID: \"e1ad29ae-ca71-48f1-a2b2-28cab1053e14\") " pod="openshift-marketplace/community-operators-dvhsp" Oct 05 21:07:47 crc kubenswrapper[4754]: I1005 21:07:47.313233 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e1ad29ae-ca71-48f1-a2b2-28cab1053e14-catalog-content\") pod \"community-operators-dvhsp\" (UID: \"e1ad29ae-ca71-48f1-a2b2-28cab1053e14\") " pod="openshift-marketplace/community-operators-dvhsp" Oct 05 21:07:47 crc kubenswrapper[4754]: I1005 21:07:47.313311 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e1ad29ae-ca71-48f1-a2b2-28cab1053e14-utilities\") pod \"community-operators-dvhsp\" (UID: \"e1ad29ae-ca71-48f1-a2b2-28cab1053e14\") " pod="openshift-marketplace/community-operators-dvhsp" Oct 05 21:07:47 crc kubenswrapper[4754]: I1005 21:07:47.351657 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-psrsg\" (UniqueName: \"kubernetes.io/projected/e1ad29ae-ca71-48f1-a2b2-28cab1053e14-kube-api-access-psrsg\") pod \"community-operators-dvhsp\" (UID: \"e1ad29ae-ca71-48f1-a2b2-28cab1053e14\") " pod="openshift-marketplace/community-operators-dvhsp" Oct 05 21:07:47 crc kubenswrapper[4754]: I1005 21:07:47.477734 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dvhsp" Oct 05 21:07:48 crc kubenswrapper[4754]: I1005 21:07:48.017770 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-dvhsp"] Oct 05 21:07:48 crc kubenswrapper[4754]: W1005 21:07:48.027900 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode1ad29ae_ca71_48f1_a2b2_28cab1053e14.slice/crio-929a70746686600d51394d8a84059a23e5f69f40d8c2098a7f8f0d85d846ccdb WatchSource:0}: Error finding container 929a70746686600d51394d8a84059a23e5f69f40d8c2098a7f8f0d85d846ccdb: Status 404 returned error can't find the container with id 929a70746686600d51394d8a84059a23e5f69f40d8c2098a7f8f0d85d846ccdb Oct 05 21:07:48 crc kubenswrapper[4754]: I1005 21:07:48.162638 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dvhsp" event={"ID":"e1ad29ae-ca71-48f1-a2b2-28cab1053e14","Type":"ContainerStarted","Data":"929a70746686600d51394d8a84059a23e5f69f40d8c2098a7f8f0d85d846ccdb"} Oct 05 21:07:48 crc kubenswrapper[4754]: I1005 21:07:48.490203 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/aded1cecd2f966ec868aa0a6ebe9c364f1aced79b6d6348d02479f11239rr6r" Oct 05 21:07:48 crc kubenswrapper[4754]: I1005 21:07:48.630928 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5f6867a7-6e2b-4f7a-bcdd-a0ff2035eee8-util\") pod \"5f6867a7-6e2b-4f7a-bcdd-a0ff2035eee8\" (UID: \"5f6867a7-6e2b-4f7a-bcdd-a0ff2035eee8\") " Oct 05 21:07:48 crc kubenswrapper[4754]: I1005 21:07:48.631155 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zdn7f\" (UniqueName: \"kubernetes.io/projected/5f6867a7-6e2b-4f7a-bcdd-a0ff2035eee8-kube-api-access-zdn7f\") pod \"5f6867a7-6e2b-4f7a-bcdd-a0ff2035eee8\" (UID: \"5f6867a7-6e2b-4f7a-bcdd-a0ff2035eee8\") " Oct 05 21:07:48 crc kubenswrapper[4754]: I1005 21:07:48.632438 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5f6867a7-6e2b-4f7a-bcdd-a0ff2035eee8-bundle\") pod \"5f6867a7-6e2b-4f7a-bcdd-a0ff2035eee8\" (UID: \"5f6867a7-6e2b-4f7a-bcdd-a0ff2035eee8\") " Oct 05 21:07:48 crc kubenswrapper[4754]: I1005 21:07:48.633056 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5f6867a7-6e2b-4f7a-bcdd-a0ff2035eee8-bundle" (OuterVolumeSpecName: "bundle") pod "5f6867a7-6e2b-4f7a-bcdd-a0ff2035eee8" (UID: "5f6867a7-6e2b-4f7a-bcdd-a0ff2035eee8"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 21:07:48 crc kubenswrapper[4754]: I1005 21:07:48.633345 4754 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5f6867a7-6e2b-4f7a-bcdd-a0ff2035eee8-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 21:07:48 crc kubenswrapper[4754]: I1005 21:07:48.643441 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5f6867a7-6e2b-4f7a-bcdd-a0ff2035eee8-kube-api-access-zdn7f" (OuterVolumeSpecName: "kube-api-access-zdn7f") pod "5f6867a7-6e2b-4f7a-bcdd-a0ff2035eee8" (UID: "5f6867a7-6e2b-4f7a-bcdd-a0ff2035eee8"). InnerVolumeSpecName "kube-api-access-zdn7f". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:07:48 crc kubenswrapper[4754]: I1005 21:07:48.661838 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5f6867a7-6e2b-4f7a-bcdd-a0ff2035eee8-util" (OuterVolumeSpecName: "util") pod "5f6867a7-6e2b-4f7a-bcdd-a0ff2035eee8" (UID: "5f6867a7-6e2b-4f7a-bcdd-a0ff2035eee8"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 21:07:48 crc kubenswrapper[4754]: I1005 21:07:48.734542 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zdn7f\" (UniqueName: \"kubernetes.io/projected/5f6867a7-6e2b-4f7a-bcdd-a0ff2035eee8-kube-api-access-zdn7f\") on node \"crc\" DevicePath \"\"" Oct 05 21:07:48 crc kubenswrapper[4754]: I1005 21:07:48.734621 4754 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5f6867a7-6e2b-4f7a-bcdd-a0ff2035eee8-util\") on node \"crc\" DevicePath \"\"" Oct 05 21:07:49 crc kubenswrapper[4754]: I1005 21:07:49.175771 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/aded1cecd2f966ec868aa0a6ebe9c364f1aced79b6d6348d02479f11239rr6r" event={"ID":"5f6867a7-6e2b-4f7a-bcdd-a0ff2035eee8","Type":"ContainerDied","Data":"7b2d10d2d64393e96db256939eb95bd8e4c5233fd52ba2323bf4d1f2632bdde5"} Oct 05 21:07:49 crc kubenswrapper[4754]: I1005 21:07:49.175839 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7b2d10d2d64393e96db256939eb95bd8e4c5233fd52ba2323bf4d1f2632bdde5" Oct 05 21:07:49 crc kubenswrapper[4754]: I1005 21:07:49.175945 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/aded1cecd2f966ec868aa0a6ebe9c364f1aced79b6d6348d02479f11239rr6r" Oct 05 21:07:49 crc kubenswrapper[4754]: I1005 21:07:49.180470 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dvhsp" event={"ID":"e1ad29ae-ca71-48f1-a2b2-28cab1053e14","Type":"ContainerDied","Data":"b48d8efaada277ba2fa2b85363cfa3dd97a7d70bbdfcae2e4f162f68d45f26cc"} Oct 05 21:07:49 crc kubenswrapper[4754]: I1005 21:07:49.180353 4754 generic.go:334] "Generic (PLEG): container finished" podID="e1ad29ae-ca71-48f1-a2b2-28cab1053e14" containerID="b48d8efaada277ba2fa2b85363cfa3dd97a7d70bbdfcae2e4f162f68d45f26cc" exitCode=0 Oct 05 21:07:50 crc kubenswrapper[4754]: I1005 21:07:50.189778 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dvhsp" event={"ID":"e1ad29ae-ca71-48f1-a2b2-28cab1053e14","Type":"ContainerStarted","Data":"2a39dc9064c3867376c561bd52cf367592864d55a7cbae85e00d6bb576cd7e9c"} Oct 05 21:07:51 crc kubenswrapper[4754]: I1005 21:07:51.197510 4754 generic.go:334] "Generic (PLEG): container finished" podID="e1ad29ae-ca71-48f1-a2b2-28cab1053e14" containerID="2a39dc9064c3867376c561bd52cf367592864d55a7cbae85e00d6bb576cd7e9c" exitCode=0 Oct 05 21:07:51 crc kubenswrapper[4754]: I1005 21:07:51.198618 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dvhsp" event={"ID":"e1ad29ae-ca71-48f1-a2b2-28cab1053e14","Type":"ContainerDied","Data":"2a39dc9064c3867376c561bd52cf367592864d55a7cbae85e00d6bb576cd7e9c"} Oct 05 21:07:52 crc kubenswrapper[4754]: I1005 21:07:52.205436 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dvhsp" event={"ID":"e1ad29ae-ca71-48f1-a2b2-28cab1053e14","Type":"ContainerStarted","Data":"132ffaae7f37291cda00391a855e9b0d5a048c54be05c77151b7b320ed7e826e"} Oct 05 21:07:52 crc kubenswrapper[4754]: I1005 21:07:52.222961 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-dvhsp" podStartSLOduration=2.644633488 podStartE2EDuration="5.222944679s" podCreationTimestamp="2025-10-05 21:07:47 +0000 UTC" firstStartedPulling="2025-10-05 21:07:49.183946734 +0000 UTC m=+793.088065484" lastFinishedPulling="2025-10-05 21:07:51.762257965 +0000 UTC m=+795.666376675" observedRunningTime="2025-10-05 21:07:52.221228384 +0000 UTC m=+796.125347084" watchObservedRunningTime="2025-10-05 21:07:52.222944679 +0000 UTC m=+796.127063389" Oct 05 21:07:52 crc kubenswrapper[4754]: I1005 21:07:52.351599 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-56b64c5fcb-gzvmj"] Oct 05 21:07:52 crc kubenswrapper[4754]: E1005 21:07:52.351847 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f6867a7-6e2b-4f7a-bcdd-a0ff2035eee8" containerName="pull" Oct 05 21:07:52 crc kubenswrapper[4754]: I1005 21:07:52.351861 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f6867a7-6e2b-4f7a-bcdd-a0ff2035eee8" containerName="pull" Oct 05 21:07:52 crc kubenswrapper[4754]: E1005 21:07:52.351875 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f6867a7-6e2b-4f7a-bcdd-a0ff2035eee8" containerName="util" Oct 05 21:07:52 crc kubenswrapper[4754]: I1005 21:07:52.351881 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f6867a7-6e2b-4f7a-bcdd-a0ff2035eee8" containerName="util" Oct 05 21:07:52 crc kubenswrapper[4754]: E1005 21:07:52.351895 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f6867a7-6e2b-4f7a-bcdd-a0ff2035eee8" containerName="extract" Oct 05 21:07:52 crc kubenswrapper[4754]: I1005 21:07:52.351901 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f6867a7-6e2b-4f7a-bcdd-a0ff2035eee8" containerName="extract" Oct 05 21:07:52 crc kubenswrapper[4754]: I1005 21:07:52.352010 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="5f6867a7-6e2b-4f7a-bcdd-a0ff2035eee8" containerName="extract" Oct 05 21:07:52 crc kubenswrapper[4754]: I1005 21:07:52.352576 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-56b64c5fcb-gzvmj" Oct 05 21:07:52 crc kubenswrapper[4754]: I1005 21:07:52.354675 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-84tx7" Oct 05 21:07:52 crc kubenswrapper[4754]: I1005 21:07:52.382764 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-56b64c5fcb-gzvmj"] Oct 05 21:07:52 crc kubenswrapper[4754]: I1005 21:07:52.491702 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9t7t2\" (UniqueName: \"kubernetes.io/projected/21fc1c35-c245-457c-a66f-97536332ed52-kube-api-access-9t7t2\") pod \"openstack-operator-controller-operator-56b64c5fcb-gzvmj\" (UID: \"21fc1c35-c245-457c-a66f-97536332ed52\") " pod="openstack-operators/openstack-operator-controller-operator-56b64c5fcb-gzvmj" Oct 05 21:07:52 crc kubenswrapper[4754]: I1005 21:07:52.592999 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9t7t2\" (UniqueName: \"kubernetes.io/projected/21fc1c35-c245-457c-a66f-97536332ed52-kube-api-access-9t7t2\") pod \"openstack-operator-controller-operator-56b64c5fcb-gzvmj\" (UID: \"21fc1c35-c245-457c-a66f-97536332ed52\") " pod="openstack-operators/openstack-operator-controller-operator-56b64c5fcb-gzvmj" Oct 05 21:07:52 crc kubenswrapper[4754]: I1005 21:07:52.618247 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9t7t2\" (UniqueName: \"kubernetes.io/projected/21fc1c35-c245-457c-a66f-97536332ed52-kube-api-access-9t7t2\") pod \"openstack-operator-controller-operator-56b64c5fcb-gzvmj\" (UID: \"21fc1c35-c245-457c-a66f-97536332ed52\") " pod="openstack-operators/openstack-operator-controller-operator-56b64c5fcb-gzvmj" Oct 05 21:07:52 crc kubenswrapper[4754]: I1005 21:07:52.669222 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-56b64c5fcb-gzvmj" Oct 05 21:07:53 crc kubenswrapper[4754]: I1005 21:07:53.132729 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-56b64c5fcb-gzvmj"] Oct 05 21:07:53 crc kubenswrapper[4754]: I1005 21:07:53.212668 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-56b64c5fcb-gzvmj" event={"ID":"21fc1c35-c245-457c-a66f-97536332ed52","Type":"ContainerStarted","Data":"f4b2162f972911cf0d8da249f1304846cda0bfa2d581ddc0ede64d70d6d8e18f"} Oct 05 21:07:57 crc kubenswrapper[4754]: I1005 21:07:57.245172 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-56b64c5fcb-gzvmj" event={"ID":"21fc1c35-c245-457c-a66f-97536332ed52","Type":"ContainerStarted","Data":"180480dd6f618c9f287241a242ee703572abe56451d19d7cc86207df4fd07eb1"} Oct 05 21:07:57 crc kubenswrapper[4754]: I1005 21:07:57.478028 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-dvhsp" Oct 05 21:07:57 crc kubenswrapper[4754]: I1005 21:07:57.478120 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-dvhsp" Oct 05 21:07:57 crc kubenswrapper[4754]: I1005 21:07:57.538270 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-dvhsp" Oct 05 21:07:58 crc kubenswrapper[4754]: I1005 21:07:58.298368 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-dvhsp" Oct 05 21:07:59 crc kubenswrapper[4754]: I1005 21:07:59.325157 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-dvhsp"] Oct 05 21:08:00 crc kubenswrapper[4754]: I1005 21:08:00.268447 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-56b64c5fcb-gzvmj" event={"ID":"21fc1c35-c245-457c-a66f-97536332ed52","Type":"ContainerStarted","Data":"3079942f0ba915a18b50eeb4533faaf44dabb95147b5635828bfdb4e4410bb14"} Oct 05 21:08:00 crc kubenswrapper[4754]: I1005 21:08:00.268611 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-dvhsp" podUID="e1ad29ae-ca71-48f1-a2b2-28cab1053e14" containerName="registry-server" containerID="cri-o://132ffaae7f37291cda00391a855e9b0d5a048c54be05c77151b7b320ed7e826e" gracePeriod=2 Oct 05 21:08:00 crc kubenswrapper[4754]: I1005 21:08:00.322774 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-56b64c5fcb-gzvmj" podStartSLOduration=1.908359596 podStartE2EDuration="8.322753994s" podCreationTimestamp="2025-10-05 21:07:52 +0000 UTC" firstStartedPulling="2025-10-05 21:07:53.142910708 +0000 UTC m=+797.047029418" lastFinishedPulling="2025-10-05 21:07:59.557305096 +0000 UTC m=+803.461423816" observedRunningTime="2025-10-05 21:08:00.316210222 +0000 UTC m=+804.220328942" watchObservedRunningTime="2025-10-05 21:08:00.322753994 +0000 UTC m=+804.226872704" Oct 05 21:08:00 crc kubenswrapper[4754]: I1005 21:08:00.728175 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dvhsp" Oct 05 21:08:00 crc kubenswrapper[4754]: I1005 21:08:00.916369 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-psrsg\" (UniqueName: \"kubernetes.io/projected/e1ad29ae-ca71-48f1-a2b2-28cab1053e14-kube-api-access-psrsg\") pod \"e1ad29ae-ca71-48f1-a2b2-28cab1053e14\" (UID: \"e1ad29ae-ca71-48f1-a2b2-28cab1053e14\") " Oct 05 21:08:00 crc kubenswrapper[4754]: I1005 21:08:00.916471 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e1ad29ae-ca71-48f1-a2b2-28cab1053e14-utilities\") pod \"e1ad29ae-ca71-48f1-a2b2-28cab1053e14\" (UID: \"e1ad29ae-ca71-48f1-a2b2-28cab1053e14\") " Oct 05 21:08:00 crc kubenswrapper[4754]: I1005 21:08:00.916524 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e1ad29ae-ca71-48f1-a2b2-28cab1053e14-catalog-content\") pod \"e1ad29ae-ca71-48f1-a2b2-28cab1053e14\" (UID: \"e1ad29ae-ca71-48f1-a2b2-28cab1053e14\") " Oct 05 21:08:00 crc kubenswrapper[4754]: I1005 21:08:00.918457 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e1ad29ae-ca71-48f1-a2b2-28cab1053e14-utilities" (OuterVolumeSpecName: "utilities") pod "e1ad29ae-ca71-48f1-a2b2-28cab1053e14" (UID: "e1ad29ae-ca71-48f1-a2b2-28cab1053e14"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 21:08:00 crc kubenswrapper[4754]: I1005 21:08:00.927792 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e1ad29ae-ca71-48f1-a2b2-28cab1053e14-kube-api-access-psrsg" (OuterVolumeSpecName: "kube-api-access-psrsg") pod "e1ad29ae-ca71-48f1-a2b2-28cab1053e14" (UID: "e1ad29ae-ca71-48f1-a2b2-28cab1053e14"). InnerVolumeSpecName "kube-api-access-psrsg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:08:01 crc kubenswrapper[4754]: I1005 21:08:01.017761 4754 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e1ad29ae-ca71-48f1-a2b2-28cab1053e14-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 21:08:01 crc kubenswrapper[4754]: I1005 21:08:01.018100 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-psrsg\" (UniqueName: \"kubernetes.io/projected/e1ad29ae-ca71-48f1-a2b2-28cab1053e14-kube-api-access-psrsg\") on node \"crc\" DevicePath \"\"" Oct 05 21:08:01 crc kubenswrapper[4754]: I1005 21:08:01.282724 4754 generic.go:334] "Generic (PLEG): container finished" podID="e1ad29ae-ca71-48f1-a2b2-28cab1053e14" containerID="132ffaae7f37291cda00391a855e9b0d5a048c54be05c77151b7b320ed7e826e" exitCode=0 Oct 05 21:08:01 crc kubenswrapper[4754]: I1005 21:08:01.282802 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dvhsp" event={"ID":"e1ad29ae-ca71-48f1-a2b2-28cab1053e14","Type":"ContainerDied","Data":"132ffaae7f37291cda00391a855e9b0d5a048c54be05c77151b7b320ed7e826e"} Oct 05 21:08:01 crc kubenswrapper[4754]: I1005 21:08:01.282843 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dvhsp" event={"ID":"e1ad29ae-ca71-48f1-a2b2-28cab1053e14","Type":"ContainerDied","Data":"929a70746686600d51394d8a84059a23e5f69f40d8c2098a7f8f0d85d846ccdb"} Oct 05 21:08:01 crc kubenswrapper[4754]: I1005 21:08:01.282864 4754 scope.go:117] "RemoveContainer" containerID="132ffaae7f37291cda00391a855e9b0d5a048c54be05c77151b7b320ed7e826e" Oct 05 21:08:01 crc kubenswrapper[4754]: I1005 21:08:01.282828 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dvhsp" Oct 05 21:08:01 crc kubenswrapper[4754]: I1005 21:08:01.283468 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-56b64c5fcb-gzvmj" Oct 05 21:08:01 crc kubenswrapper[4754]: I1005 21:08:01.314829 4754 scope.go:117] "RemoveContainer" containerID="2a39dc9064c3867376c561bd52cf367592864d55a7cbae85e00d6bb576cd7e9c" Oct 05 21:08:01 crc kubenswrapper[4754]: I1005 21:08:01.340004 4754 scope.go:117] "RemoveContainer" containerID="b48d8efaada277ba2fa2b85363cfa3dd97a7d70bbdfcae2e4f162f68d45f26cc" Oct 05 21:08:01 crc kubenswrapper[4754]: I1005 21:08:01.345912 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e1ad29ae-ca71-48f1-a2b2-28cab1053e14-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e1ad29ae-ca71-48f1-a2b2-28cab1053e14" (UID: "e1ad29ae-ca71-48f1-a2b2-28cab1053e14"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 21:08:01 crc kubenswrapper[4754]: I1005 21:08:01.369898 4754 scope.go:117] "RemoveContainer" containerID="132ffaae7f37291cda00391a855e9b0d5a048c54be05c77151b7b320ed7e826e" Oct 05 21:08:01 crc kubenswrapper[4754]: E1005 21:08:01.370358 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"132ffaae7f37291cda00391a855e9b0d5a048c54be05c77151b7b320ed7e826e\": container with ID starting with 132ffaae7f37291cda00391a855e9b0d5a048c54be05c77151b7b320ed7e826e not found: ID does not exist" containerID="132ffaae7f37291cda00391a855e9b0d5a048c54be05c77151b7b320ed7e826e" Oct 05 21:08:01 crc kubenswrapper[4754]: I1005 21:08:01.370397 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"132ffaae7f37291cda00391a855e9b0d5a048c54be05c77151b7b320ed7e826e"} err="failed to get container status \"132ffaae7f37291cda00391a855e9b0d5a048c54be05c77151b7b320ed7e826e\": rpc error: code = NotFound desc = could not find container \"132ffaae7f37291cda00391a855e9b0d5a048c54be05c77151b7b320ed7e826e\": container with ID starting with 132ffaae7f37291cda00391a855e9b0d5a048c54be05c77151b7b320ed7e826e not found: ID does not exist" Oct 05 21:08:01 crc kubenswrapper[4754]: I1005 21:08:01.370423 4754 scope.go:117] "RemoveContainer" containerID="2a39dc9064c3867376c561bd52cf367592864d55a7cbae85e00d6bb576cd7e9c" Oct 05 21:08:01 crc kubenswrapper[4754]: E1005 21:08:01.370887 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2a39dc9064c3867376c561bd52cf367592864d55a7cbae85e00d6bb576cd7e9c\": container with ID starting with 2a39dc9064c3867376c561bd52cf367592864d55a7cbae85e00d6bb576cd7e9c not found: ID does not exist" containerID="2a39dc9064c3867376c561bd52cf367592864d55a7cbae85e00d6bb576cd7e9c" Oct 05 21:08:01 crc kubenswrapper[4754]: I1005 21:08:01.370925 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2a39dc9064c3867376c561bd52cf367592864d55a7cbae85e00d6bb576cd7e9c"} err="failed to get container status \"2a39dc9064c3867376c561bd52cf367592864d55a7cbae85e00d6bb576cd7e9c\": rpc error: code = NotFound desc = could not find container \"2a39dc9064c3867376c561bd52cf367592864d55a7cbae85e00d6bb576cd7e9c\": container with ID starting with 2a39dc9064c3867376c561bd52cf367592864d55a7cbae85e00d6bb576cd7e9c not found: ID does not exist" Oct 05 21:08:01 crc kubenswrapper[4754]: I1005 21:08:01.370943 4754 scope.go:117] "RemoveContainer" containerID="b48d8efaada277ba2fa2b85363cfa3dd97a7d70bbdfcae2e4f162f68d45f26cc" Oct 05 21:08:01 crc kubenswrapper[4754]: E1005 21:08:01.371372 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b48d8efaada277ba2fa2b85363cfa3dd97a7d70bbdfcae2e4f162f68d45f26cc\": container with ID starting with b48d8efaada277ba2fa2b85363cfa3dd97a7d70bbdfcae2e4f162f68d45f26cc not found: ID does not exist" containerID="b48d8efaada277ba2fa2b85363cfa3dd97a7d70bbdfcae2e4f162f68d45f26cc" Oct 05 21:08:01 crc kubenswrapper[4754]: I1005 21:08:01.371440 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b48d8efaada277ba2fa2b85363cfa3dd97a7d70bbdfcae2e4f162f68d45f26cc"} err="failed to get container status \"b48d8efaada277ba2fa2b85363cfa3dd97a7d70bbdfcae2e4f162f68d45f26cc\": rpc error: code = NotFound desc = could not find container \"b48d8efaada277ba2fa2b85363cfa3dd97a7d70bbdfcae2e4f162f68d45f26cc\": container with ID starting with b48d8efaada277ba2fa2b85363cfa3dd97a7d70bbdfcae2e4f162f68d45f26cc not found: ID does not exist" Oct 05 21:08:01 crc kubenswrapper[4754]: I1005 21:08:01.424775 4754 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e1ad29ae-ca71-48f1-a2b2-28cab1053e14-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 21:08:01 crc kubenswrapper[4754]: I1005 21:08:01.629446 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-dvhsp"] Oct 05 21:08:01 crc kubenswrapper[4754]: I1005 21:08:01.639105 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-dvhsp"] Oct 05 21:08:02 crc kubenswrapper[4754]: I1005 21:08:02.301904 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-56b64c5fcb-gzvmj" Oct 05 21:08:02 crc kubenswrapper[4754]: I1005 21:08:02.853533 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e1ad29ae-ca71-48f1-a2b2-28cab1053e14" path="/var/lib/kubelet/pods/e1ad29ae-ca71-48f1-a2b2-28cab1053e14/volumes" Oct 05 21:08:17 crc kubenswrapper[4754]: I1005 21:08:17.506573 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-5f7c849b98-n9ksx"] Oct 05 21:08:17 crc kubenswrapper[4754]: E1005 21:08:17.508161 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1ad29ae-ca71-48f1-a2b2-28cab1053e14" containerName="extract-utilities" Oct 05 21:08:17 crc kubenswrapper[4754]: I1005 21:08:17.508232 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1ad29ae-ca71-48f1-a2b2-28cab1053e14" containerName="extract-utilities" Oct 05 21:08:17 crc kubenswrapper[4754]: E1005 21:08:17.508302 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1ad29ae-ca71-48f1-a2b2-28cab1053e14" containerName="extract-content" Oct 05 21:08:17 crc kubenswrapper[4754]: I1005 21:08:17.508362 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1ad29ae-ca71-48f1-a2b2-28cab1053e14" containerName="extract-content" Oct 05 21:08:17 crc kubenswrapper[4754]: E1005 21:08:17.508418 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1ad29ae-ca71-48f1-a2b2-28cab1053e14" containerName="registry-server" Oct 05 21:08:17 crc kubenswrapper[4754]: I1005 21:08:17.508467 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1ad29ae-ca71-48f1-a2b2-28cab1053e14" containerName="registry-server" Oct 05 21:08:17 crc kubenswrapper[4754]: I1005 21:08:17.508649 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1ad29ae-ca71-48f1-a2b2-28cab1053e14" containerName="registry-server" Oct 05 21:08:17 crc kubenswrapper[4754]: I1005 21:08:17.509329 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-5f7c849b98-n9ksx" Oct 05 21:08:17 crc kubenswrapper[4754]: I1005 21:08:17.516782 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-5bmn8" Oct 05 21:08:17 crc kubenswrapper[4754]: I1005 21:08:17.538674 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-75dfd9b554-fxx5s"] Oct 05 21:08:17 crc kubenswrapper[4754]: I1005 21:08:17.539658 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-fxx5s" Oct 05 21:08:17 crc kubenswrapper[4754]: I1005 21:08:17.542478 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-rpjd8" Oct 05 21:08:17 crc kubenswrapper[4754]: I1005 21:08:17.549822 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-5f7c849b98-n9ksx"] Oct 05 21:08:17 crc kubenswrapper[4754]: I1005 21:08:17.569529 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-7d4d4f8d-kcxvd"] Oct 05 21:08:17 crc kubenswrapper[4754]: I1005 21:08:17.570518 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-7d4d4f8d-kcxvd" Oct 05 21:08:17 crc kubenswrapper[4754]: I1005 21:08:17.575402 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-5568b5d68-f5rjh"] Oct 05 21:08:17 crc kubenswrapper[4754]: I1005 21:08:17.576463 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-5568b5d68-f5rjh" Oct 05 21:08:17 crc kubenswrapper[4754]: I1005 21:08:17.578262 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-75dfd9b554-fxx5s"] Oct 05 21:08:17 crc kubenswrapper[4754]: I1005 21:08:17.578400 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-jp98m" Oct 05 21:08:17 crc kubenswrapper[4754]: I1005 21:08:17.583144 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-m8dhp" Oct 05 21:08:17 crc kubenswrapper[4754]: I1005 21:08:17.603761 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-5568b5d68-f5rjh"] Oct 05 21:08:17 crc kubenswrapper[4754]: I1005 21:08:17.628603 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-8f58bc9db-xsrhv"] Oct 05 21:08:17 crc kubenswrapper[4754]: I1005 21:08:17.629836 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-8f58bc9db-xsrhv" Oct 05 21:08:17 crc kubenswrapper[4754]: I1005 21:08:17.631699 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-g4hgz" Oct 05 21:08:17 crc kubenswrapper[4754]: I1005 21:08:17.638098 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-7d4d4f8d-kcxvd"] Oct 05 21:08:17 crc kubenswrapper[4754]: I1005 21:08:17.658060 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-8f58bc9db-xsrhv"] Oct 05 21:08:17 crc kubenswrapper[4754]: I1005 21:08:17.668382 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-966sj\" (UniqueName: \"kubernetes.io/projected/3ec442d5-220a-4d2f-8f62-a13ff86a3229-kube-api-access-966sj\") pod \"designate-operator-controller-manager-75dfd9b554-fxx5s\" (UID: \"3ec442d5-220a-4d2f-8f62-a13ff86a3229\") " pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-fxx5s" Oct 05 21:08:17 crc kubenswrapper[4754]: I1005 21:08:17.668469 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qr6mp\" (UniqueName: \"kubernetes.io/projected/9a1cc24e-8371-44cb-bd37-5b765853fdff-kube-api-access-qr6mp\") pod \"barbican-operator-controller-manager-5f7c849b98-n9ksx\" (UID: \"9a1cc24e-8371-44cb-bd37-5b765853fdff\") " pod="openstack-operators/barbican-operator-controller-manager-5f7c849b98-n9ksx" Oct 05 21:08:17 crc kubenswrapper[4754]: I1005 21:08:17.676774 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-658588b8c9-z7tq9"] Oct 05 21:08:17 crc kubenswrapper[4754]: I1005 21:08:17.677755 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-658588b8c9-z7tq9" Oct 05 21:08:17 crc kubenswrapper[4754]: I1005 21:08:17.680032 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Oct 05 21:08:17 crc kubenswrapper[4754]: I1005 21:08:17.680298 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-2lbjg" Oct 05 21:08:17 crc kubenswrapper[4754]: I1005 21:08:17.705344 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-658588b8c9-z7tq9"] Oct 05 21:08:17 crc kubenswrapper[4754]: I1005 21:08:17.713101 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-54876c876f-svzwt"] Oct 05 21:08:17 crc kubenswrapper[4754]: I1005 21:08:17.714165 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-54876c876f-svzwt" Oct 05 21:08:17 crc kubenswrapper[4754]: I1005 21:08:17.721813 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-625mw" Oct 05 21:08:17 crc kubenswrapper[4754]: I1005 21:08:17.785595 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-966sj\" (UniqueName: \"kubernetes.io/projected/3ec442d5-220a-4d2f-8f62-a13ff86a3229-kube-api-access-966sj\") pod \"designate-operator-controller-manager-75dfd9b554-fxx5s\" (UID: \"3ec442d5-220a-4d2f-8f62-a13ff86a3229\") " pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-fxx5s" Oct 05 21:08:17 crc kubenswrapper[4754]: I1005 21:08:17.785649 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bfghj\" (UniqueName: \"kubernetes.io/projected/82184642-9d7b-4ada-b7b2-efde6c91cb14-kube-api-access-bfghj\") pod \"heat-operator-controller-manager-8f58bc9db-xsrhv\" (UID: \"82184642-9d7b-4ada-b7b2-efde6c91cb14\") " pod="openstack-operators/heat-operator-controller-manager-8f58bc9db-xsrhv" Oct 05 21:08:17 crc kubenswrapper[4754]: I1005 21:08:17.785670 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q95vh\" (UniqueName: \"kubernetes.io/projected/47cafc06-0f75-48ee-bae5-4e30352f4572-kube-api-access-q95vh\") pod \"cinder-operator-controller-manager-7d4d4f8d-kcxvd\" (UID: \"47cafc06-0f75-48ee-bae5-4e30352f4572\") " pod="openstack-operators/cinder-operator-controller-manager-7d4d4f8d-kcxvd" Oct 05 21:08:17 crc kubenswrapper[4754]: I1005 21:08:17.785688 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wgxjc\" (UniqueName: \"kubernetes.io/projected/62159d2d-ec19-4b44-89df-846faa5757dc-kube-api-access-wgxjc\") pod \"glance-operator-controller-manager-5568b5d68-f5rjh\" (UID: \"62159d2d-ec19-4b44-89df-846faa5757dc\") " pod="openstack-operators/glance-operator-controller-manager-5568b5d68-f5rjh" Oct 05 21:08:17 crc kubenswrapper[4754]: I1005 21:08:17.785733 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qr6mp\" (UniqueName: \"kubernetes.io/projected/9a1cc24e-8371-44cb-bd37-5b765853fdff-kube-api-access-qr6mp\") pod \"barbican-operator-controller-manager-5f7c849b98-n9ksx\" (UID: \"9a1cc24e-8371-44cb-bd37-5b765853fdff\") " pod="openstack-operators/barbican-operator-controller-manager-5f7c849b98-n9ksx" Oct 05 21:08:17 crc kubenswrapper[4754]: I1005 21:08:17.788683 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-54876c876f-svzwt"] Oct 05 21:08:17 crc kubenswrapper[4754]: I1005 21:08:17.794192 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-649675d675-v7285"] Oct 05 21:08:17 crc kubenswrapper[4754]: I1005 21:08:17.795323 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-649675d675-v7285" Oct 05 21:08:17 crc kubenswrapper[4754]: I1005 21:08:17.798286 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-brhcd" Oct 05 21:08:17 crc kubenswrapper[4754]: I1005 21:08:17.836005 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-655d88ccb9-l56qv"] Oct 05 21:08:17 crc kubenswrapper[4754]: I1005 21:08:17.836454 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qr6mp\" (UniqueName: \"kubernetes.io/projected/9a1cc24e-8371-44cb-bd37-5b765853fdff-kube-api-access-qr6mp\") pod \"barbican-operator-controller-manager-5f7c849b98-n9ksx\" (UID: \"9a1cc24e-8371-44cb-bd37-5b765853fdff\") " pod="openstack-operators/barbican-operator-controller-manager-5f7c849b98-n9ksx" Oct 05 21:08:17 crc kubenswrapper[4754]: I1005 21:08:17.837296 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-655d88ccb9-l56qv" Oct 05 21:08:17 crc kubenswrapper[4754]: I1005 21:08:17.843632 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-fxk6f" Oct 05 21:08:17 crc kubenswrapper[4754]: I1005 21:08:17.847866 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-966sj\" (UniqueName: \"kubernetes.io/projected/3ec442d5-220a-4d2f-8f62-a13ff86a3229-kube-api-access-966sj\") pod \"designate-operator-controller-manager-75dfd9b554-fxx5s\" (UID: \"3ec442d5-220a-4d2f-8f62-a13ff86a3229\") " pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-fxx5s" Oct 05 21:08:17 crc kubenswrapper[4754]: I1005 21:08:17.865229 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-65d89cfd9f-msh87"] Oct 05 21:08:17 crc kubenswrapper[4754]: I1005 21:08:17.866268 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-65d89cfd9f-msh87" Oct 05 21:08:17 crc kubenswrapper[4754]: I1005 21:08:17.870264 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-fxx5s" Oct 05 21:08:17 crc kubenswrapper[4754]: I1005 21:08:17.872319 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-vccdl" Oct 05 21:08:17 crc kubenswrapper[4754]: I1005 21:08:17.874300 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-649675d675-v7285"] Oct 05 21:08:17 crc kubenswrapper[4754]: I1005 21:08:17.879621 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-655d88ccb9-l56qv"] Oct 05 21:08:17 crc kubenswrapper[4754]: I1005 21:08:17.896418 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-24bn6\" (UniqueName: \"kubernetes.io/projected/4bb20421-ce34-44ee-8740-82eedd4716f3-kube-api-access-24bn6\") pod \"infra-operator-controller-manager-658588b8c9-z7tq9\" (UID: \"4bb20421-ce34-44ee-8740-82eedd4716f3\") " pod="openstack-operators/infra-operator-controller-manager-658588b8c9-z7tq9" Oct 05 21:08:17 crc kubenswrapper[4754]: I1005 21:08:17.896492 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m6jzx\" (UniqueName: \"kubernetes.io/projected/f560133d-b3b5-4791-b69a-a700367f0d96-kube-api-access-m6jzx\") pod \"horizon-operator-controller-manager-54876c876f-svzwt\" (UID: \"f560133d-b3b5-4791-b69a-a700367f0d96\") " pod="openstack-operators/horizon-operator-controller-manager-54876c876f-svzwt" Oct 05 21:08:17 crc kubenswrapper[4754]: I1005 21:08:17.896674 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bfghj\" (UniqueName: \"kubernetes.io/projected/82184642-9d7b-4ada-b7b2-efde6c91cb14-kube-api-access-bfghj\") pod \"heat-operator-controller-manager-8f58bc9db-xsrhv\" (UID: \"82184642-9d7b-4ada-b7b2-efde6c91cb14\") " pod="openstack-operators/heat-operator-controller-manager-8f58bc9db-xsrhv" Oct 05 21:08:17 crc kubenswrapper[4754]: I1005 21:08:17.896694 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q95vh\" (UniqueName: \"kubernetes.io/projected/47cafc06-0f75-48ee-bae5-4e30352f4572-kube-api-access-q95vh\") pod \"cinder-operator-controller-manager-7d4d4f8d-kcxvd\" (UID: \"47cafc06-0f75-48ee-bae5-4e30352f4572\") " pod="openstack-operators/cinder-operator-controller-manager-7d4d4f8d-kcxvd" Oct 05 21:08:17 crc kubenswrapper[4754]: I1005 21:08:17.896710 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4bb20421-ce34-44ee-8740-82eedd4716f3-cert\") pod \"infra-operator-controller-manager-658588b8c9-z7tq9\" (UID: \"4bb20421-ce34-44ee-8740-82eedd4716f3\") " pod="openstack-operators/infra-operator-controller-manager-658588b8c9-z7tq9" Oct 05 21:08:17 crc kubenswrapper[4754]: I1005 21:08:17.896733 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wgxjc\" (UniqueName: \"kubernetes.io/projected/62159d2d-ec19-4b44-89df-846faa5757dc-kube-api-access-wgxjc\") pod \"glance-operator-controller-manager-5568b5d68-f5rjh\" (UID: \"62159d2d-ec19-4b44-89df-846faa5757dc\") " pod="openstack-operators/glance-operator-controller-manager-5568b5d68-f5rjh" Oct 05 21:08:17 crc kubenswrapper[4754]: I1005 21:08:17.897750 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-65d89cfd9f-msh87"] Oct 05 21:08:17 crc kubenswrapper[4754]: I1005 21:08:17.937682 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q95vh\" (UniqueName: \"kubernetes.io/projected/47cafc06-0f75-48ee-bae5-4e30352f4572-kube-api-access-q95vh\") pod \"cinder-operator-controller-manager-7d4d4f8d-kcxvd\" (UID: \"47cafc06-0f75-48ee-bae5-4e30352f4572\") " pod="openstack-operators/cinder-operator-controller-manager-7d4d4f8d-kcxvd" Oct 05 21:08:17 crc kubenswrapper[4754]: I1005 21:08:17.940480 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-456m8"] Oct 05 21:08:17 crc kubenswrapper[4754]: I1005 21:08:17.945692 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-456m8" Oct 05 21:08:17 crc kubenswrapper[4754]: I1005 21:08:17.951104 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bfghj\" (UniqueName: \"kubernetes.io/projected/82184642-9d7b-4ada-b7b2-efde6c91cb14-kube-api-access-bfghj\") pod \"heat-operator-controller-manager-8f58bc9db-xsrhv\" (UID: \"82184642-9d7b-4ada-b7b2-efde6c91cb14\") " pod="openstack-operators/heat-operator-controller-manager-8f58bc9db-xsrhv" Oct 05 21:08:17 crc kubenswrapper[4754]: I1005 21:08:17.954886 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-456m8"] Oct 05 21:08:17 crc kubenswrapper[4754]: I1005 21:08:17.962688 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wgxjc\" (UniqueName: \"kubernetes.io/projected/62159d2d-ec19-4b44-89df-846faa5757dc-kube-api-access-wgxjc\") pod \"glance-operator-controller-manager-5568b5d68-f5rjh\" (UID: \"62159d2d-ec19-4b44-89df-846faa5757dc\") " pod="openstack-operators/glance-operator-controller-manager-5568b5d68-f5rjh" Oct 05 21:08:17 crc kubenswrapper[4754]: I1005 21:08:17.989756 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-6cd6d7bdf5-f5jxx"] Oct 05 21:08:17 crc kubenswrapper[4754]: I1005 21:08:17.990800 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-6cd6d7bdf5-f5jxx" Oct 05 21:08:18 crc kubenswrapper[4754]: I1005 21:08:17.998883 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-6cd6d7bdf5-f5jxx"] Oct 05 21:08:18 crc kubenswrapper[4754]: I1005 21:08:17.999025 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-2kkqv" Oct 05 21:08:18 crc kubenswrapper[4754]: I1005 21:08:17.999171 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kcq9f\" (UniqueName: \"kubernetes.io/projected/5a1312fe-c6d6-41de-a442-b044db904a6d-kube-api-access-kcq9f\") pod \"mariadb-operator-controller-manager-6cd6d7bdf5-f5jxx\" (UID: \"5a1312fe-c6d6-41de-a442-b044db904a6d\") " pod="openstack-operators/mariadb-operator-controller-manager-6cd6d7bdf5-f5jxx" Oct 05 21:08:18 crc kubenswrapper[4754]: I1005 21:08:17.999224 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zbcs6\" (UniqueName: \"kubernetes.io/projected/5757c596-103f-4a00-ac16-ed9e6c9e4719-kube-api-access-zbcs6\") pod \"manila-operator-controller-manager-65d89cfd9f-msh87\" (UID: \"5757c596-103f-4a00-ac16-ed9e6c9e4719\") " pod="openstack-operators/manila-operator-controller-manager-65d89cfd9f-msh87" Oct 05 21:08:18 crc kubenswrapper[4754]: I1005 21:08:17.999248 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x825r\" (UniqueName: \"kubernetes.io/projected/2735f261-cad9-4dff-9136-931a45c85ac5-kube-api-access-x825r\") pod \"keystone-operator-controller-manager-655d88ccb9-l56qv\" (UID: \"2735f261-cad9-4dff-9136-931a45c85ac5\") " pod="openstack-operators/keystone-operator-controller-manager-655d88ccb9-l56qv" Oct 05 21:08:18 crc kubenswrapper[4754]: I1005 21:08:17.999307 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-24bn6\" (UniqueName: \"kubernetes.io/projected/4bb20421-ce34-44ee-8740-82eedd4716f3-kube-api-access-24bn6\") pod \"infra-operator-controller-manager-658588b8c9-z7tq9\" (UID: \"4bb20421-ce34-44ee-8740-82eedd4716f3\") " pod="openstack-operators/infra-operator-controller-manager-658588b8c9-z7tq9" Oct 05 21:08:18 crc kubenswrapper[4754]: I1005 21:08:17.999374 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m6jzx\" (UniqueName: \"kubernetes.io/projected/f560133d-b3b5-4791-b69a-a700367f0d96-kube-api-access-m6jzx\") pod \"horizon-operator-controller-manager-54876c876f-svzwt\" (UID: \"f560133d-b3b5-4791-b69a-a700367f0d96\") " pod="openstack-operators/horizon-operator-controller-manager-54876c876f-svzwt" Oct 05 21:08:18 crc kubenswrapper[4754]: I1005 21:08:17.999426 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9xbvr\" (UniqueName: \"kubernetes.io/projected/3ddbeeee-8a3e-4f3c-97e0-c7d8d5929345-kube-api-access-9xbvr\") pod \"ironic-operator-controller-manager-649675d675-v7285\" (UID: \"3ddbeeee-8a3e-4f3c-97e0-c7d8d5929345\") " pod="openstack-operators/ironic-operator-controller-manager-649675d675-v7285" Oct 05 21:08:18 crc kubenswrapper[4754]: I1005 21:08:17.999478 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4bb20421-ce34-44ee-8740-82eedd4716f3-cert\") pod \"infra-operator-controller-manager-658588b8c9-z7tq9\" (UID: \"4bb20421-ce34-44ee-8740-82eedd4716f3\") " pod="openstack-operators/infra-operator-controller-manager-658588b8c9-z7tq9" Oct 05 21:08:18 crc kubenswrapper[4754]: E1005 21:08:17.999595 4754 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Oct 05 21:08:18 crc kubenswrapper[4754]: E1005 21:08:17.999645 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4bb20421-ce34-44ee-8740-82eedd4716f3-cert podName:4bb20421-ce34-44ee-8740-82eedd4716f3 nodeName:}" failed. No retries permitted until 2025-10-05 21:08:18.499623798 +0000 UTC m=+822.403742508 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/4bb20421-ce34-44ee-8740-82eedd4716f3-cert") pod "infra-operator-controller-manager-658588b8c9-z7tq9" (UID: "4bb20421-ce34-44ee-8740-82eedd4716f3") : secret "infra-operator-webhook-server-cert" not found Oct 05 21:08:18 crc kubenswrapper[4754]: I1005 21:08:18.011746 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-8d984cc4d-7xzhb"] Oct 05 21:08:18 crc kubenswrapper[4754]: I1005 21:08:18.013086 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-8d984cc4d-7xzhb" Oct 05 21:08:18 crc kubenswrapper[4754]: I1005 21:08:18.017638 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-g89tm" Oct 05 21:08:18 crc kubenswrapper[4754]: I1005 21:08:18.040039 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-7468f855d8-w8z85"] Oct 05 21:08:18 crc kubenswrapper[4754]: I1005 21:08:18.040970 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-7468f855d8-w8z85" Oct 05 21:08:18 crc kubenswrapper[4754]: I1005 21:08:18.043893 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m6jzx\" (UniqueName: \"kubernetes.io/projected/f560133d-b3b5-4791-b69a-a700367f0d96-kube-api-access-m6jzx\") pod \"horizon-operator-controller-manager-54876c876f-svzwt\" (UID: \"f560133d-b3b5-4791-b69a-a700367f0d96\") " pod="openstack-operators/horizon-operator-controller-manager-54876c876f-svzwt" Oct 05 21:08:18 crc kubenswrapper[4754]: I1005 21:08:18.044692 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-24bn6\" (UniqueName: \"kubernetes.io/projected/4bb20421-ce34-44ee-8740-82eedd4716f3-kube-api-access-24bn6\") pod \"infra-operator-controller-manager-658588b8c9-z7tq9\" (UID: \"4bb20421-ce34-44ee-8740-82eedd4716f3\") " pod="openstack-operators/infra-operator-controller-manager-658588b8c9-z7tq9" Oct 05 21:08:18 crc kubenswrapper[4754]: I1005 21:08:18.049729 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-8r2mt" Oct 05 21:08:18 crc kubenswrapper[4754]: I1005 21:08:18.052543 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-8d984cc4d-7xzhb"] Oct 05 21:08:18 crc kubenswrapper[4754]: I1005 21:08:18.052872 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-54876c876f-svzwt" Oct 05 21:08:18 crc kubenswrapper[4754]: I1005 21:08:18.057245 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-7c7fc454ff-4fb5b"] Oct 05 21:08:18 crc kubenswrapper[4754]: I1005 21:08:18.058444 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-7c7fc454ff-4fb5b" Oct 05 21:08:18 crc kubenswrapper[4754]: I1005 21:08:18.059951 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-7468f855d8-w8z85"] Oct 05 21:08:18 crc kubenswrapper[4754]: I1005 21:08:18.060692 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-gqn8r" Oct 05 21:08:18 crc kubenswrapper[4754]: I1005 21:08:18.078933 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-7c7fc454ff-4fb5b"] Oct 05 21:08:18 crc kubenswrapper[4754]: I1005 21:08:18.101698 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9xbvr\" (UniqueName: \"kubernetes.io/projected/3ddbeeee-8a3e-4f3c-97e0-c7d8d5929345-kube-api-access-9xbvr\") pod \"ironic-operator-controller-manager-649675d675-v7285\" (UID: \"3ddbeeee-8a3e-4f3c-97e0-c7d8d5929345\") " pod="openstack-operators/ironic-operator-controller-manager-649675d675-v7285" Oct 05 21:08:18 crc kubenswrapper[4754]: I1005 21:08:18.101764 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bmvvr\" (UniqueName: \"kubernetes.io/projected/aee8b142-dea8-4ab3-87e9-14f3d3a75543-kube-api-access-bmvvr\") pod \"redhat-operators-456m8\" (UID: \"aee8b142-dea8-4ab3-87e9-14f3d3a75543\") " pod="openshift-marketplace/redhat-operators-456m8" Oct 05 21:08:18 crc kubenswrapper[4754]: I1005 21:08:18.101788 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aee8b142-dea8-4ab3-87e9-14f3d3a75543-catalog-content\") pod \"redhat-operators-456m8\" (UID: \"aee8b142-dea8-4ab3-87e9-14f3d3a75543\") " pod="openshift-marketplace/redhat-operators-456m8" Oct 05 21:08:18 crc kubenswrapper[4754]: I1005 21:08:18.101821 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aee8b142-dea8-4ab3-87e9-14f3d3a75543-utilities\") pod \"redhat-operators-456m8\" (UID: \"aee8b142-dea8-4ab3-87e9-14f3d3a75543\") " pod="openshift-marketplace/redhat-operators-456m8" Oct 05 21:08:18 crc kubenswrapper[4754]: I1005 21:08:18.101862 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kcq9f\" (UniqueName: \"kubernetes.io/projected/5a1312fe-c6d6-41de-a442-b044db904a6d-kube-api-access-kcq9f\") pod \"mariadb-operator-controller-manager-6cd6d7bdf5-f5jxx\" (UID: \"5a1312fe-c6d6-41de-a442-b044db904a6d\") " pod="openstack-operators/mariadb-operator-controller-manager-6cd6d7bdf5-f5jxx" Oct 05 21:08:18 crc kubenswrapper[4754]: I1005 21:08:18.101899 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zbcs6\" (UniqueName: \"kubernetes.io/projected/5757c596-103f-4a00-ac16-ed9e6c9e4719-kube-api-access-zbcs6\") pod \"manila-operator-controller-manager-65d89cfd9f-msh87\" (UID: \"5757c596-103f-4a00-ac16-ed9e6c9e4719\") " pod="openstack-operators/manila-operator-controller-manager-65d89cfd9f-msh87" Oct 05 21:08:18 crc kubenswrapper[4754]: I1005 21:08:18.101918 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x825r\" (UniqueName: \"kubernetes.io/projected/2735f261-cad9-4dff-9136-931a45c85ac5-kube-api-access-x825r\") pod \"keystone-operator-controller-manager-655d88ccb9-l56qv\" (UID: \"2735f261-cad9-4dff-9136-931a45c85ac5\") " pod="openstack-operators/keystone-operator-controller-manager-655d88ccb9-l56qv" Oct 05 21:08:18 crc kubenswrapper[4754]: I1005 21:08:18.110331 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665cfq946"] Oct 05 21:08:18 crc kubenswrapper[4754]: I1005 21:08:18.111354 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665cfq946" Oct 05 21:08:18 crc kubenswrapper[4754]: I1005 21:08:18.123556 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Oct 05 21:08:18 crc kubenswrapper[4754]: I1005 21:08:18.123759 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-gd786" Oct 05 21:08:18 crc kubenswrapper[4754]: I1005 21:08:18.128349 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x825r\" (UniqueName: \"kubernetes.io/projected/2735f261-cad9-4dff-9136-931a45c85ac5-kube-api-access-x825r\") pod \"keystone-operator-controller-manager-655d88ccb9-l56qv\" (UID: \"2735f261-cad9-4dff-9136-931a45c85ac5\") " pod="openstack-operators/keystone-operator-controller-manager-655d88ccb9-l56qv" Oct 05 21:08:18 crc kubenswrapper[4754]: I1005 21:08:18.140793 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-5f7c849b98-n9ksx" Oct 05 21:08:18 crc kubenswrapper[4754]: I1005 21:08:18.146636 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665cfq946"] Oct 05 21:08:18 crc kubenswrapper[4754]: I1005 21:08:18.160225 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kcq9f\" (UniqueName: \"kubernetes.io/projected/5a1312fe-c6d6-41de-a442-b044db904a6d-kube-api-access-kcq9f\") pod \"mariadb-operator-controller-manager-6cd6d7bdf5-f5jxx\" (UID: \"5a1312fe-c6d6-41de-a442-b044db904a6d\") " pod="openstack-operators/mariadb-operator-controller-manager-6cd6d7bdf5-f5jxx" Oct 05 21:08:18 crc kubenswrapper[4754]: I1005 21:08:18.161891 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zbcs6\" (UniqueName: \"kubernetes.io/projected/5757c596-103f-4a00-ac16-ed9e6c9e4719-kube-api-access-zbcs6\") pod \"manila-operator-controller-manager-65d89cfd9f-msh87\" (UID: \"5757c596-103f-4a00-ac16-ed9e6c9e4719\") " pod="openstack-operators/manila-operator-controller-manager-65d89cfd9f-msh87" Oct 05 21:08:18 crc kubenswrapper[4754]: I1005 21:08:18.185110 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9xbvr\" (UniqueName: \"kubernetes.io/projected/3ddbeeee-8a3e-4f3c-97e0-c7d8d5929345-kube-api-access-9xbvr\") pod \"ironic-operator-controller-manager-649675d675-v7285\" (UID: \"3ddbeeee-8a3e-4f3c-97e0-c7d8d5929345\") " pod="openstack-operators/ironic-operator-controller-manager-649675d675-v7285" Oct 05 21:08:18 crc kubenswrapper[4754]: I1005 21:08:18.187302 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-7d4d4f8d-kcxvd" Oct 05 21:08:18 crc kubenswrapper[4754]: I1005 21:08:18.191791 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-655d88ccb9-l56qv" Oct 05 21:08:18 crc kubenswrapper[4754]: I1005 21:08:18.201036 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-5568b5d68-f5rjh" Oct 05 21:08:18 crc kubenswrapper[4754]: I1005 21:08:18.202342 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-54689d9f88-4lrzz"] Oct 05 21:08:18 crc kubenswrapper[4754]: I1005 21:08:18.203107 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vmrg7\" (UniqueName: \"kubernetes.io/projected/ff3ba6ea-c2d8-4a3e-b0a0-58fd6df5b196-kube-api-access-vmrg7\") pod \"neutron-operator-controller-manager-8d984cc4d-7xzhb\" (UID: \"ff3ba6ea-c2d8-4a3e-b0a0-58fd6df5b196\") " pod="openstack-operators/neutron-operator-controller-manager-8d984cc4d-7xzhb" Oct 05 21:08:18 crc kubenswrapper[4754]: I1005 21:08:18.203170 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n6f52\" (UniqueName: \"kubernetes.io/projected/c336d534-f819-4859-8e15-5f15b68a36ad-kube-api-access-n6f52\") pod \"openstack-baremetal-operator-controller-manager-5dfbbd665cfq946\" (UID: \"c336d534-f819-4859-8e15-5f15b68a36ad\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665cfq946" Oct 05 21:08:18 crc kubenswrapper[4754]: I1005 21:08:18.203250 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bmvvr\" (UniqueName: \"kubernetes.io/projected/aee8b142-dea8-4ab3-87e9-14f3d3a75543-kube-api-access-bmvvr\") pod \"redhat-operators-456m8\" (UID: \"aee8b142-dea8-4ab3-87e9-14f3d3a75543\") " pod="openshift-marketplace/redhat-operators-456m8" Oct 05 21:08:18 crc kubenswrapper[4754]: I1005 21:08:18.203286 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aee8b142-dea8-4ab3-87e9-14f3d3a75543-catalog-content\") pod \"redhat-operators-456m8\" (UID: \"aee8b142-dea8-4ab3-87e9-14f3d3a75543\") " pod="openshift-marketplace/redhat-operators-456m8" Oct 05 21:08:18 crc kubenswrapper[4754]: I1005 21:08:18.203308 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fvkwj\" (UniqueName: \"kubernetes.io/projected/e0b701ef-8b8d-4717-aa27-07233f6b6c15-kube-api-access-fvkwj\") pod \"nova-operator-controller-manager-7c7fc454ff-4fb5b\" (UID: \"e0b701ef-8b8d-4717-aa27-07233f6b6c15\") " pod="openstack-operators/nova-operator-controller-manager-7c7fc454ff-4fb5b" Oct 05 21:08:18 crc kubenswrapper[4754]: I1005 21:08:18.203326 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aee8b142-dea8-4ab3-87e9-14f3d3a75543-utilities\") pod \"redhat-operators-456m8\" (UID: \"aee8b142-dea8-4ab3-87e9-14f3d3a75543\") " pod="openshift-marketplace/redhat-operators-456m8" Oct 05 21:08:18 crc kubenswrapper[4754]: I1005 21:08:18.203353 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w6n8v\" (UniqueName: \"kubernetes.io/projected/cb4a4798-9283-4633-9b3b-e2d72faa221f-kube-api-access-w6n8v\") pod \"octavia-operator-controller-manager-7468f855d8-w8z85\" (UID: \"cb4a4798-9283-4633-9b3b-e2d72faa221f\") " pod="openstack-operators/octavia-operator-controller-manager-7468f855d8-w8z85" Oct 05 21:08:18 crc kubenswrapper[4754]: I1005 21:08:18.203390 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c336d534-f819-4859-8e15-5f15b68a36ad-cert\") pod \"openstack-baremetal-operator-controller-manager-5dfbbd665cfq946\" (UID: \"c336d534-f819-4859-8e15-5f15b68a36ad\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665cfq946" Oct 05 21:08:18 crc kubenswrapper[4754]: I1005 21:08:18.203403 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-54689d9f88-4lrzz" Oct 05 21:08:18 crc kubenswrapper[4754]: I1005 21:08:18.204355 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aee8b142-dea8-4ab3-87e9-14f3d3a75543-catalog-content\") pod \"redhat-operators-456m8\" (UID: \"aee8b142-dea8-4ab3-87e9-14f3d3a75543\") " pod="openshift-marketplace/redhat-operators-456m8" Oct 05 21:08:18 crc kubenswrapper[4754]: I1005 21:08:18.204567 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aee8b142-dea8-4ab3-87e9-14f3d3a75543-utilities\") pod \"redhat-operators-456m8\" (UID: \"aee8b142-dea8-4ab3-87e9-14f3d3a75543\") " pod="openshift-marketplace/redhat-operators-456m8" Oct 05 21:08:18 crc kubenswrapper[4754]: I1005 21:08:18.209401 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-rq2bd" Oct 05 21:08:18 crc kubenswrapper[4754]: I1005 21:08:18.248234 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-8f58bc9db-xsrhv" Oct 05 21:08:18 crc kubenswrapper[4754]: I1005 21:08:18.257193 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-54689d9f88-4lrzz"] Oct 05 21:08:18 crc kubenswrapper[4754]: I1005 21:08:18.290686 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bmvvr\" (UniqueName: \"kubernetes.io/projected/aee8b142-dea8-4ab3-87e9-14f3d3a75543-kube-api-access-bmvvr\") pod \"redhat-operators-456m8\" (UID: \"aee8b142-dea8-4ab3-87e9-14f3d3a75543\") " pod="openshift-marketplace/redhat-operators-456m8" Oct 05 21:08:18 crc kubenswrapper[4754]: I1005 21:08:18.299602 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-579449c7d5-zcct6"] Oct 05 21:08:18 crc kubenswrapper[4754]: I1005 21:08:18.300988 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-579449c7d5-zcct6" Oct 05 21:08:18 crc kubenswrapper[4754]: I1005 21:08:18.301998 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-65d89cfd9f-msh87" Oct 05 21:08:18 crc kubenswrapper[4754]: I1005 21:08:18.304510 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fvkwj\" (UniqueName: \"kubernetes.io/projected/e0b701ef-8b8d-4717-aa27-07233f6b6c15-kube-api-access-fvkwj\") pod \"nova-operator-controller-manager-7c7fc454ff-4fb5b\" (UID: \"e0b701ef-8b8d-4717-aa27-07233f6b6c15\") " pod="openstack-operators/nova-operator-controller-manager-7c7fc454ff-4fb5b" Oct 05 21:08:18 crc kubenswrapper[4754]: I1005 21:08:18.304575 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w6n8v\" (UniqueName: \"kubernetes.io/projected/cb4a4798-9283-4633-9b3b-e2d72faa221f-kube-api-access-w6n8v\") pod \"octavia-operator-controller-manager-7468f855d8-w8z85\" (UID: \"cb4a4798-9283-4633-9b3b-e2d72faa221f\") " pod="openstack-operators/octavia-operator-controller-manager-7468f855d8-w8z85" Oct 05 21:08:18 crc kubenswrapper[4754]: I1005 21:08:18.304610 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c336d534-f819-4859-8e15-5f15b68a36ad-cert\") pod \"openstack-baremetal-operator-controller-manager-5dfbbd665cfq946\" (UID: \"c336d534-f819-4859-8e15-5f15b68a36ad\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665cfq946" Oct 05 21:08:18 crc kubenswrapper[4754]: I1005 21:08:18.304650 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vmrg7\" (UniqueName: \"kubernetes.io/projected/ff3ba6ea-c2d8-4a3e-b0a0-58fd6df5b196-kube-api-access-vmrg7\") pod \"neutron-operator-controller-manager-8d984cc4d-7xzhb\" (UID: \"ff3ba6ea-c2d8-4a3e-b0a0-58fd6df5b196\") " pod="openstack-operators/neutron-operator-controller-manager-8d984cc4d-7xzhb" Oct 05 21:08:18 crc kubenswrapper[4754]: I1005 21:08:18.304671 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n6f52\" (UniqueName: \"kubernetes.io/projected/c336d534-f819-4859-8e15-5f15b68a36ad-kube-api-access-n6f52\") pod \"openstack-baremetal-operator-controller-manager-5dfbbd665cfq946\" (UID: \"c336d534-f819-4859-8e15-5f15b68a36ad\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665cfq946" Oct 05 21:08:18 crc kubenswrapper[4754]: I1005 21:08:18.304696 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mrgrx\" (UniqueName: \"kubernetes.io/projected/6316859b-8ab7-4b46-b1fa-c7f893d39a95-kube-api-access-mrgrx\") pod \"placement-operator-controller-manager-54689d9f88-4lrzz\" (UID: \"6316859b-8ab7-4b46-b1fa-c7f893d39a95\") " pod="openstack-operators/placement-operator-controller-manager-54689d9f88-4lrzz" Oct 05 21:08:18 crc kubenswrapper[4754]: E1005 21:08:18.305077 4754 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 05 21:08:18 crc kubenswrapper[4754]: E1005 21:08:18.305125 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c336d534-f819-4859-8e15-5f15b68a36ad-cert podName:c336d534-f819-4859-8e15-5f15b68a36ad nodeName:}" failed. No retries permitted until 2025-10-05 21:08:18.805109191 +0000 UTC m=+822.709227901 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/c336d534-f819-4859-8e15-5f15b68a36ad-cert") pod "openstack-baremetal-operator-controller-manager-5dfbbd665cfq946" (UID: "c336d534-f819-4859-8e15-5f15b68a36ad") : secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 05 21:08:18 crc kubenswrapper[4754]: I1005 21:08:18.305620 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-njw8l" Oct 05 21:08:18 crc kubenswrapper[4754]: I1005 21:08:18.313630 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-579449c7d5-zcct6"] Oct 05 21:08:18 crc kubenswrapper[4754]: I1005 21:08:18.333677 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-456m8" Oct 05 21:08:18 crc kubenswrapper[4754]: I1005 21:08:18.358282 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w6n8v\" (UniqueName: \"kubernetes.io/projected/cb4a4798-9283-4633-9b3b-e2d72faa221f-kube-api-access-w6n8v\") pod \"octavia-operator-controller-manager-7468f855d8-w8z85\" (UID: \"cb4a4798-9283-4633-9b3b-e2d72faa221f\") " pod="openstack-operators/octavia-operator-controller-manager-7468f855d8-w8z85" Oct 05 21:08:18 crc kubenswrapper[4754]: I1005 21:08:18.359633 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-6cd6d7bdf5-f5jxx" Oct 05 21:08:18 crc kubenswrapper[4754]: I1005 21:08:18.360414 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vmrg7\" (UniqueName: \"kubernetes.io/projected/ff3ba6ea-c2d8-4a3e-b0a0-58fd6df5b196-kube-api-access-vmrg7\") pod \"neutron-operator-controller-manager-8d984cc4d-7xzhb\" (UID: \"ff3ba6ea-c2d8-4a3e-b0a0-58fd6df5b196\") " pod="openstack-operators/neutron-operator-controller-manager-8d984cc4d-7xzhb" Oct 05 21:08:18 crc kubenswrapper[4754]: I1005 21:08:18.363355 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n6f52\" (UniqueName: \"kubernetes.io/projected/c336d534-f819-4859-8e15-5f15b68a36ad-kube-api-access-n6f52\") pod \"openstack-baremetal-operator-controller-manager-5dfbbd665cfq946\" (UID: \"c336d534-f819-4859-8e15-5f15b68a36ad\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665cfq946" Oct 05 21:08:18 crc kubenswrapper[4754]: I1005 21:08:18.363435 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-6859f9b676-hgkv4"] Oct 05 21:08:18 crc kubenswrapper[4754]: I1005 21:08:18.374376 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fvkwj\" (UniqueName: \"kubernetes.io/projected/e0b701ef-8b8d-4717-aa27-07233f6b6c15-kube-api-access-fvkwj\") pod \"nova-operator-controller-manager-7c7fc454ff-4fb5b\" (UID: \"e0b701ef-8b8d-4717-aa27-07233f6b6c15\") " pod="openstack-operators/nova-operator-controller-manager-7c7fc454ff-4fb5b" Oct 05 21:08:18 crc kubenswrapper[4754]: I1005 21:08:18.384970 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-8d984cc4d-7xzhb" Oct 05 21:08:18 crc kubenswrapper[4754]: I1005 21:08:18.417752 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-7468f855d8-w8z85" Oct 05 21:08:18 crc kubenswrapper[4754]: I1005 21:08:18.456729 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mrgrx\" (UniqueName: \"kubernetes.io/projected/6316859b-8ab7-4b46-b1fa-c7f893d39a95-kube-api-access-mrgrx\") pod \"placement-operator-controller-manager-54689d9f88-4lrzz\" (UID: \"6316859b-8ab7-4b46-b1fa-c7f893d39a95\") " pod="openstack-operators/placement-operator-controller-manager-54689d9f88-4lrzz" Oct 05 21:08:18 crc kubenswrapper[4754]: I1005 21:08:18.456858 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kk8vw\" (UniqueName: \"kubernetes.io/projected/b18dee24-0d55-4805-8d76-623180998686-kube-api-access-kk8vw\") pod \"ovn-operator-controller-manager-579449c7d5-zcct6\" (UID: \"b18dee24-0d55-4805-8d76-623180998686\") " pod="openstack-operators/ovn-operator-controller-manager-579449c7d5-zcct6" Oct 05 21:08:18 crc kubenswrapper[4754]: I1005 21:08:18.457895 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-7c7fc454ff-4fb5b" Oct 05 21:08:18 crc kubenswrapper[4754]: I1005 21:08:18.470867 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-649675d675-v7285" Oct 05 21:08:18 crc kubenswrapper[4754]: I1005 21:08:18.521391 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-5d4d74dd89-h44bd"] Oct 05 21:08:18 crc kubenswrapper[4754]: I1005 21:08:18.522809 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-5d4d74dd89-h44bd" Oct 05 21:08:18 crc kubenswrapper[4754]: I1005 21:08:18.527796 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-6859f9b676-hgkv4" Oct 05 21:08:18 crc kubenswrapper[4754]: I1005 21:08:18.536244 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-6859f9b676-hgkv4"] Oct 05 21:08:18 crc kubenswrapper[4754]: I1005 21:08:18.540251 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-8x96z" Oct 05 21:08:18 crc kubenswrapper[4754]: I1005 21:08:18.550034 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-gthnf" Oct 05 21:08:18 crc kubenswrapper[4754]: I1005 21:08:18.562886 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kk8vw\" (UniqueName: \"kubernetes.io/projected/b18dee24-0d55-4805-8d76-623180998686-kube-api-access-kk8vw\") pod \"ovn-operator-controller-manager-579449c7d5-zcct6\" (UID: \"b18dee24-0d55-4805-8d76-623180998686\") " pod="openstack-operators/ovn-operator-controller-manager-579449c7d5-zcct6" Oct 05 21:08:18 crc kubenswrapper[4754]: I1005 21:08:18.562967 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4bb20421-ce34-44ee-8740-82eedd4716f3-cert\") pod \"infra-operator-controller-manager-658588b8c9-z7tq9\" (UID: \"4bb20421-ce34-44ee-8740-82eedd4716f3\") " pod="openstack-operators/infra-operator-controller-manager-658588b8c9-z7tq9" Oct 05 21:08:18 crc kubenswrapper[4754]: I1005 21:08:18.618344 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-5d4d74dd89-h44bd"] Oct 05 21:08:18 crc kubenswrapper[4754]: I1005 21:08:18.624422 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mrgrx\" (UniqueName: \"kubernetes.io/projected/6316859b-8ab7-4b46-b1fa-c7f893d39a95-kube-api-access-mrgrx\") pod \"placement-operator-controller-manager-54689d9f88-4lrzz\" (UID: \"6316859b-8ab7-4b46-b1fa-c7f893d39a95\") " pod="openstack-operators/placement-operator-controller-manager-54689d9f88-4lrzz" Oct 05 21:08:18 crc kubenswrapper[4754]: I1005 21:08:18.625097 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4bb20421-ce34-44ee-8740-82eedd4716f3-cert\") pod \"infra-operator-controller-manager-658588b8c9-z7tq9\" (UID: \"4bb20421-ce34-44ee-8740-82eedd4716f3\") " pod="openstack-operators/infra-operator-controller-manager-658588b8c9-z7tq9" Oct 05 21:08:18 crc kubenswrapper[4754]: I1005 21:08:18.638761 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-5cd5cb47d7-fx25z"] Oct 05 21:08:18 crc kubenswrapper[4754]: I1005 21:08:18.639936 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-fx25z" Oct 05 21:08:18 crc kubenswrapper[4754]: I1005 21:08:18.642793 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kk8vw\" (UniqueName: \"kubernetes.io/projected/b18dee24-0d55-4805-8d76-623180998686-kube-api-access-kk8vw\") pod \"ovn-operator-controller-manager-579449c7d5-zcct6\" (UID: \"b18dee24-0d55-4805-8d76-623180998686\") " pod="openstack-operators/ovn-operator-controller-manager-579449c7d5-zcct6" Oct 05 21:08:18 crc kubenswrapper[4754]: I1005 21:08:18.645701 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-6rhqq" Oct 05 21:08:18 crc kubenswrapper[4754]: I1005 21:08:18.662677 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-579449c7d5-zcct6" Oct 05 21:08:18 crc kubenswrapper[4754]: I1005 21:08:18.664311 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-42qgx\" (UniqueName: \"kubernetes.io/projected/f48f3071-cd94-49dd-bfa0-1ec74f495c72-kube-api-access-42qgx\") pod \"telemetry-operator-controller-manager-5d4d74dd89-h44bd\" (UID: \"f48f3071-cd94-49dd-bfa0-1ec74f495c72\") " pod="openstack-operators/telemetry-operator-controller-manager-5d4d74dd89-h44bd" Oct 05 21:08:18 crc kubenswrapper[4754]: I1005 21:08:18.664389 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g95tf\" (UniqueName: \"kubernetes.io/projected/2b729715-7fd9-4ca5-b4dd-cb0eb0034aac-kube-api-access-g95tf\") pod \"swift-operator-controller-manager-6859f9b676-hgkv4\" (UID: \"2b729715-7fd9-4ca5-b4dd-cb0eb0034aac\") " pod="openstack-operators/swift-operator-controller-manager-6859f9b676-hgkv4" Oct 05 21:08:18 crc kubenswrapper[4754]: I1005 21:08:18.693883 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-6cbc6dd547-zq79b"] Oct 05 21:08:18 crc kubenswrapper[4754]: I1005 21:08:18.699105 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-6cbc6dd547-zq79b" Oct 05 21:08:18 crc kubenswrapper[4754]: I1005 21:08:18.701070 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-7kmx6" Oct 05 21:08:18 crc kubenswrapper[4754]: I1005 21:08:18.706601 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5cd5cb47d7-fx25z"] Oct 05 21:08:18 crc kubenswrapper[4754]: I1005 21:08:18.737343 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-6cbc6dd547-zq79b"] Oct 05 21:08:18 crc kubenswrapper[4754]: I1005 21:08:18.796264 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2jwmp\" (UniqueName: \"kubernetes.io/projected/b1cc32e7-f619-4a0e-b0f9-195e36c82d01-kube-api-access-2jwmp\") pod \"test-operator-controller-manager-5cd5cb47d7-fx25z\" (UID: \"b1cc32e7-f619-4a0e-b0f9-195e36c82d01\") " pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-fx25z" Oct 05 21:08:18 crc kubenswrapper[4754]: I1005 21:08:18.796597 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-42qgx\" (UniqueName: \"kubernetes.io/projected/f48f3071-cd94-49dd-bfa0-1ec74f495c72-kube-api-access-42qgx\") pod \"telemetry-operator-controller-manager-5d4d74dd89-h44bd\" (UID: \"f48f3071-cd94-49dd-bfa0-1ec74f495c72\") " pod="openstack-operators/telemetry-operator-controller-manager-5d4d74dd89-h44bd" Oct 05 21:08:18 crc kubenswrapper[4754]: I1005 21:08:18.796648 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g95tf\" (UniqueName: \"kubernetes.io/projected/2b729715-7fd9-4ca5-b4dd-cb0eb0034aac-kube-api-access-g95tf\") pod \"swift-operator-controller-manager-6859f9b676-hgkv4\" (UID: \"2b729715-7fd9-4ca5-b4dd-cb0eb0034aac\") " pod="openstack-operators/swift-operator-controller-manager-6859f9b676-hgkv4" Oct 05 21:08:18 crc kubenswrapper[4754]: I1005 21:08:18.811908 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-55895b89bc-zzbvh"] Oct 05 21:08:18 crc kubenswrapper[4754]: I1005 21:08:18.822002 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-55895b89bc-zzbvh" Oct 05 21:08:18 crc kubenswrapper[4754]: I1005 21:08:18.830294 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g95tf\" (UniqueName: \"kubernetes.io/projected/2b729715-7fd9-4ca5-b4dd-cb0eb0034aac-kube-api-access-g95tf\") pod \"swift-operator-controller-manager-6859f9b676-hgkv4\" (UID: \"2b729715-7fd9-4ca5-b4dd-cb0eb0034aac\") " pod="openstack-operators/swift-operator-controller-manager-6859f9b676-hgkv4" Oct 05 21:08:18 crc kubenswrapper[4754]: I1005 21:08:18.830697 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-hv5h8" Oct 05 21:08:18 crc kubenswrapper[4754]: I1005 21:08:18.831450 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Oct 05 21:08:18 crc kubenswrapper[4754]: I1005 21:08:18.843229 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-42qgx\" (UniqueName: \"kubernetes.io/projected/f48f3071-cd94-49dd-bfa0-1ec74f495c72-kube-api-access-42qgx\") pod \"telemetry-operator-controller-manager-5d4d74dd89-h44bd\" (UID: \"f48f3071-cd94-49dd-bfa0-1ec74f495c72\") " pod="openstack-operators/telemetry-operator-controller-manager-5d4d74dd89-h44bd" Oct 05 21:08:18 crc kubenswrapper[4754]: I1005 21:08:18.858004 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-54689d9f88-4lrzz" Oct 05 21:08:18 crc kubenswrapper[4754]: I1005 21:08:18.897445 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-658588b8c9-z7tq9" Oct 05 21:08:19 crc kubenswrapper[4754]: I1005 21:08:18.898870 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2jwmp\" (UniqueName: \"kubernetes.io/projected/b1cc32e7-f619-4a0e-b0f9-195e36c82d01-kube-api-access-2jwmp\") pod \"test-operator-controller-manager-5cd5cb47d7-fx25z\" (UID: \"b1cc32e7-f619-4a0e-b0f9-195e36c82d01\") " pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-fx25z" Oct 05 21:08:19 crc kubenswrapper[4754]: I1005 21:08:18.898931 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c336d534-f819-4859-8e15-5f15b68a36ad-cert\") pod \"openstack-baremetal-operator-controller-manager-5dfbbd665cfq946\" (UID: \"c336d534-f819-4859-8e15-5f15b68a36ad\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665cfq946" Oct 05 21:08:19 crc kubenswrapper[4754]: I1005 21:08:18.898967 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zf796\" (UniqueName: \"kubernetes.io/projected/cea6487f-b088-4f9e-a42d-6a3517df7669-kube-api-access-zf796\") pod \"watcher-operator-controller-manager-6cbc6dd547-zq79b\" (UID: \"cea6487f-b088-4f9e-a42d-6a3517df7669\") " pod="openstack-operators/watcher-operator-controller-manager-6cbc6dd547-zq79b" Oct 05 21:08:19 crc kubenswrapper[4754]: E1005 21:08:18.899458 4754 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 05 21:08:19 crc kubenswrapper[4754]: E1005 21:08:18.899533 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c336d534-f819-4859-8e15-5f15b68a36ad-cert podName:c336d534-f819-4859-8e15-5f15b68a36ad nodeName:}" failed. No retries permitted until 2025-10-05 21:08:19.899485462 +0000 UTC m=+823.803604172 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/c336d534-f819-4859-8e15-5f15b68a36ad-cert") pod "openstack-baremetal-operator-controller-manager-5dfbbd665cfq946" (UID: "c336d534-f819-4859-8e15-5f15b68a36ad") : secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 05 21:08:19 crc kubenswrapper[4754]: I1005 21:08:18.930931 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-5d4d74dd89-h44bd" Oct 05 21:08:19 crc kubenswrapper[4754]: I1005 21:08:18.961036 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2jwmp\" (UniqueName: \"kubernetes.io/projected/b1cc32e7-f619-4a0e-b0f9-195e36c82d01-kube-api-access-2jwmp\") pod \"test-operator-controller-manager-5cd5cb47d7-fx25z\" (UID: \"b1cc32e7-f619-4a0e-b0f9-195e36c82d01\") " pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-fx25z" Oct 05 21:08:19 crc kubenswrapper[4754]: I1005 21:08:18.980135 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-6859f9b676-hgkv4" Oct 05 21:08:19 crc kubenswrapper[4754]: I1005 21:08:19.003371 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zf796\" (UniqueName: \"kubernetes.io/projected/cea6487f-b088-4f9e-a42d-6a3517df7669-kube-api-access-zf796\") pod \"watcher-operator-controller-manager-6cbc6dd547-zq79b\" (UID: \"cea6487f-b088-4f9e-a42d-6a3517df7669\") " pod="openstack-operators/watcher-operator-controller-manager-6cbc6dd547-zq79b" Oct 05 21:08:19 crc kubenswrapper[4754]: I1005 21:08:19.003415 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4sr9p\" (UniqueName: \"kubernetes.io/projected/f131e747-362b-4442-8377-0c00aedfd8ae-kube-api-access-4sr9p\") pod \"openstack-operator-controller-manager-55895b89bc-zzbvh\" (UID: \"f131e747-362b-4442-8377-0c00aedfd8ae\") " pod="openstack-operators/openstack-operator-controller-manager-55895b89bc-zzbvh" Oct 05 21:08:19 crc kubenswrapper[4754]: I1005 21:08:19.003543 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f131e747-362b-4442-8377-0c00aedfd8ae-cert\") pod \"openstack-operator-controller-manager-55895b89bc-zzbvh\" (UID: \"f131e747-362b-4442-8377-0c00aedfd8ae\") " pod="openstack-operators/openstack-operator-controller-manager-55895b89bc-zzbvh" Oct 05 21:08:19 crc kubenswrapper[4754]: I1005 21:08:19.059598 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zf796\" (UniqueName: \"kubernetes.io/projected/cea6487f-b088-4f9e-a42d-6a3517df7669-kube-api-access-zf796\") pod \"watcher-operator-controller-manager-6cbc6dd547-zq79b\" (UID: \"cea6487f-b088-4f9e-a42d-6a3517df7669\") " pod="openstack-operators/watcher-operator-controller-manager-6cbc6dd547-zq79b" Oct 05 21:08:19 crc kubenswrapper[4754]: I1005 21:08:19.107416 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f131e747-362b-4442-8377-0c00aedfd8ae-cert\") pod \"openstack-operator-controller-manager-55895b89bc-zzbvh\" (UID: \"f131e747-362b-4442-8377-0c00aedfd8ae\") " pod="openstack-operators/openstack-operator-controller-manager-55895b89bc-zzbvh" Oct 05 21:08:19 crc kubenswrapper[4754]: I1005 21:08:19.107501 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4sr9p\" (UniqueName: \"kubernetes.io/projected/f131e747-362b-4442-8377-0c00aedfd8ae-kube-api-access-4sr9p\") pod \"openstack-operator-controller-manager-55895b89bc-zzbvh\" (UID: \"f131e747-362b-4442-8377-0c00aedfd8ae\") " pod="openstack-operators/openstack-operator-controller-manager-55895b89bc-zzbvh" Oct 05 21:08:19 crc kubenswrapper[4754]: E1005 21:08:19.107914 4754 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Oct 05 21:08:19 crc kubenswrapper[4754]: E1005 21:08:19.107969 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f131e747-362b-4442-8377-0c00aedfd8ae-cert podName:f131e747-362b-4442-8377-0c00aedfd8ae nodeName:}" failed. No retries permitted until 2025-10-05 21:08:19.607942806 +0000 UTC m=+823.512061506 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/f131e747-362b-4442-8377-0c00aedfd8ae-cert") pod "openstack-operator-controller-manager-55895b89bc-zzbvh" (UID: "f131e747-362b-4442-8377-0c00aedfd8ae") : secret "webhook-server-cert" not found Oct 05 21:08:19 crc kubenswrapper[4754]: I1005 21:08:19.133075 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4sr9p\" (UniqueName: \"kubernetes.io/projected/f131e747-362b-4442-8377-0c00aedfd8ae-kube-api-access-4sr9p\") pod \"openstack-operator-controller-manager-55895b89bc-zzbvh\" (UID: \"f131e747-362b-4442-8377-0c00aedfd8ae\") " pod="openstack-operators/openstack-operator-controller-manager-55895b89bc-zzbvh" Oct 05 21:08:19 crc kubenswrapper[4754]: I1005 21:08:19.133704 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-55895b89bc-zzbvh"] Oct 05 21:08:19 crc kubenswrapper[4754]: I1005 21:08:19.133748 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-hkb96"] Oct 05 21:08:19 crc kubenswrapper[4754]: I1005 21:08:19.134508 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-hkb96"] Oct 05 21:08:19 crc kubenswrapper[4754]: I1005 21:08:19.134540 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-75dfd9b554-fxx5s"] Oct 05 21:08:19 crc kubenswrapper[4754]: I1005 21:08:19.134612 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-hkb96" Oct 05 21:08:19 crc kubenswrapper[4754]: I1005 21:08:19.136690 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-wb5gb" Oct 05 21:08:19 crc kubenswrapper[4754]: I1005 21:08:19.211653 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wbkkf\" (UniqueName: \"kubernetes.io/projected/97f32c3b-569a-4f15-9f7c-fb18299dec30-kube-api-access-wbkkf\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-hkb96\" (UID: \"97f32c3b-569a-4f15-9f7c-fb18299dec30\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-hkb96" Oct 05 21:08:19 crc kubenswrapper[4754]: I1005 21:08:19.285864 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-fx25z" Oct 05 21:08:19 crc kubenswrapper[4754]: I1005 21:08:19.313253 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wbkkf\" (UniqueName: \"kubernetes.io/projected/97f32c3b-569a-4f15-9f7c-fb18299dec30-kube-api-access-wbkkf\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-hkb96\" (UID: \"97f32c3b-569a-4f15-9f7c-fb18299dec30\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-hkb96" Oct 05 21:08:19 crc kubenswrapper[4754]: I1005 21:08:19.320105 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-6cbc6dd547-zq79b" Oct 05 21:08:19 crc kubenswrapper[4754]: I1005 21:08:19.336310 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-54876c876f-svzwt"] Oct 05 21:08:19 crc kubenswrapper[4754]: I1005 21:08:19.365755 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wbkkf\" (UniqueName: \"kubernetes.io/projected/97f32c3b-569a-4f15-9f7c-fb18299dec30-kube-api-access-wbkkf\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-hkb96\" (UID: \"97f32c3b-569a-4f15-9f7c-fb18299dec30\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-hkb96" Oct 05 21:08:19 crc kubenswrapper[4754]: I1005 21:08:19.390171 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-hkb96" Oct 05 21:08:19 crc kubenswrapper[4754]: W1005 21:08:19.476741 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf560133d_b3b5_4791_b69a_a700367f0d96.slice/crio-f9a936c6e3997fb1440695f52dc1ea5565a8dbfdb8bfc5499e1e7e203ac04f4d WatchSource:0}: Error finding container f9a936c6e3997fb1440695f52dc1ea5565a8dbfdb8bfc5499e1e7e203ac04f4d: Status 404 returned error can't find the container with id f9a936c6e3997fb1440695f52dc1ea5565a8dbfdb8bfc5499e1e7e203ac04f4d Oct 05 21:08:19 crc kubenswrapper[4754]: I1005 21:08:19.599153 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-fxx5s" event={"ID":"3ec442d5-220a-4d2f-8f62-a13ff86a3229","Type":"ContainerStarted","Data":"946100fd6ce3ddf94195d72c4b8df7f207ead57f067df6b63482b5c7580948c6"} Oct 05 21:08:19 crc kubenswrapper[4754]: I1005 21:08:19.625083 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f131e747-362b-4442-8377-0c00aedfd8ae-cert\") pod \"openstack-operator-controller-manager-55895b89bc-zzbvh\" (UID: \"f131e747-362b-4442-8377-0c00aedfd8ae\") " pod="openstack-operators/openstack-operator-controller-manager-55895b89bc-zzbvh" Oct 05 21:08:19 crc kubenswrapper[4754]: I1005 21:08:19.644762 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-54876c876f-svzwt" event={"ID":"f560133d-b3b5-4791-b69a-a700367f0d96","Type":"ContainerStarted","Data":"f9a936c6e3997fb1440695f52dc1ea5565a8dbfdb8bfc5499e1e7e203ac04f4d"} Oct 05 21:08:19 crc kubenswrapper[4754]: I1005 21:08:19.645762 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f131e747-362b-4442-8377-0c00aedfd8ae-cert\") pod \"openstack-operator-controller-manager-55895b89bc-zzbvh\" (UID: \"f131e747-362b-4442-8377-0c00aedfd8ae\") " pod="openstack-operators/openstack-operator-controller-manager-55895b89bc-zzbvh" Oct 05 21:08:19 crc kubenswrapper[4754]: I1005 21:08:19.655929 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-55895b89bc-zzbvh" Oct 05 21:08:19 crc kubenswrapper[4754]: I1005 21:08:19.709660 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-5568b5d68-f5rjh"] Oct 05 21:08:19 crc kubenswrapper[4754]: I1005 21:08:19.734349 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-5f7c849b98-n9ksx"] Oct 05 21:08:19 crc kubenswrapper[4754]: W1005 21:08:19.801298 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9a1cc24e_8371_44cb_bd37_5b765853fdff.slice/crio-3c3478e9dcc758540a00008015f6f7b7b67cb739d2e4155897746181fff6cad1 WatchSource:0}: Error finding container 3c3478e9dcc758540a00008015f6f7b7b67cb739d2e4155897746181fff6cad1: Status 404 returned error can't find the container with id 3c3478e9dcc758540a00008015f6f7b7b67cb739d2e4155897746181fff6cad1 Oct 05 21:08:19 crc kubenswrapper[4754]: I1005 21:08:19.929156 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c336d534-f819-4859-8e15-5f15b68a36ad-cert\") pod \"openstack-baremetal-operator-controller-manager-5dfbbd665cfq946\" (UID: \"c336d534-f819-4859-8e15-5f15b68a36ad\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665cfq946" Oct 05 21:08:19 crc kubenswrapper[4754]: I1005 21:08:19.955427 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c336d534-f819-4859-8e15-5f15b68a36ad-cert\") pod \"openstack-baremetal-operator-controller-manager-5dfbbd665cfq946\" (UID: \"c336d534-f819-4859-8e15-5f15b68a36ad\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665cfq946" Oct 05 21:08:19 crc kubenswrapper[4754]: I1005 21:08:19.966009 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665cfq946" Oct 05 21:08:20 crc kubenswrapper[4754]: I1005 21:08:20.058259 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-7d4d4f8d-kcxvd"] Oct 05 21:08:20 crc kubenswrapper[4754]: I1005 21:08:20.065812 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-655d88ccb9-l56qv"] Oct 05 21:08:20 crc kubenswrapper[4754]: I1005 21:08:20.096313 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-8f58bc9db-xsrhv"] Oct 05 21:08:20 crc kubenswrapper[4754]: W1005 21:08:20.111969 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2735f261_cad9_4dff_9136_931a45c85ac5.slice/crio-541cb0ee7ab80f1b515d554a73fcd413cb9f9dc09e723b26162cd4efe5458e6e WatchSource:0}: Error finding container 541cb0ee7ab80f1b515d554a73fcd413cb9f9dc09e723b26162cd4efe5458e6e: Status 404 returned error can't find the container with id 541cb0ee7ab80f1b515d554a73fcd413cb9f9dc09e723b26162cd4efe5458e6e Oct 05 21:08:20 crc kubenswrapper[4754]: I1005 21:08:20.207912 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-6cd6d7bdf5-f5jxx"] Oct 05 21:08:20 crc kubenswrapper[4754]: W1005 21:08:20.210686 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5a1312fe_c6d6_41de_a442_b044db904a6d.slice/crio-fbc706f9435caa1a706ec45ae98b493154389ebcda9eb1a7e73f1e46d11e6372 WatchSource:0}: Error finding container fbc706f9435caa1a706ec45ae98b493154389ebcda9eb1a7e73f1e46d11e6372: Status 404 returned error can't find the container with id fbc706f9435caa1a706ec45ae98b493154389ebcda9eb1a7e73f1e46d11e6372 Oct 05 21:08:20 crc kubenswrapper[4754]: I1005 21:08:20.534779 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-456m8"] Oct 05 21:08:20 crc kubenswrapper[4754]: W1005 21:08:20.563825 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaee8b142_dea8_4ab3_87e9_14f3d3a75543.slice/crio-43d9575fbca631b0b9c86dabf6db03f572663a8c9fc41e9b57e634d93ad5a71a WatchSource:0}: Error finding container 43d9575fbca631b0b9c86dabf6db03f572663a8c9fc41e9b57e634d93ad5a71a: Status 404 returned error can't find the container with id 43d9575fbca631b0b9c86dabf6db03f572663a8c9fc41e9b57e634d93ad5a71a Oct 05 21:08:20 crc kubenswrapper[4754]: I1005 21:08:20.577951 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-8d984cc4d-7xzhb"] Oct 05 21:08:20 crc kubenswrapper[4754]: I1005 21:08:20.590794 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-649675d675-v7285"] Oct 05 21:08:20 crc kubenswrapper[4754]: I1005 21:08:20.604245 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-579449c7d5-zcct6"] Oct 05 21:08:20 crc kubenswrapper[4754]: W1005 21:08:20.606409 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podff3ba6ea_c2d8_4a3e_b0a0_58fd6df5b196.slice/crio-f0c91280d19f7427fdf5e804eb10486c6d91d0bb80d5e504cf7b0dddce4216f4 WatchSource:0}: Error finding container f0c91280d19f7427fdf5e804eb10486c6d91d0bb80d5e504cf7b0dddce4216f4: Status 404 returned error can't find the container with id f0c91280d19f7427fdf5e804eb10486c6d91d0bb80d5e504cf7b0dddce4216f4 Oct 05 21:08:20 crc kubenswrapper[4754]: I1005 21:08:20.627005 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-54689d9f88-4lrzz"] Oct 05 21:08:20 crc kubenswrapper[4754]: I1005 21:08:20.654859 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-655d88ccb9-l56qv" event={"ID":"2735f261-cad9-4dff-9136-931a45c85ac5","Type":"ContainerStarted","Data":"541cb0ee7ab80f1b515d554a73fcd413cb9f9dc09e723b26162cd4efe5458e6e"} Oct 05 21:08:20 crc kubenswrapper[4754]: I1005 21:08:20.656294 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-8d984cc4d-7xzhb" event={"ID":"ff3ba6ea-c2d8-4a3e-b0a0-58fd6df5b196","Type":"ContainerStarted","Data":"f0c91280d19f7427fdf5e804eb10486c6d91d0bb80d5e504cf7b0dddce4216f4"} Oct 05 21:08:20 crc kubenswrapper[4754]: I1005 21:08:20.657717 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-5568b5d68-f5rjh" event={"ID":"62159d2d-ec19-4b44-89df-846faa5757dc","Type":"ContainerStarted","Data":"64077ee9aea91b0f47732275293563e9422773772f861b25ec3fb8e9c4b06073"} Oct 05 21:08:20 crc kubenswrapper[4754]: I1005 21:08:20.659322 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-649675d675-v7285" event={"ID":"3ddbeeee-8a3e-4f3c-97e0-c7d8d5929345","Type":"ContainerStarted","Data":"2cb3e8ddc278411e3179ba63ea871f9f9c414f1c4ac6993d6a4eae242c60bf9d"} Oct 05 21:08:20 crc kubenswrapper[4754]: I1005 21:08:20.660237 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-5f7c849b98-n9ksx" event={"ID":"9a1cc24e-8371-44cb-bd37-5b765853fdff","Type":"ContainerStarted","Data":"3c3478e9dcc758540a00008015f6f7b7b67cb739d2e4155897746181fff6cad1"} Oct 05 21:08:20 crc kubenswrapper[4754]: I1005 21:08:20.663534 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-6859f9b676-hgkv4"] Oct 05 21:08:20 crc kubenswrapper[4754]: I1005 21:08:20.670179 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-8f58bc9db-xsrhv" event={"ID":"82184642-9d7b-4ada-b7b2-efde6c91cb14","Type":"ContainerStarted","Data":"69d069ed842765406a54a24f120ab3e8c0260f3974b792345d36c6cd166fffc6"} Oct 05 21:08:20 crc kubenswrapper[4754]: W1005 21:08:20.670276 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb18dee24_0d55_4805_8d76_623180998686.slice/crio-914c3603d1a4a34d607d7e9c855c1abfcc28dcbd5ef14e107e9561aaacbf5718 WatchSource:0}: Error finding container 914c3603d1a4a34d607d7e9c855c1abfcc28dcbd5ef14e107e9561aaacbf5718: Status 404 returned error can't find the container with id 914c3603d1a4a34d607d7e9c855c1abfcc28dcbd5ef14e107e9561aaacbf5718 Oct 05 21:08:20 crc kubenswrapper[4754]: I1005 21:08:20.671430 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-7d4d4f8d-kcxvd" event={"ID":"47cafc06-0f75-48ee-bae5-4e30352f4572","Type":"ContainerStarted","Data":"b0e812ccb751e35ce0abd8f5c46584ab6019883e4b3561d3be9399f663627a61"} Oct 05 21:08:20 crc kubenswrapper[4754]: I1005 21:08:20.672966 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-456m8" event={"ID":"aee8b142-dea8-4ab3-87e9-14f3d3a75543","Type":"ContainerStarted","Data":"43d9575fbca631b0b9c86dabf6db03f572663a8c9fc41e9b57e634d93ad5a71a"} Oct 05 21:08:20 crc kubenswrapper[4754]: W1005 21:08:20.675180 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6316859b_8ab7_4b46_b1fa_c7f893d39a95.slice/crio-f060fd690273da29559a50e13b2c7012ffa46fedaeca007ef25681cdb00d1685 WatchSource:0}: Error finding container f060fd690273da29559a50e13b2c7012ffa46fedaeca007ef25681cdb00d1685: Status 404 returned error can't find the container with id f060fd690273da29559a50e13b2c7012ffa46fedaeca007ef25681cdb00d1685 Oct 05 21:08:20 crc kubenswrapper[4754]: I1005 21:08:20.675771 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-65d89cfd9f-msh87"] Oct 05 21:08:20 crc kubenswrapper[4754]: I1005 21:08:20.679938 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-6cd6d7bdf5-f5jxx" event={"ID":"5a1312fe-c6d6-41de-a442-b044db904a6d","Type":"ContainerStarted","Data":"fbc706f9435caa1a706ec45ae98b493154389ebcda9eb1a7e73f1e46d11e6372"} Oct 05 21:08:20 crc kubenswrapper[4754]: W1005 21:08:20.683635 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5757c596_103f_4a00_ac16_ed9e6c9e4719.slice/crio-d5bc437450a76209aee9c1cf907be0a2c03d03210d18fc0eedec889ebd3e2091 WatchSource:0}: Error finding container d5bc437450a76209aee9c1cf907be0a2c03d03210d18fc0eedec889ebd3e2091: Status 404 returned error can't find the container with id d5bc437450a76209aee9c1cf907be0a2c03d03210d18fc0eedec889ebd3e2091 Oct 05 21:08:20 crc kubenswrapper[4754]: I1005 21:08:20.919831 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-5d4d74dd89-h44bd"] Oct 05 21:08:20 crc kubenswrapper[4754]: I1005 21:08:20.989908 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-658588b8c9-z7tq9"] Oct 05 21:08:20 crc kubenswrapper[4754]: I1005 21:08:20.998665 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-7468f855d8-w8z85"] Oct 05 21:08:21 crc kubenswrapper[4754]: I1005 21:08:21.044798 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-7c7fc454ff-4fb5b"] Oct 05 21:08:21 crc kubenswrapper[4754]: I1005 21:08:21.055651 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5cd5cb47d7-fx25z"] Oct 05 21:08:21 crc kubenswrapper[4754]: E1005 21:08:21.106434 4754 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:0daf76cc40ab619ae266b11defcc1b65beb22d859369e7b1b04de9169089a4cb,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-2jwmp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5cd5cb47d7-fx25z_openstack-operators(b1cc32e7-f619-4a0e-b0f9-195e36c82d01): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 05 21:08:21 crc kubenswrapper[4754]: I1005 21:08:21.150110 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-hkb96"] Oct 05 21:08:21 crc kubenswrapper[4754]: I1005 21:08:21.165362 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-55895b89bc-zzbvh"] Oct 05 21:08:21 crc kubenswrapper[4754]: I1005 21:08:21.186010 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665cfq946"] Oct 05 21:08:21 crc kubenswrapper[4754]: I1005 21:08:21.187244 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-6cbc6dd547-zq79b"] Oct 05 21:08:21 crc kubenswrapper[4754]: E1005 21:08:21.221515 4754 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:64f57b2b59dea2bd9fae91490c5bec2687131884a049e6579819d9f951b877c6,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-zf796,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-6cbc6dd547-zq79b_openstack-operators(cea6487f-b088-4f9e-a42d-6a3517df7669): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 05 21:08:21 crc kubenswrapper[4754]: E1005 21:08:21.225375 4754 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:bcd1acac74e68eea5a9c3b7ba1bcb29d3a5b43423fc23c19ad4715bdac41f799,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:true,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-baremetal-operator-agent:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_ANSIBLEEE_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-ansibleee-runner:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_EVALUATOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-evaluator:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_LISTENER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-listener:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_NOTIFIER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-notifier:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_APACHE_IMAGE_URL_DEFAULT,Value:registry.redhat.io/ubi9/httpd-24:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_KEYSTONE_LISTENER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-keystone-listener:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-worker:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_CENTRAL_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_COMPUTE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_IPMI_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-ipmi:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_MYSQLD_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/prometheus/mysqld-exporter:v0.15.1,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_NOTIFICATION_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-notification:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_SGCORE_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/sg-core:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_BACKUP_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-backup:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-scheduler:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_VOLUME_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-volume:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_BACKENDBIND9_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-backend-bind9:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_CENTRAL_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-central:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_MDNS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-mdns:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_PRODUCER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-producer:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_UNBOUND_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-unbound:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-worker:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_FRR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-frr:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_ISCSID_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-iscsid:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_KEPLER_IMAGE_URL_DEFAULT,Value:quay.io/sustainable_computing_io/kepler:release-0.7.12,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_LOGROTATE_CROND_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cron:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_MULTIPATHD_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-multipathd:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_DHCP_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_METADATA_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_OVN_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-ovn-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_SRIOV_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-sriov-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NODE_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/prometheus/node-exporter:v1.5.0,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_OVN_BGP_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-bgp-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_PODMAN_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/navidys/prometheus-podman-exporter:v1.10.1,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_GLANCE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-glance-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_CFNAPI_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-api-cfn:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_ENGINE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-engine:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HORIZON_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_INFRA_MEMCACHED_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-memcached:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_INFRA_REDIS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-redis:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_CONDUCTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-conductor:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_INSPECTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-inspector:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_NEUTRON_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-neutron-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_PXE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-pxe:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_PYTHON_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/ironic-python-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_KEYSTONE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-keystone:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_KSM_IMAGE_URL_DEFAULT,Value:registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-scheduler:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_SHARE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-share:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MARIADB_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-mariadb:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NET_UTILS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-netutils:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NEUTRON_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_COMPUTE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-compute:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_CONDUCTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-conductor:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_NOVNC_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-novncproxy:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-scheduler:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_HEALTHMANAGER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-health-manager:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_HOUSEKEEPING_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-housekeeping:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_RSYSLOG_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-rsyslog:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-worker:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_CLIENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-openstackclient:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_MUST_GATHER_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-must-gather:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_NETWORK_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-network-exporter:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OS_CONTAINER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/edpm-hardened-uefi:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_CONTROLLER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_CONTROLLER_OVS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-base:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_NB_DBCLUSTER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-nb-db-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_NORTHD_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-northd:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_SB_DBCLUSTER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-sb-db-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_PLACEMENT_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-placement-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_RABBITMQ_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_ACCOUNT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-account:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_CONTAINER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-container:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_OBJECT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-object:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_PROXY_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-proxy-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_TEST_TEMPEST_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_APPLIER_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-applier:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_DECISION_ENGINE_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-decision-engine:current-podified,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cert,ReadOnly:true,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-n6f52,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-baremetal-operator-controller-manager-5dfbbd665cfq946_openstack-operators(c336d534-f819-4859-8e15-5f15b68a36ad): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 05 21:08:21 crc kubenswrapper[4754]: W1005 21:08:21.248229 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf131e747_362b_4442_8377_0c00aedfd8ae.slice/crio-f78276d47f1d51079d2e8424e7587ee593c7b3c07972090718e03a0b3f31ccf1 WatchSource:0}: Error finding container f78276d47f1d51079d2e8424e7587ee593c7b3c07972090718e03a0b3f31ccf1: Status 404 returned error can't find the container with id f78276d47f1d51079d2e8424e7587ee593c7b3c07972090718e03a0b3f31ccf1 Oct 05 21:08:21 crc kubenswrapper[4754]: E1005 21:08:21.441144 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-fx25z" podUID="b1cc32e7-f619-4a0e-b0f9-195e36c82d01" Oct 05 21:08:21 crc kubenswrapper[4754]: E1005 21:08:21.538098 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/watcher-operator-controller-manager-6cbc6dd547-zq79b" podUID="cea6487f-b088-4f9e-a42d-6a3517df7669" Oct 05 21:08:21 crc kubenswrapper[4754]: E1005 21:08:21.605171 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665cfq946" podUID="c336d534-f819-4859-8e15-5f15b68a36ad" Oct 05 21:08:21 crc kubenswrapper[4754]: I1005 21:08:21.714912 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665cfq946" event={"ID":"c336d534-f819-4859-8e15-5f15b68a36ad","Type":"ContainerStarted","Data":"2dbde177a90fd706d75743ccc6f7fde0a0402002ff4fed80aa8caa3180fe920c"} Oct 05 21:08:21 crc kubenswrapper[4754]: I1005 21:08:21.714962 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665cfq946" event={"ID":"c336d534-f819-4859-8e15-5f15b68a36ad","Type":"ContainerStarted","Data":"7122e59f1f3cf531f68e0db5f1f8495758a10c50011b4c57b2eea66214ec00aa"} Oct 05 21:08:21 crc kubenswrapper[4754]: E1005 21:08:21.717719 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:bcd1acac74e68eea5a9c3b7ba1bcb29d3a5b43423fc23c19ad4715bdac41f799\\\"\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665cfq946" podUID="c336d534-f819-4859-8e15-5f15b68a36ad" Oct 05 21:08:21 crc kubenswrapper[4754]: I1005 21:08:21.720835 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-7468f855d8-w8z85" event={"ID":"cb4a4798-9283-4633-9b3b-e2d72faa221f","Type":"ContainerStarted","Data":"b4345e9426882ae7082fca36740e85a1005335ee6c7bdd7d9743e9bae03d3113"} Oct 05 21:08:21 crc kubenswrapper[4754]: I1005 21:08:21.737804 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-hkb96" event={"ID":"97f32c3b-569a-4f15-9f7c-fb18299dec30","Type":"ContainerStarted","Data":"acc68fb20d1732f8e19ff04565a9a919ee4a3f0989794109b82683496071f159"} Oct 05 21:08:21 crc kubenswrapper[4754]: I1005 21:08:21.739117 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-579449c7d5-zcct6" event={"ID":"b18dee24-0d55-4805-8d76-623180998686","Type":"ContainerStarted","Data":"914c3603d1a4a34d607d7e9c855c1abfcc28dcbd5ef14e107e9561aaacbf5718"} Oct 05 21:08:21 crc kubenswrapper[4754]: I1005 21:08:21.743005 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-6cbc6dd547-zq79b" event={"ID":"cea6487f-b088-4f9e-a42d-6a3517df7669","Type":"ContainerStarted","Data":"93f98fdb819b6e689a85cc8486e8c7b0e16687ab54682ceb563515f724fb78ed"} Oct 05 21:08:21 crc kubenswrapper[4754]: I1005 21:08:21.743063 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-6cbc6dd547-zq79b" event={"ID":"cea6487f-b088-4f9e-a42d-6a3517df7669","Type":"ContainerStarted","Data":"af2f5d5f31a4058aa89be6c3c16f0f22f27fd2c956fc1b85309dac1def5477df"} Oct 05 21:08:21 crc kubenswrapper[4754]: E1005 21:08:21.745112 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:64f57b2b59dea2bd9fae91490c5bec2687131884a049e6579819d9f951b877c6\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-6cbc6dd547-zq79b" podUID="cea6487f-b088-4f9e-a42d-6a3517df7669" Oct 05 21:08:21 crc kubenswrapper[4754]: I1005 21:08:21.747469 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-fx25z" event={"ID":"b1cc32e7-f619-4a0e-b0f9-195e36c82d01","Type":"ContainerStarted","Data":"67b8b17d1830389aea3eacdde5d9e0f8ea7b004abfc6c9465ceb4d536349ad45"} Oct 05 21:08:21 crc kubenswrapper[4754]: I1005 21:08:21.747518 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-fx25z" event={"ID":"b1cc32e7-f619-4a0e-b0f9-195e36c82d01","Type":"ContainerStarted","Data":"bdd55e920674f95287d221f4004063e5d863106773a7e18e38594b2c539749df"} Oct 05 21:08:21 crc kubenswrapper[4754]: E1005 21:08:21.749478 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:0daf76cc40ab619ae266b11defcc1b65beb22d859369e7b1b04de9169089a4cb\\\"\"" pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-fx25z" podUID="b1cc32e7-f619-4a0e-b0f9-195e36c82d01" Oct 05 21:08:21 crc kubenswrapper[4754]: I1005 21:08:21.751119 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-658588b8c9-z7tq9" event={"ID":"4bb20421-ce34-44ee-8740-82eedd4716f3","Type":"ContainerStarted","Data":"4bc2f1a6a7cfd71f0d42b0c8529949b170c5285aa06b7320e1c00f788a633e3b"} Oct 05 21:08:21 crc kubenswrapper[4754]: I1005 21:08:21.763125 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-7c7fc454ff-4fb5b" event={"ID":"e0b701ef-8b8d-4717-aa27-07233f6b6c15","Type":"ContainerStarted","Data":"38c28e645aca6554b3f60b413c87a90a8adcfc2b59df6e04fb6fd61fd43c55b3"} Oct 05 21:08:21 crc kubenswrapper[4754]: I1005 21:08:21.775174 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-55895b89bc-zzbvh" event={"ID":"f131e747-362b-4442-8377-0c00aedfd8ae","Type":"ContainerStarted","Data":"787df531bfb1bf71be94899d0e62566e12bdefabc7612ffa27b60ed62f788e6f"} Oct 05 21:08:21 crc kubenswrapper[4754]: I1005 21:08:21.775219 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-55895b89bc-zzbvh" event={"ID":"f131e747-362b-4442-8377-0c00aedfd8ae","Type":"ContainerStarted","Data":"a1d0af893d3a2bd9a6be8f428eeccdcc651dbb0d46cbfb62a464e18c44be9429"} Oct 05 21:08:21 crc kubenswrapper[4754]: I1005 21:08:21.775229 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-55895b89bc-zzbvh" event={"ID":"f131e747-362b-4442-8377-0c00aedfd8ae","Type":"ContainerStarted","Data":"f78276d47f1d51079d2e8424e7587ee593c7b3c07972090718e03a0b3f31ccf1"} Oct 05 21:08:21 crc kubenswrapper[4754]: I1005 21:08:21.775878 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-55895b89bc-zzbvh" Oct 05 21:08:21 crc kubenswrapper[4754]: I1005 21:08:21.777429 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-5d4d74dd89-h44bd" event={"ID":"f48f3071-cd94-49dd-bfa0-1ec74f495c72","Type":"ContainerStarted","Data":"1d62609a255024be07dfa71f85786ed297f99ea94c943fa105bda6d21aec1e23"} Oct 05 21:08:21 crc kubenswrapper[4754]: I1005 21:08:21.780926 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-65d89cfd9f-msh87" event={"ID":"5757c596-103f-4a00-ac16-ed9e6c9e4719","Type":"ContainerStarted","Data":"d5bc437450a76209aee9c1cf907be0a2c03d03210d18fc0eedec889ebd3e2091"} Oct 05 21:08:21 crc kubenswrapper[4754]: I1005 21:08:21.781935 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-54689d9f88-4lrzz" event={"ID":"6316859b-8ab7-4b46-b1fa-c7f893d39a95","Type":"ContainerStarted","Data":"f060fd690273da29559a50e13b2c7012ffa46fedaeca007ef25681cdb00d1685"} Oct 05 21:08:21 crc kubenswrapper[4754]: I1005 21:08:21.783050 4754 generic.go:334] "Generic (PLEG): container finished" podID="aee8b142-dea8-4ab3-87e9-14f3d3a75543" containerID="cf9684df762a16a28d139678817b7a635cac7b5268b7c6f9a53d8cfbcf0c147a" exitCode=0 Oct 05 21:08:21 crc kubenswrapper[4754]: I1005 21:08:21.783090 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-456m8" event={"ID":"aee8b142-dea8-4ab3-87e9-14f3d3a75543","Type":"ContainerDied","Data":"cf9684df762a16a28d139678817b7a635cac7b5268b7c6f9a53d8cfbcf0c147a"} Oct 05 21:08:21 crc kubenswrapper[4754]: I1005 21:08:21.789478 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-6859f9b676-hgkv4" event={"ID":"2b729715-7fd9-4ca5-b4dd-cb0eb0034aac","Type":"ContainerStarted","Data":"966d9cbcceb3a65a0766de08c8aecddd7fb68caae980855bff451c5cc77452b5"} Oct 05 21:08:21 crc kubenswrapper[4754]: I1005 21:08:21.905878 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-55895b89bc-zzbvh" podStartSLOduration=3.9058556319999997 podStartE2EDuration="3.905855632s" podCreationTimestamp="2025-10-05 21:08:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 21:08:21.904783064 +0000 UTC m=+825.808901784" watchObservedRunningTime="2025-10-05 21:08:21.905855632 +0000 UTC m=+825.809974342" Oct 05 21:08:22 crc kubenswrapper[4754]: E1005 21:08:22.812801 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:bcd1acac74e68eea5a9c3b7ba1bcb29d3a5b43423fc23c19ad4715bdac41f799\\\"\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665cfq946" podUID="c336d534-f819-4859-8e15-5f15b68a36ad" Oct 05 21:08:22 crc kubenswrapper[4754]: E1005 21:08:22.815306 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:0daf76cc40ab619ae266b11defcc1b65beb22d859369e7b1b04de9169089a4cb\\\"\"" pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-fx25z" podUID="b1cc32e7-f619-4a0e-b0f9-195e36c82d01" Oct 05 21:08:22 crc kubenswrapper[4754]: E1005 21:08:22.815622 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:64f57b2b59dea2bd9fae91490c5bec2687131884a049e6579819d9f951b877c6\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-6cbc6dd547-zq79b" podUID="cea6487f-b088-4f9e-a42d-6a3517df7669" Oct 05 21:08:29 crc kubenswrapper[4754]: I1005 21:08:29.666415 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-55895b89bc-zzbvh" Oct 05 21:08:30 crc kubenswrapper[4754]: I1005 21:08:30.217005 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-rz4fg"] Oct 05 21:08:30 crc kubenswrapper[4754]: I1005 21:08:30.219111 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rz4fg" Oct 05 21:08:30 crc kubenswrapper[4754]: I1005 21:08:30.227778 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rz4fg"] Oct 05 21:08:30 crc kubenswrapper[4754]: I1005 21:08:30.314142 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92d8d285-7fd2-49a6-80b2-c72270bf1164-utilities\") pod \"redhat-marketplace-rz4fg\" (UID: \"92d8d285-7fd2-49a6-80b2-c72270bf1164\") " pod="openshift-marketplace/redhat-marketplace-rz4fg" Oct 05 21:08:30 crc kubenswrapper[4754]: I1005 21:08:30.314325 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92d8d285-7fd2-49a6-80b2-c72270bf1164-catalog-content\") pod \"redhat-marketplace-rz4fg\" (UID: \"92d8d285-7fd2-49a6-80b2-c72270bf1164\") " pod="openshift-marketplace/redhat-marketplace-rz4fg" Oct 05 21:08:30 crc kubenswrapper[4754]: I1005 21:08:30.314358 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7kf8d\" (UniqueName: \"kubernetes.io/projected/92d8d285-7fd2-49a6-80b2-c72270bf1164-kube-api-access-7kf8d\") pod \"redhat-marketplace-rz4fg\" (UID: \"92d8d285-7fd2-49a6-80b2-c72270bf1164\") " pod="openshift-marketplace/redhat-marketplace-rz4fg" Oct 05 21:08:30 crc kubenswrapper[4754]: I1005 21:08:30.415241 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92d8d285-7fd2-49a6-80b2-c72270bf1164-catalog-content\") pod \"redhat-marketplace-rz4fg\" (UID: \"92d8d285-7fd2-49a6-80b2-c72270bf1164\") " pod="openshift-marketplace/redhat-marketplace-rz4fg" Oct 05 21:08:30 crc kubenswrapper[4754]: I1005 21:08:30.415293 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7kf8d\" (UniqueName: \"kubernetes.io/projected/92d8d285-7fd2-49a6-80b2-c72270bf1164-kube-api-access-7kf8d\") pod \"redhat-marketplace-rz4fg\" (UID: \"92d8d285-7fd2-49a6-80b2-c72270bf1164\") " pod="openshift-marketplace/redhat-marketplace-rz4fg" Oct 05 21:08:30 crc kubenswrapper[4754]: I1005 21:08:30.415363 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92d8d285-7fd2-49a6-80b2-c72270bf1164-utilities\") pod \"redhat-marketplace-rz4fg\" (UID: \"92d8d285-7fd2-49a6-80b2-c72270bf1164\") " pod="openshift-marketplace/redhat-marketplace-rz4fg" Oct 05 21:08:30 crc kubenswrapper[4754]: I1005 21:08:30.416281 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92d8d285-7fd2-49a6-80b2-c72270bf1164-utilities\") pod \"redhat-marketplace-rz4fg\" (UID: \"92d8d285-7fd2-49a6-80b2-c72270bf1164\") " pod="openshift-marketplace/redhat-marketplace-rz4fg" Oct 05 21:08:30 crc kubenswrapper[4754]: I1005 21:08:30.422196 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92d8d285-7fd2-49a6-80b2-c72270bf1164-catalog-content\") pod \"redhat-marketplace-rz4fg\" (UID: \"92d8d285-7fd2-49a6-80b2-c72270bf1164\") " pod="openshift-marketplace/redhat-marketplace-rz4fg" Oct 05 21:08:30 crc kubenswrapper[4754]: I1005 21:08:30.434669 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7kf8d\" (UniqueName: \"kubernetes.io/projected/92d8d285-7fd2-49a6-80b2-c72270bf1164-kube-api-access-7kf8d\") pod \"redhat-marketplace-rz4fg\" (UID: \"92d8d285-7fd2-49a6-80b2-c72270bf1164\") " pod="openshift-marketplace/redhat-marketplace-rz4fg" Oct 05 21:08:30 crc kubenswrapper[4754]: I1005 21:08:30.550728 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rz4fg" Oct 05 21:08:34 crc kubenswrapper[4754]: E1005 21:08:34.121103 4754 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/ovn-operator@sha256:f923b76c1dd8fde02a5faf8a0a433cfacfb7b743f371de64a12e30d6efcde254" Oct 05 21:08:34 crc kubenswrapper[4754]: E1005 21:08:34.121784 4754 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ovn-operator@sha256:f923b76c1dd8fde02a5faf8a0a433cfacfb7b743f371de64a12e30d6efcde254,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-kk8vw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-579449c7d5-zcct6_openstack-operators(b18dee24-0d55-4805-8d76-623180998686): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 05 21:08:34 crc kubenswrapper[4754]: E1005 21:08:34.524543 4754 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2" Oct 05 21:08:34 crc kubenswrapper[4754]: E1005 21:08:34.524814 4754 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-wbkkf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-5f97d8c699-hkb96_openstack-operators(97f32c3b-569a-4f15-9f7c-fb18299dec30): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 05 21:08:34 crc kubenswrapper[4754]: E1005 21:08:34.526038 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-hkb96" podUID="97f32c3b-569a-4f15-9f7c-fb18299dec30" Oct 05 21:08:34 crc kubenswrapper[4754]: E1005 21:08:34.913065 4754 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/octavia-operator@sha256:da5c3078d80878d66c616e6f8a0bb909f95d971cde2c612f96fded064113e182" Oct 05 21:08:34 crc kubenswrapper[4754]: E1005 21:08:34.913331 4754 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/octavia-operator@sha256:da5c3078d80878d66c616e6f8a0bb909f95d971cde2c612f96fded064113e182,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-w6n8v,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-7468f855d8-w8z85_openstack-operators(cb4a4798-9283-4633-9b3b-e2d72faa221f): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 05 21:08:34 crc kubenswrapper[4754]: E1005 21:08:34.930979 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-hkb96" podUID="97f32c3b-569a-4f15-9f7c-fb18299dec30" Oct 05 21:08:35 crc kubenswrapper[4754]: I1005 21:08:35.942987 4754 generic.go:334] "Generic (PLEG): container finished" podID="aee8b142-dea8-4ab3-87e9-14f3d3a75543" containerID="2ca19477259a6da6812707cc7775fbd87030fa5c33d188c619770772bf12e063" exitCode=0 Oct 05 21:08:35 crc kubenswrapper[4754]: I1005 21:08:35.943074 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-456m8" event={"ID":"aee8b142-dea8-4ab3-87e9-14f3d3a75543","Type":"ContainerDied","Data":"2ca19477259a6da6812707cc7775fbd87030fa5c33d188c619770772bf12e063"} Oct 05 21:08:38 crc kubenswrapper[4754]: E1005 21:08:38.035902 4754 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/ironic-operator@sha256:785670b14b19ffd7e0799dcf3e3e275329fa822d4a604eace09574f8bb1f8162" Oct 05 21:08:38 crc kubenswrapper[4754]: E1005 21:08:38.036294 4754 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ironic-operator@sha256:785670b14b19ffd7e0799dcf3e3e275329fa822d4a604eace09574f8bb1f8162,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-9xbvr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ironic-operator-controller-manager-649675d675-v7285_openstack-operators(3ddbeeee-8a3e-4f3c-97e0-c7d8d5929345): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 05 21:08:38 crc kubenswrapper[4754]: I1005 21:08:38.546682 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-fl6cv"] Oct 05 21:08:38 crc kubenswrapper[4754]: I1005 21:08:38.548546 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fl6cv" Oct 05 21:08:38 crc kubenswrapper[4754]: I1005 21:08:38.554436 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-fl6cv"] Oct 05 21:08:38 crc kubenswrapper[4754]: I1005 21:08:38.666531 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2xxjr\" (UniqueName: \"kubernetes.io/projected/8a49c2cc-d61f-404e-946b-bb74451f895d-kube-api-access-2xxjr\") pod \"certified-operators-fl6cv\" (UID: \"8a49c2cc-d61f-404e-946b-bb74451f895d\") " pod="openshift-marketplace/certified-operators-fl6cv" Oct 05 21:08:38 crc kubenswrapper[4754]: I1005 21:08:38.666650 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8a49c2cc-d61f-404e-946b-bb74451f895d-catalog-content\") pod \"certified-operators-fl6cv\" (UID: \"8a49c2cc-d61f-404e-946b-bb74451f895d\") " pod="openshift-marketplace/certified-operators-fl6cv" Oct 05 21:08:38 crc kubenswrapper[4754]: I1005 21:08:38.666705 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8a49c2cc-d61f-404e-946b-bb74451f895d-utilities\") pod \"certified-operators-fl6cv\" (UID: \"8a49c2cc-d61f-404e-946b-bb74451f895d\") " pod="openshift-marketplace/certified-operators-fl6cv" Oct 05 21:08:38 crc kubenswrapper[4754]: I1005 21:08:38.768528 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2xxjr\" (UniqueName: \"kubernetes.io/projected/8a49c2cc-d61f-404e-946b-bb74451f895d-kube-api-access-2xxjr\") pod \"certified-operators-fl6cv\" (UID: \"8a49c2cc-d61f-404e-946b-bb74451f895d\") " pod="openshift-marketplace/certified-operators-fl6cv" Oct 05 21:08:38 crc kubenswrapper[4754]: I1005 21:08:38.768618 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8a49c2cc-d61f-404e-946b-bb74451f895d-catalog-content\") pod \"certified-operators-fl6cv\" (UID: \"8a49c2cc-d61f-404e-946b-bb74451f895d\") " pod="openshift-marketplace/certified-operators-fl6cv" Oct 05 21:08:38 crc kubenswrapper[4754]: I1005 21:08:38.768885 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8a49c2cc-d61f-404e-946b-bb74451f895d-utilities\") pod \"certified-operators-fl6cv\" (UID: \"8a49c2cc-d61f-404e-946b-bb74451f895d\") " pod="openshift-marketplace/certified-operators-fl6cv" Oct 05 21:08:38 crc kubenswrapper[4754]: I1005 21:08:38.770027 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8a49c2cc-d61f-404e-946b-bb74451f895d-utilities\") pod \"certified-operators-fl6cv\" (UID: \"8a49c2cc-d61f-404e-946b-bb74451f895d\") " pod="openshift-marketplace/certified-operators-fl6cv" Oct 05 21:08:38 crc kubenswrapper[4754]: I1005 21:08:38.770030 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8a49c2cc-d61f-404e-946b-bb74451f895d-catalog-content\") pod \"certified-operators-fl6cv\" (UID: \"8a49c2cc-d61f-404e-946b-bb74451f895d\") " pod="openshift-marketplace/certified-operators-fl6cv" Oct 05 21:08:38 crc kubenswrapper[4754]: I1005 21:08:38.792412 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2xxjr\" (UniqueName: \"kubernetes.io/projected/8a49c2cc-d61f-404e-946b-bb74451f895d-kube-api-access-2xxjr\") pod \"certified-operators-fl6cv\" (UID: \"8a49c2cc-d61f-404e-946b-bb74451f895d\") " pod="openshift-marketplace/certified-operators-fl6cv" Oct 05 21:08:38 crc kubenswrapper[4754]: I1005 21:08:38.885055 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fl6cv" Oct 05 21:08:39 crc kubenswrapper[4754]: E1005 21:08:39.204162 4754 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/swift-operator@sha256:637bb7b9ac308bc1e323391a3593b824f688090a856c83385814c17a571b1eed" Oct 05 21:08:39 crc kubenswrapper[4754]: E1005 21:08:39.204412 4754 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:637bb7b9ac308bc1e323391a3593b824f688090a856c83385814c17a571b1eed,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-g95tf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-6859f9b676-hgkv4_openstack-operators(2b729715-7fd9-4ca5-b4dd-cb0eb0034aac): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 05 21:08:40 crc kubenswrapper[4754]: E1005 21:08:40.266387 4754 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/infra-operator@sha256:b6cef68bfaacdf992a9fa1a6b03a848a48c18cbb6ed12d95561b4b37d858b99f" Oct 05 21:08:40 crc kubenswrapper[4754]: E1005 21:08:40.268627 4754 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/infra-operator@sha256:b6cef68bfaacdf992a9fa1a6b03a848a48c18cbb6ed12d95561b4b37d858b99f,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{600 -3} {} 600m DecimalSI},memory: {{2147483648 0} {} 2Gi BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{536870912 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cert,ReadOnly:true,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-24bn6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod infra-operator-controller-manager-658588b8c9-z7tq9_openstack-operators(4bb20421-ce34-44ee-8740-82eedd4716f3): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 05 21:08:40 crc kubenswrapper[4754]: E1005 21:08:40.879740 4754 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/nova-operator@sha256:e4c4ff39c54c0af231fb781759ab50ed86285c74d38bdea43fa75646b762d842" Oct 05 21:08:40 crc kubenswrapper[4754]: E1005 21:08:40.880036 4754 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:e4c4ff39c54c0af231fb781759ab50ed86285c74d38bdea43fa75646b762d842,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-fvkwj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-7c7fc454ff-4fb5b_openstack-operators(e0b701ef-8b8d-4717-aa27-07233f6b6c15): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 05 21:08:41 crc kubenswrapper[4754]: E1005 21:08:41.368964 4754 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/manila-operator@sha256:063aae1458289d1090a77c74c2b978b9eb978b0e4062c399f0cb5434a8dd2757" Oct 05 21:08:41 crc kubenswrapper[4754]: E1005 21:08:41.369522 4754 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/manila-operator@sha256:063aae1458289d1090a77c74c2b978b9eb978b0e4062c399f0cb5434a8dd2757,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-zbcs6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod manila-operator-controller-manager-65d89cfd9f-msh87_openstack-operators(5757c596-103f-4a00-ac16-ed9e6c9e4719): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 05 21:08:41 crc kubenswrapper[4754]: E1005 21:08:41.783122 4754 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/placement-operator@sha256:adc23c5fd1aece2b16dc8e22ceed628f9a719455e39d3f98c77544665c6749e1" Oct 05 21:08:41 crc kubenswrapper[4754]: E1005 21:08:41.783290 4754 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:adc23c5fd1aece2b16dc8e22ceed628f9a719455e39d3f98c77544665c6749e1,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-mrgrx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-54689d9f88-4lrzz_openstack-operators(6316859b-8ab7-4b46-b1fa-c7f893d39a95): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 05 21:08:42 crc kubenswrapper[4754]: E1005 21:08:42.267636 4754 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/mariadb-operator@sha256:5f96b563a63494082323bfced089d6589e0c89db43c6a39a2e912c79b1a278fe" Oct 05 21:08:42 crc kubenswrapper[4754]: E1005 21:08:42.268157 4754 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/mariadb-operator@sha256:5f96b563a63494082323bfced089d6589e0c89db43c6a39a2e912c79b1a278fe,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-kcq9f,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod mariadb-operator-controller-manager-6cd6d7bdf5-f5jxx_openstack-operators(5a1312fe-c6d6-41de-a442-b044db904a6d): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 05 21:08:46 crc kubenswrapper[4754]: E1005 21:08:46.368091 4754 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/barbican-operator@sha256:f9fc3cf4084a325d7f5f9773bfcc2b839ccff1c72e61fdd8f410a7ef46497f75" Oct 05 21:08:46 crc kubenswrapper[4754]: E1005 21:08:46.369045 4754 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/barbican-operator@sha256:f9fc3cf4084a325d7f5f9773bfcc2b839ccff1c72e61fdd8f410a7ef46497f75,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-qr6mp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-operator-controller-manager-5f7c849b98-n9ksx_openstack-operators(9a1cc24e-8371-44cb-bd37-5b765853fdff): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 05 21:08:46 crc kubenswrapper[4754]: E1005 21:08:46.835378 4754 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/neutron-operator@sha256:dfd044635f9df9ed1d249387fa622177db35cdc72475e1c570617b8d17c64862" Oct 05 21:08:46 crc kubenswrapper[4754]: E1005 21:08:46.836711 4754 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/neutron-operator@sha256:dfd044635f9df9ed1d249387fa622177db35cdc72475e1c570617b8d17c64862,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-vmrg7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-8d984cc4d-7xzhb_openstack-operators(ff3ba6ea-c2d8-4a3e-b0a0-58fd6df5b196): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 05 21:08:47 crc kubenswrapper[4754]: E1005 21:08:47.381805 4754 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/telemetry-operator@sha256:bf55026ba10b80e1e24733078bd204cef8766d21a305fd000707a1e3b30ff52e" Oct 05 21:08:47 crc kubenswrapper[4754]: E1005 21:08:47.382054 4754 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/telemetry-operator@sha256:bf55026ba10b80e1e24733078bd204cef8766d21a305fd000707a1e3b30ff52e,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-42qgx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-5d4d74dd89-h44bd_openstack-operators(f48f3071-cd94-49dd-bfa0-1ec74f495c72): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 05 21:08:48 crc kubenswrapper[4754]: I1005 21:08:48.945415 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rz4fg"] Oct 05 21:08:49 crc kubenswrapper[4754]: E1005 21:08:49.215066 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/octavia-operator-controller-manager-7468f855d8-w8z85" podUID="cb4a4798-9283-4633-9b3b-e2d72faa221f" Oct 05 21:08:49 crc kubenswrapper[4754]: I1005 21:08:49.215608 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-fl6cv"] Oct 05 21:08:49 crc kubenswrapper[4754]: W1005 21:08:49.340346 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8a49c2cc_d61f_404e_946b_bb74451f895d.slice/crio-72b713d766a0be917aa3e435f4ea8ae4b520050643fd4ac6fca3c1289a10bd08 WatchSource:0}: Error finding container 72b713d766a0be917aa3e435f4ea8ae4b520050643fd4ac6fca3c1289a10bd08: Status 404 returned error can't find the container with id 72b713d766a0be917aa3e435f4ea8ae4b520050643fd4ac6fca3c1289a10bd08 Oct 05 21:08:49 crc kubenswrapper[4754]: E1005 21:08:49.766461 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/ovn-operator-controller-manager-579449c7d5-zcct6" podUID="b18dee24-0d55-4805-8d76-623180998686" Oct 05 21:08:49 crc kubenswrapper[4754]: E1005 21:08:49.787088 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/manila-operator-controller-manager-65d89cfd9f-msh87" podUID="5757c596-103f-4a00-ac16-ed9e6c9e4719" Oct 05 21:08:49 crc kubenswrapper[4754]: E1005 21:08:49.848031 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/telemetry-operator-controller-manager-5d4d74dd89-h44bd" podUID="f48f3071-cd94-49dd-bfa0-1ec74f495c72" Oct 05 21:08:50 crc kubenswrapper[4754]: I1005 21:08:50.068319 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fl6cv" event={"ID":"8a49c2cc-d61f-404e-946b-bb74451f895d","Type":"ContainerStarted","Data":"72b713d766a0be917aa3e435f4ea8ae4b520050643fd4ac6fca3c1289a10bd08"} Oct 05 21:08:50 crc kubenswrapper[4754]: I1005 21:08:50.069459 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-5d4d74dd89-h44bd" event={"ID":"f48f3071-cd94-49dd-bfa0-1ec74f495c72","Type":"ContainerStarted","Data":"9570b6dfd04476c9fa8dbe41bf6345b2004ac82d3ff0d7a2a995d35a44944657"} Oct 05 21:08:50 crc kubenswrapper[4754]: I1005 21:08:50.073927 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-65d89cfd9f-msh87" event={"ID":"5757c596-103f-4a00-ac16-ed9e6c9e4719","Type":"ContainerStarted","Data":"e87b517dc243122d414113112e1ddeab81544d49be58c1be9143036044540f4d"} Oct 05 21:08:50 crc kubenswrapper[4754]: I1005 21:08:50.079506 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665cfq946" event={"ID":"c336d534-f819-4859-8e15-5f15b68a36ad","Type":"ContainerStarted","Data":"d1360348812e002c0d1d5077c7b29176615e9e95da93107fcd36f6ece88c5059"} Oct 05 21:08:50 crc kubenswrapper[4754]: I1005 21:08:50.080338 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665cfq946" Oct 05 21:08:50 crc kubenswrapper[4754]: I1005 21:08:50.112969 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-655d88ccb9-l56qv" event={"ID":"2735f261-cad9-4dff-9136-931a45c85ac5","Type":"ContainerStarted","Data":"4b1a3513549c5875d294672a5a9f00c781ef2c512e7114d9b690653669a28d9a"} Oct 05 21:08:50 crc kubenswrapper[4754]: I1005 21:08:50.123872 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-579449c7d5-zcct6" event={"ID":"b18dee24-0d55-4805-8d76-623180998686","Type":"ContainerStarted","Data":"0e5e76c069a6a4de1781cc116c88c16bb78dcf18af4c9e58f77f99e4fae02b3b"} Oct 05 21:08:50 crc kubenswrapper[4754]: I1005 21:08:50.124840 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665cfq946" podStartSLOduration=5.5621726129999995 podStartE2EDuration="33.12482359s" podCreationTimestamp="2025-10-05 21:08:17 +0000 UTC" firstStartedPulling="2025-10-05 21:08:21.223269212 +0000 UTC m=+825.127387922" lastFinishedPulling="2025-10-05 21:08:48.785920189 +0000 UTC m=+852.690038899" observedRunningTime="2025-10-05 21:08:50.121281458 +0000 UTC m=+854.025400168" watchObservedRunningTime="2025-10-05 21:08:50.12482359 +0000 UTC m=+854.028942300" Oct 05 21:08:50 crc kubenswrapper[4754]: I1005 21:08:50.125682 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rz4fg" event={"ID":"92d8d285-7fd2-49a6-80b2-c72270bf1164","Type":"ContainerStarted","Data":"27dabd6a42fcc4669fda2e7bcbf264e49859a363597fe148ed4b527c78756a78"} Oct 05 21:08:50 crc kubenswrapper[4754]: I1005 21:08:50.127476 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-54876c876f-svzwt" event={"ID":"f560133d-b3b5-4791-b69a-a700367f0d96","Type":"ContainerStarted","Data":"e845bde9eedb89376041cb8809b814105add1c73638510acc5d4e930a06583ae"} Oct 05 21:08:50 crc kubenswrapper[4754]: I1005 21:08:50.132594 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-456m8" event={"ID":"aee8b142-dea8-4ab3-87e9-14f3d3a75543","Type":"ContainerStarted","Data":"fb35c6fd17cd23d67f4fbdd6d4e4977d77bbb398d847e6b5ad80461080ecc8a7"} Oct 05 21:08:50 crc kubenswrapper[4754]: I1005 21:08:50.134777 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-8f58bc9db-xsrhv" event={"ID":"82184642-9d7b-4ada-b7b2-efde6c91cb14","Type":"ContainerStarted","Data":"c87b0f6d5911407f8cf1b5ef5b1c939cea5f41e4114188be5b83059e3bb1a162"} Oct 05 21:08:50 crc kubenswrapper[4754]: I1005 21:08:50.137218 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-7468f855d8-w8z85" event={"ID":"cb4a4798-9283-4633-9b3b-e2d72faa221f","Type":"ContainerStarted","Data":"90ff7dfbccbf55264558847982dc0e26a0b3b62961d5f97436ffe90a297d5fc9"} Oct 05 21:08:50 crc kubenswrapper[4754]: I1005 21:08:50.139414 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-7d4d4f8d-kcxvd" event={"ID":"47cafc06-0f75-48ee-bae5-4e30352f4572","Type":"ContainerStarted","Data":"8ce5b9e4ce921bd9b53f38d182fecaf2a4e168aeb24f6f3ad012eaeed6015fb4"} Oct 05 21:08:50 crc kubenswrapper[4754]: I1005 21:08:50.153746 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-5568b5d68-f5rjh" event={"ID":"62159d2d-ec19-4b44-89df-846faa5757dc","Type":"ContainerStarted","Data":"11d8b217023ece1e7484cf6671591bccd070d84704b4605d6834e99ac3ad3a95"} Oct 05 21:08:50 crc kubenswrapper[4754]: I1005 21:08:50.160693 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-fxx5s" event={"ID":"3ec442d5-220a-4d2f-8f62-a13ff86a3229","Type":"ContainerStarted","Data":"bc5b05bda8f27800e1c37ddcf6d5ad0a595f63cf14a699f2c05472fd5d06ccea"} Oct 05 21:08:50 crc kubenswrapper[4754]: I1005 21:08:50.175483 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-456m8" podStartSLOduration=6.174479323 podStartE2EDuration="33.175464315s" podCreationTimestamp="2025-10-05 21:08:17 +0000 UTC" firstStartedPulling="2025-10-05 21:08:21.785972955 +0000 UTC m=+825.690091665" lastFinishedPulling="2025-10-05 21:08:48.786957947 +0000 UTC m=+852.691076657" observedRunningTime="2025-10-05 21:08:50.174467559 +0000 UTC m=+854.078586269" watchObservedRunningTime="2025-10-05 21:08:50.175464315 +0000 UTC m=+854.079583025" Oct 05 21:08:51 crc kubenswrapper[4754]: I1005 21:08:51.168418 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-6cd6d7bdf5-f5jxx" event={"ID":"5a1312fe-c6d6-41de-a442-b044db904a6d","Type":"ContainerStarted","Data":"19ecd25d01d0da353ad9577be02b1f23b7dba469392f5c95f8d51338aac0dee4"} Oct 05 21:08:51 crc kubenswrapper[4754]: E1005 21:08:51.317365 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/manila-operator@sha256:063aae1458289d1090a77c74c2b978b9eb978b0e4062c399f0cb5434a8dd2757\\\"\"" pod="openstack-operators/manila-operator-controller-manager-65d89cfd9f-msh87" podUID="5757c596-103f-4a00-ac16-ed9e6c9e4719" Oct 05 21:08:51 crc kubenswrapper[4754]: E1005 21:08:51.317456 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:bf55026ba10b80e1e24733078bd204cef8766d21a305fd000707a1e3b30ff52e\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-5d4d74dd89-h44bd" podUID="f48f3071-cd94-49dd-bfa0-1ec74f495c72" Oct 05 21:08:51 crc kubenswrapper[4754]: E1005 21:08:51.897293 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/mariadb-operator-controller-manager-6cd6d7bdf5-f5jxx" podUID="5a1312fe-c6d6-41de-a442-b044db904a6d" Oct 05 21:08:51 crc kubenswrapper[4754]: E1005 21:08:51.999355 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/neutron-operator-controller-manager-8d984cc4d-7xzhb" podUID="ff3ba6ea-c2d8-4a3e-b0a0-58fd6df5b196" Oct 05 21:08:51 crc kubenswrapper[4754]: E1005 21:08:51.999377 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/nova-operator-controller-manager-7c7fc454ff-4fb5b" podUID="e0b701ef-8b8d-4717-aa27-07233f6b6c15" Oct 05 21:08:51 crc kubenswrapper[4754]: E1005 21:08:51.999558 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/ironic-operator-controller-manager-649675d675-v7285" podUID="3ddbeeee-8a3e-4f3c-97e0-c7d8d5929345" Oct 05 21:08:52 crc kubenswrapper[4754]: E1005 21:08:52.121446 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/placement-operator-controller-manager-54689d9f88-4lrzz" podUID="6316859b-8ab7-4b46-b1fa-c7f893d39a95" Oct 05 21:08:52 crc kubenswrapper[4754]: E1005 21:08:52.148973 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/infra-operator-controller-manager-658588b8c9-z7tq9" podUID="4bb20421-ce34-44ee-8740-82eedd4716f3" Oct 05 21:08:52 crc kubenswrapper[4754]: I1005 21:08:52.190437 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-7c7fc454ff-4fb5b" event={"ID":"e0b701ef-8b8d-4717-aa27-07233f6b6c15","Type":"ContainerStarted","Data":"aeaf2391b132107fcde5d50a23cc5a785e613f0a1df325d3a7991970c2994f62"} Oct 05 21:08:52 crc kubenswrapper[4754]: I1005 21:08:52.199134 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-fxx5s" event={"ID":"3ec442d5-220a-4d2f-8f62-a13ff86a3229","Type":"ContainerStarted","Data":"e927fd50117b9ba885ef12619b6efcd1fe3c67638a3dcc381b1e0770e928324e"} Oct 05 21:08:52 crc kubenswrapper[4754]: I1005 21:08:52.201965 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-5f7c849b98-n9ksx" event={"ID":"9a1cc24e-8371-44cb-bd37-5b765853fdff","Type":"ContainerStarted","Data":"a16bd2030973b13bdc36b0df11dc908597c6870bfd93d45d298709ececd20baa"} Oct 05 21:08:52 crc kubenswrapper[4754]: I1005 21:08:52.205740 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-658588b8c9-z7tq9" event={"ID":"4bb20421-ce34-44ee-8740-82eedd4716f3","Type":"ContainerStarted","Data":"ceb8d47d542504986b49ad533c780aa1d8123b1749c13ee083fffd188201a0ff"} Oct 05 21:08:52 crc kubenswrapper[4754]: I1005 21:08:52.216724 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-fx25z" event={"ID":"b1cc32e7-f619-4a0e-b0f9-195e36c82d01","Type":"ContainerStarted","Data":"90cc62857967b123264f1cbc9bb704108f5a95a7d76a27e8d563565c7a4caa85"} Oct 05 21:08:52 crc kubenswrapper[4754]: I1005 21:08:52.218542 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-hkb96" event={"ID":"97f32c3b-569a-4f15-9f7c-fb18299dec30","Type":"ContainerStarted","Data":"1757a2a44043a8d17b35df6d848b6342220a9d142b676b6a4fde5727b6e24042"} Oct 05 21:08:52 crc kubenswrapper[4754]: I1005 21:08:52.221812 4754 generic.go:334] "Generic (PLEG): container finished" podID="92d8d285-7fd2-49a6-80b2-c72270bf1164" containerID="019301fbc8086a02212eeecc11c6bda9d83447d02cc0057f30170f3b0ef3f430" exitCode=0 Oct 05 21:08:52 crc kubenswrapper[4754]: I1005 21:08:52.221907 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rz4fg" event={"ID":"92d8d285-7fd2-49a6-80b2-c72270bf1164","Type":"ContainerDied","Data":"019301fbc8086a02212eeecc11c6bda9d83447d02cc0057f30170f3b0ef3f430"} Oct 05 21:08:52 crc kubenswrapper[4754]: I1005 21:08:52.235050 4754 generic.go:334] "Generic (PLEG): container finished" podID="8a49c2cc-d61f-404e-946b-bb74451f895d" containerID="1d265c1fd4de76309817aba1e212a9f6666f0ec33471058ab8942dc4eba4f84a" exitCode=0 Oct 05 21:08:52 crc kubenswrapper[4754]: I1005 21:08:52.235393 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fl6cv" event={"ID":"8a49c2cc-d61f-404e-946b-bb74451f895d","Type":"ContainerDied","Data":"1d265c1fd4de76309817aba1e212a9f6666f0ec33471058ab8942dc4eba4f84a"} Oct 05 21:08:52 crc kubenswrapper[4754]: I1005 21:08:52.244781 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-6859f9b676-hgkv4" event={"ID":"2b729715-7fd9-4ca5-b4dd-cb0eb0034aac","Type":"ContainerStarted","Data":"d5183aa2176f5daaa134b4dbc5ed61aea09fe577208acadeb623803d96d2d96e"} Oct 05 21:08:52 crc kubenswrapper[4754]: I1005 21:08:52.260256 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-7d4d4f8d-kcxvd" event={"ID":"47cafc06-0f75-48ee-bae5-4e30352f4572","Type":"ContainerStarted","Data":"0186c82032389ed76a90457dde86563f02966692d450b45e47d8ff2c39aee783"} Oct 05 21:08:52 crc kubenswrapper[4754]: I1005 21:08:52.261051 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-7d4d4f8d-kcxvd" Oct 05 21:08:52 crc kubenswrapper[4754]: I1005 21:08:52.277330 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-6cbc6dd547-zq79b" event={"ID":"cea6487f-b088-4f9e-a42d-6a3517df7669","Type":"ContainerStarted","Data":"48b95fd262a7937ea048baf2d7a8a4e54a46a1ecb58d10170d55177cc5e78b7f"} Oct 05 21:08:52 crc kubenswrapper[4754]: I1005 21:08:52.282973 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-54689d9f88-4lrzz" event={"ID":"6316859b-8ab7-4b46-b1fa-c7f893d39a95","Type":"ContainerStarted","Data":"bbd8e92dd955973f23b09d2947af34a74c40921e47e7fb37dc86f3bf8f3d7b50"} Oct 05 21:08:52 crc kubenswrapper[4754]: I1005 21:08:52.290458 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-7d4d4f8d-kcxvd" podStartSLOduration=6.787872844 podStartE2EDuration="35.290430542s" podCreationTimestamp="2025-10-05 21:08:17 +0000 UTC" firstStartedPulling="2025-10-05 21:08:20.11209089 +0000 UTC m=+824.016209590" lastFinishedPulling="2025-10-05 21:08:48.614648568 +0000 UTC m=+852.518767288" observedRunningTime="2025-10-05 21:08:52.288321047 +0000 UTC m=+856.192439787" watchObservedRunningTime="2025-10-05 21:08:52.290430542 +0000 UTC m=+856.194549252" Oct 05 21:08:52 crc kubenswrapper[4754]: I1005 21:08:52.302783 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-5568b5d68-f5rjh" event={"ID":"62159d2d-ec19-4b44-89df-846faa5757dc","Type":"ContainerStarted","Data":"eee52832d65e25cd1790d8455afa97fe53da1acc270c2da09ff9b255b3ce77ae"} Oct 05 21:08:52 crc kubenswrapper[4754]: E1005 21:08:52.309277 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/swift-operator-controller-manager-6859f9b676-hgkv4" podUID="2b729715-7fd9-4ca5-b4dd-cb0eb0034aac" Oct 05 21:08:52 crc kubenswrapper[4754]: I1005 21:08:52.316171 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-8f58bc9db-xsrhv" event={"ID":"82184642-9d7b-4ada-b7b2-efde6c91cb14","Type":"ContainerStarted","Data":"c1cd56e177ff1566e97af29268166bea7157c33594fb32d36ce329ea69581de9"} Oct 05 21:08:52 crc kubenswrapper[4754]: I1005 21:08:52.316903 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-8f58bc9db-xsrhv" Oct 05 21:08:52 crc kubenswrapper[4754]: I1005 21:08:52.339393 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-54876c876f-svzwt" event={"ID":"f560133d-b3b5-4791-b69a-a700367f0d96","Type":"ContainerStarted","Data":"1e730ca4f67cc1cee4fc452b9a239abb5674793401916bd060b488898a29c6e6"} Oct 05 21:08:52 crc kubenswrapper[4754]: I1005 21:08:52.339465 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-54876c876f-svzwt" Oct 05 21:08:52 crc kubenswrapper[4754]: I1005 21:08:52.364509 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-8f58bc9db-xsrhv" podStartSLOduration=6.885607831 podStartE2EDuration="35.36446867s" podCreationTimestamp="2025-10-05 21:08:17 +0000 UTC" firstStartedPulling="2025-10-05 21:08:20.134930727 +0000 UTC m=+824.039049437" lastFinishedPulling="2025-10-05 21:08:48.613791566 +0000 UTC m=+852.517910276" observedRunningTime="2025-10-05 21:08:52.357556849 +0000 UTC m=+856.261675559" watchObservedRunningTime="2025-10-05 21:08:52.36446867 +0000 UTC m=+856.268587370" Oct 05 21:08:52 crc kubenswrapper[4754]: I1005 21:08:52.365577 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-655d88ccb9-l56qv" event={"ID":"2735f261-cad9-4dff-9136-931a45c85ac5","Type":"ContainerStarted","Data":"ec6daf8ffe0444cfb96cf4e41db329b83a67d1dcdb5d263d2c8ca3909ebf7129"} Oct 05 21:08:52 crc kubenswrapper[4754]: I1005 21:08:52.366268 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-655d88ccb9-l56qv" Oct 05 21:08:52 crc kubenswrapper[4754]: I1005 21:08:52.372538 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-8d984cc4d-7xzhb" event={"ID":"ff3ba6ea-c2d8-4a3e-b0a0-58fd6df5b196","Type":"ContainerStarted","Data":"a2ffbd00505fb3b15ec2d2c091f6319acea0fae818c36184b918bc8430f15e2c"} Oct 05 21:08:52 crc kubenswrapper[4754]: E1005 21:08:52.373701 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/neutron-operator@sha256:dfd044635f9df9ed1d249387fa622177db35cdc72475e1c570617b8d17c64862\\\"\"" pod="openstack-operators/neutron-operator-controller-manager-8d984cc4d-7xzhb" podUID="ff3ba6ea-c2d8-4a3e-b0a0-58fd6df5b196" Oct 05 21:08:52 crc kubenswrapper[4754]: I1005 21:08:52.375206 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-649675d675-v7285" event={"ID":"3ddbeeee-8a3e-4f3c-97e0-c7d8d5929345","Type":"ContainerStarted","Data":"dbb334fd35297fa442f6b79801371ff2bd3ed59d527e5036ef49bbbe83421221"} Oct 05 21:08:52 crc kubenswrapper[4754]: E1005 21:08:52.376737 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:bf55026ba10b80e1e24733078bd204cef8766d21a305fd000707a1e3b30ff52e\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-5d4d74dd89-h44bd" podUID="f48f3071-cd94-49dd-bfa0-1ec74f495c72" Oct 05 21:08:52 crc kubenswrapper[4754]: I1005 21:08:52.407016 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-54876c876f-svzwt" podStartSLOduration=6.286713501 podStartE2EDuration="35.406994692s" podCreationTimestamp="2025-10-05 21:08:17 +0000 UTC" firstStartedPulling="2025-10-05 21:08:19.494181732 +0000 UTC m=+823.398300442" lastFinishedPulling="2025-10-05 21:08:48.614462913 +0000 UTC m=+852.518581633" observedRunningTime="2025-10-05 21:08:52.392097882 +0000 UTC m=+856.296216592" watchObservedRunningTime="2025-10-05 21:08:52.406994692 +0000 UTC m=+856.311113402" Oct 05 21:08:52 crc kubenswrapper[4754]: E1005 21:08:52.431537 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/barbican-operator-controller-manager-5f7c849b98-n9ksx" podUID="9a1cc24e-8371-44cb-bd37-5b765853fdff" Oct 05 21:08:52 crc kubenswrapper[4754]: I1005 21:08:52.552017 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-655d88ccb9-l56qv" podStartSLOduration=7.063420634 podStartE2EDuration="35.551993286s" podCreationTimestamp="2025-10-05 21:08:17 +0000 UTC" firstStartedPulling="2025-10-05 21:08:20.124600177 +0000 UTC m=+824.028718887" lastFinishedPulling="2025-10-05 21:08:48.613172819 +0000 UTC m=+852.517291539" observedRunningTime="2025-10-05 21:08:52.549403858 +0000 UTC m=+856.453522568" watchObservedRunningTime="2025-10-05 21:08:52.551993286 +0000 UTC m=+856.456111996" Oct 05 21:08:53 crc kubenswrapper[4754]: I1005 21:08:53.449029 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-5568b5d68-f5rjh" podStartSLOduration=7.5878892350000005 podStartE2EDuration="36.449004265s" podCreationTimestamp="2025-10-05 21:08:17 +0000 UTC" firstStartedPulling="2025-10-05 21:08:19.753767354 +0000 UTC m=+823.657886064" lastFinishedPulling="2025-10-05 21:08:48.614882384 +0000 UTC m=+852.519001094" observedRunningTime="2025-10-05 21:08:53.415435167 +0000 UTC m=+857.319553897" watchObservedRunningTime="2025-10-05 21:08:53.449004265 +0000 UTC m=+857.353122985" Oct 05 21:08:53 crc kubenswrapper[4754]: I1005 21:08:53.494578 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-fxx5s" podStartSLOduration=7.115651349 podStartE2EDuration="36.494547707s" podCreationTimestamp="2025-10-05 21:08:17 +0000 UTC" firstStartedPulling="2025-10-05 21:08:19.234741734 +0000 UTC m=+823.138860434" lastFinishedPulling="2025-10-05 21:08:48.613638072 +0000 UTC m=+852.517756792" observedRunningTime="2025-10-05 21:08:53.49008904 +0000 UTC m=+857.394207760" watchObservedRunningTime="2025-10-05 21:08:53.494547707 +0000 UTC m=+857.398666417" Oct 05 21:08:53 crc kubenswrapper[4754]: I1005 21:08:53.522032 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-hkb96" podStartSLOduration=7.918407826 podStartE2EDuration="35.521978564s" podCreationTimestamp="2025-10-05 21:08:18 +0000 UTC" firstStartedPulling="2025-10-05 21:08:21.217673516 +0000 UTC m=+825.121792226" lastFinishedPulling="2025-10-05 21:08:48.821244254 +0000 UTC m=+852.725362964" observedRunningTime="2025-10-05 21:08:53.51300151 +0000 UTC m=+857.417120240" watchObservedRunningTime="2025-10-05 21:08:53.521978564 +0000 UTC m=+857.426097314" Oct 05 21:08:53 crc kubenswrapper[4754]: E1005 21:08:53.584737 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/barbican-operator@sha256:f9fc3cf4084a325d7f5f9773bfcc2b839ccff1c72e61fdd8f410a7ef46497f75\\\"\"" pod="openstack-operators/barbican-operator-controller-manager-5f7c849b98-n9ksx" podUID="9a1cc24e-8371-44cb-bd37-5b765853fdff" Oct 05 21:08:53 crc kubenswrapper[4754]: I1005 21:08:53.604570 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-fx25z" podStartSLOduration=7.903271911 podStartE2EDuration="35.604548375s" podCreationTimestamp="2025-10-05 21:08:18 +0000 UTC" firstStartedPulling="2025-10-05 21:08:21.106312392 +0000 UTC m=+825.010431102" lastFinishedPulling="2025-10-05 21:08:48.807588856 +0000 UTC m=+852.711707566" observedRunningTime="2025-10-05 21:08:53.601168966 +0000 UTC m=+857.505287676" watchObservedRunningTime="2025-10-05 21:08:53.604548375 +0000 UTC m=+857.508667085" Oct 05 21:08:53 crc kubenswrapper[4754]: I1005 21:08:53.620390 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-6cbc6dd547-zq79b" podStartSLOduration=8.033995441 podStartE2EDuration="35.620364399s" podCreationTimestamp="2025-10-05 21:08:18 +0000 UTC" firstStartedPulling="2025-10-05 21:08:21.221363852 +0000 UTC m=+825.125482552" lastFinishedPulling="2025-10-05 21:08:48.80773281 +0000 UTC m=+852.711851510" observedRunningTime="2025-10-05 21:08:53.614072104 +0000 UTC m=+857.518190814" watchObservedRunningTime="2025-10-05 21:08:53.620364399 +0000 UTC m=+857.524483149" Oct 05 21:08:53 crc kubenswrapper[4754]: E1005 21:08:53.624330 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/neutron-operator@sha256:dfd044635f9df9ed1d249387fa622177db35cdc72475e1c570617b8d17c64862\\\"\"" pod="openstack-operators/neutron-operator-controller-manager-8d984cc4d-7xzhb" podUID="ff3ba6ea-c2d8-4a3e-b0a0-58fd6df5b196" Oct 05 21:08:54 crc kubenswrapper[4754]: I1005 21:08:54.397127 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-655d88ccb9-l56qv" Oct 05 21:08:54 crc kubenswrapper[4754]: I1005 21:08:54.397656 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-8f58bc9db-xsrhv" Oct 05 21:08:54 crc kubenswrapper[4754]: I1005 21:08:54.398740 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-7d4d4f8d-kcxvd" Oct 05 21:08:55 crc kubenswrapper[4754]: I1005 21:08:55.410625 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-649675d675-v7285" event={"ID":"3ddbeeee-8a3e-4f3c-97e0-c7d8d5929345","Type":"ContainerStarted","Data":"22e4cce77ae13e083ec2333a82f55953baff2fb62997d56bed77653b6b9e9e82"} Oct 05 21:08:55 crc kubenswrapper[4754]: I1005 21:08:55.411138 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-649675d675-v7285" Oct 05 21:08:55 crc kubenswrapper[4754]: I1005 21:08:55.414452 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-65d89cfd9f-msh87" event={"ID":"5757c596-103f-4a00-ac16-ed9e6c9e4719","Type":"ContainerStarted","Data":"491991f7aa4d80bd8e05eee742c5ed9c66f303c038ce07e680c2c8996487a228"} Oct 05 21:08:55 crc kubenswrapper[4754]: I1005 21:08:55.414799 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-65d89cfd9f-msh87" Oct 05 21:08:55 crc kubenswrapper[4754]: I1005 21:08:55.419973 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-54689d9f88-4lrzz" event={"ID":"6316859b-8ab7-4b46-b1fa-c7f893d39a95","Type":"ContainerStarted","Data":"fd0eeed343a798c5ca982963cb7f4c5e5b86a870a167af64d1af263d265792c7"} Oct 05 21:08:55 crc kubenswrapper[4754]: I1005 21:08:55.420267 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-54689d9f88-4lrzz" Oct 05 21:08:55 crc kubenswrapper[4754]: I1005 21:08:55.437216 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-579449c7d5-zcct6" event={"ID":"b18dee24-0d55-4805-8d76-623180998686","Type":"ContainerStarted","Data":"84bd3faf5782b22bcf4b9a3eb53d9a62949d6cfb3167086e0bc7a6f7514d5670"} Oct 05 21:08:55 crc kubenswrapper[4754]: I1005 21:08:55.437401 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-649675d675-v7285" podStartSLOduration=4.3459694429999995 podStartE2EDuration="38.43738747s" podCreationTimestamp="2025-10-05 21:08:17 +0000 UTC" firstStartedPulling="2025-10-05 21:08:20.663003434 +0000 UTC m=+824.567122144" lastFinishedPulling="2025-10-05 21:08:54.754421471 +0000 UTC m=+858.658540171" observedRunningTime="2025-10-05 21:08:55.435219574 +0000 UTC m=+859.339338304" watchObservedRunningTime="2025-10-05 21:08:55.43738747 +0000 UTC m=+859.341506180" Oct 05 21:08:55 crc kubenswrapper[4754]: I1005 21:08:55.437475 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-579449c7d5-zcct6" Oct 05 21:08:55 crc kubenswrapper[4754]: I1005 21:08:55.443974 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-6859f9b676-hgkv4" event={"ID":"2b729715-7fd9-4ca5-b4dd-cb0eb0034aac","Type":"ContainerStarted","Data":"445a8205b43d915615d2b596500dd24cd8434bb0da683af3c83bee3d5f914a99"} Oct 05 21:08:55 crc kubenswrapper[4754]: I1005 21:08:55.444686 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-6859f9b676-hgkv4" Oct 05 21:08:55 crc kubenswrapper[4754]: I1005 21:08:55.451363 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-7468f855d8-w8z85" event={"ID":"cb4a4798-9283-4633-9b3b-e2d72faa221f","Type":"ContainerStarted","Data":"9ca80573d86b3ed9bf53b74191e858b72b15fe00c6ef86f0603eb1b08218fa07"} Oct 05 21:08:55 crc kubenswrapper[4754]: I1005 21:08:55.452217 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-7468f855d8-w8z85" Oct 05 21:08:55 crc kubenswrapper[4754]: I1005 21:08:55.457161 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-658588b8c9-z7tq9" event={"ID":"4bb20421-ce34-44ee-8740-82eedd4716f3","Type":"ContainerStarted","Data":"6d824c9854f2b1f00eccc4163b42b8b3c2e1b607594e22f2734585813dbce71e"} Oct 05 21:08:55 crc kubenswrapper[4754]: I1005 21:08:55.457944 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-658588b8c9-z7tq9" Oct 05 21:08:55 crc kubenswrapper[4754]: I1005 21:08:55.462405 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rz4fg" event={"ID":"92d8d285-7fd2-49a6-80b2-c72270bf1164","Type":"ContainerStarted","Data":"319a83a7eccd933fcd2cf1d953c9aeb4c55b86744ba95b122118d8491f28dcf9"} Oct 05 21:08:55 crc kubenswrapper[4754]: I1005 21:08:55.468296 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-7c7fc454ff-4fb5b" event={"ID":"e0b701ef-8b8d-4717-aa27-07233f6b6c15","Type":"ContainerStarted","Data":"5597a18f437722a58014b0e72e3846401fd7a445c39ceef6b9ed00986720ebb4"} Oct 05 21:08:55 crc kubenswrapper[4754]: I1005 21:08:55.468670 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-7c7fc454ff-4fb5b" Oct 05 21:08:55 crc kubenswrapper[4754]: I1005 21:08:55.469275 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-65d89cfd9f-msh87" podStartSLOduration=4.352164296 podStartE2EDuration="38.469250394s" podCreationTimestamp="2025-10-05 21:08:17 +0000 UTC" firstStartedPulling="2025-10-05 21:08:20.688536912 +0000 UTC m=+824.592655622" lastFinishedPulling="2025-10-05 21:08:54.80562301 +0000 UTC m=+858.709741720" observedRunningTime="2025-10-05 21:08:55.465117926 +0000 UTC m=+859.369236636" watchObservedRunningTime="2025-10-05 21:08:55.469250394 +0000 UTC m=+859.373369104" Oct 05 21:08:55 crc kubenswrapper[4754]: I1005 21:08:55.475873 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-6cd6d7bdf5-f5jxx" event={"ID":"5a1312fe-c6d6-41de-a442-b044db904a6d","Type":"ContainerStarted","Data":"9e049dc716d5dec68a1e7764c6f46bbac6966d8d97350ac05e40020dad72d975"} Oct 05 21:08:55 crc kubenswrapper[4754]: I1005 21:08:55.494288 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-54689d9f88-4lrzz" podStartSLOduration=3.362902846 podStartE2EDuration="37.494264888s" podCreationTimestamp="2025-10-05 21:08:18 +0000 UTC" firstStartedPulling="2025-10-05 21:08:20.679997909 +0000 UTC m=+824.584116619" lastFinishedPulling="2025-10-05 21:08:54.811359951 +0000 UTC m=+858.715478661" observedRunningTime="2025-10-05 21:08:55.491360152 +0000 UTC m=+859.395478852" watchObservedRunningTime="2025-10-05 21:08:55.494264888 +0000 UTC m=+859.398383598" Oct 05 21:08:55 crc kubenswrapper[4754]: I1005 21:08:55.521970 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-7468f855d8-w8z85" podStartSLOduration=4.879289867 podStartE2EDuration="38.521944853s" podCreationTimestamp="2025-10-05 21:08:17 +0000 UTC" firstStartedPulling="2025-10-05 21:08:21.106015394 +0000 UTC m=+825.010134104" lastFinishedPulling="2025-10-05 21:08:54.74867038 +0000 UTC m=+858.652789090" observedRunningTime="2025-10-05 21:08:55.516300175 +0000 UTC m=+859.420418885" watchObservedRunningTime="2025-10-05 21:08:55.521944853 +0000 UTC m=+859.426063563" Oct 05 21:08:55 crc kubenswrapper[4754]: I1005 21:08:55.554890 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-579449c7d5-zcct6" podStartSLOduration=4.606068114 podStartE2EDuration="37.554866384s" podCreationTimestamp="2025-10-05 21:08:18 +0000 UTC" firstStartedPulling="2025-10-05 21:08:20.675978004 +0000 UTC m=+824.580096714" lastFinishedPulling="2025-10-05 21:08:53.624776274 +0000 UTC m=+857.528894984" observedRunningTime="2025-10-05 21:08:55.550861719 +0000 UTC m=+859.454980429" watchObservedRunningTime="2025-10-05 21:08:55.554866384 +0000 UTC m=+859.458985094" Oct 05 21:08:55 crc kubenswrapper[4754]: I1005 21:08:55.644405 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-658588b8c9-z7tq9" podStartSLOduration=4.941206447 podStartE2EDuration="38.644379566s" podCreationTimestamp="2025-10-05 21:08:17 +0000 UTC" firstStartedPulling="2025-10-05 21:08:21.090790816 +0000 UTC m=+824.994909526" lastFinishedPulling="2025-10-05 21:08:54.793963935 +0000 UTC m=+858.698082645" observedRunningTime="2025-10-05 21:08:55.603320842 +0000 UTC m=+859.507439552" watchObservedRunningTime="2025-10-05 21:08:55.644379566 +0000 UTC m=+859.548498276" Oct 05 21:08:55 crc kubenswrapper[4754]: I1005 21:08:55.648844 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-6859f9b676-hgkv4" podStartSLOduration=3.509586165 podStartE2EDuration="37.648817682s" podCreationTimestamp="2025-10-05 21:08:18 +0000 UTC" firstStartedPulling="2025-10-05 21:08:20.706597385 +0000 UTC m=+824.610716095" lastFinishedPulling="2025-10-05 21:08:54.845828902 +0000 UTC m=+858.749947612" observedRunningTime="2025-10-05 21:08:55.643608716 +0000 UTC m=+859.547727426" watchObservedRunningTime="2025-10-05 21:08:55.648817682 +0000 UTC m=+859.552936392" Oct 05 21:08:55 crc kubenswrapper[4754]: I1005 21:08:55.827470 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-6cd6d7bdf5-f5jxx" podStartSLOduration=4.228798927 podStartE2EDuration="38.827450376s" podCreationTimestamp="2025-10-05 21:08:17 +0000 UTC" firstStartedPulling="2025-10-05 21:08:20.215198397 +0000 UTC m=+824.119317107" lastFinishedPulling="2025-10-05 21:08:54.813849846 +0000 UTC m=+858.717968556" observedRunningTime="2025-10-05 21:08:55.824969611 +0000 UTC m=+859.729088311" watchObservedRunningTime="2025-10-05 21:08:55.827450376 +0000 UTC m=+859.731569086" Oct 05 21:08:55 crc kubenswrapper[4754]: I1005 21:08:55.880429 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-7c7fc454ff-4fb5b" podStartSLOduration=5.23751386 podStartE2EDuration="38.880409662s" podCreationTimestamp="2025-10-05 21:08:17 +0000 UTC" firstStartedPulling="2025-10-05 21:08:21.105724087 +0000 UTC m=+825.009842797" lastFinishedPulling="2025-10-05 21:08:54.748619889 +0000 UTC m=+858.652738599" observedRunningTime="2025-10-05 21:08:55.8772702 +0000 UTC m=+859.781388910" watchObservedRunningTime="2025-10-05 21:08:55.880409662 +0000 UTC m=+859.784528372" Oct 05 21:08:56 crc kubenswrapper[4754]: I1005 21:08:56.487355 4754 generic.go:334] "Generic (PLEG): container finished" podID="92d8d285-7fd2-49a6-80b2-c72270bf1164" containerID="319a83a7eccd933fcd2cf1d953c9aeb4c55b86744ba95b122118d8491f28dcf9" exitCode=0 Oct 05 21:08:56 crc kubenswrapper[4754]: I1005 21:08:56.487440 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rz4fg" event={"ID":"92d8d285-7fd2-49a6-80b2-c72270bf1164","Type":"ContainerDied","Data":"319a83a7eccd933fcd2cf1d953c9aeb4c55b86744ba95b122118d8491f28dcf9"} Oct 05 21:08:56 crc kubenswrapper[4754]: I1005 21:08:56.495575 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fl6cv" event={"ID":"8a49c2cc-d61f-404e-946b-bb74451f895d","Type":"ContainerStarted","Data":"605ebedf259ab9bbc3a4491c6c01012f58371d58385324dda12d7075faf0957e"} Oct 05 21:08:57 crc kubenswrapper[4754]: I1005 21:08:57.503157 4754 generic.go:334] "Generic (PLEG): container finished" podID="8a49c2cc-d61f-404e-946b-bb74451f895d" containerID="605ebedf259ab9bbc3a4491c6c01012f58371d58385324dda12d7075faf0957e" exitCode=0 Oct 05 21:08:57 crc kubenswrapper[4754]: I1005 21:08:57.503284 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fl6cv" event={"ID":"8a49c2cc-d61f-404e-946b-bb74451f895d","Type":"ContainerDied","Data":"605ebedf259ab9bbc3a4491c6c01012f58371d58385324dda12d7075faf0957e"} Oct 05 21:08:57 crc kubenswrapper[4754]: I1005 21:08:57.505713 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rz4fg" event={"ID":"92d8d285-7fd2-49a6-80b2-c72270bf1164","Type":"ContainerStarted","Data":"9e594042f83cbe456e8700dd35ebc3149b6f47e7ddf92d9f04bad2553b63f7ca"} Oct 05 21:08:57 crc kubenswrapper[4754]: I1005 21:08:57.871664 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-fxx5s" Oct 05 21:08:57 crc kubenswrapper[4754]: I1005 21:08:57.876872 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-fxx5s" Oct 05 21:08:57 crc kubenswrapper[4754]: I1005 21:08:57.906133 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-rz4fg" podStartSLOduration=24.465597784 podStartE2EDuration="27.906100222s" podCreationTimestamp="2025-10-05 21:08:30 +0000 UTC" firstStartedPulling="2025-10-05 21:08:53.624833906 +0000 UTC m=+857.528952606" lastFinishedPulling="2025-10-05 21:08:57.065336344 +0000 UTC m=+860.969455044" observedRunningTime="2025-10-05 21:08:57.554976975 +0000 UTC m=+861.459095695" watchObservedRunningTime="2025-10-05 21:08:57.906100222 +0000 UTC m=+861.810218962" Oct 05 21:08:58 crc kubenswrapper[4754]: I1005 21:08:58.056784 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-54876c876f-svzwt" Oct 05 21:08:58 crc kubenswrapper[4754]: I1005 21:08:58.202693 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-5568b5d68-f5rjh" Oct 05 21:08:58 crc kubenswrapper[4754]: I1005 21:08:58.208752 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-5568b5d68-f5rjh" Oct 05 21:08:58 crc kubenswrapper[4754]: I1005 21:08:58.334767 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-456m8" Oct 05 21:08:58 crc kubenswrapper[4754]: I1005 21:08:58.334872 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-456m8" Oct 05 21:08:58 crc kubenswrapper[4754]: I1005 21:08:58.360179 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-6cd6d7bdf5-f5jxx" Oct 05 21:08:58 crc kubenswrapper[4754]: I1005 21:08:58.522513 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fl6cv" event={"ID":"8a49c2cc-d61f-404e-946b-bb74451f895d","Type":"ContainerStarted","Data":"c36e29b107fa44f342cfdce736365f51bfd3b8d03d11ee8823189be7f3c4b554"} Oct 05 21:08:58 crc kubenswrapper[4754]: I1005 21:08:58.553213 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-fl6cv" podStartSLOduration=15.916740433 podStartE2EDuration="20.553182173s" podCreationTimestamp="2025-10-05 21:08:38 +0000 UTC" firstStartedPulling="2025-10-05 21:08:53.584593153 +0000 UTC m=+857.488711863" lastFinishedPulling="2025-10-05 21:08:58.221034893 +0000 UTC m=+862.125153603" observedRunningTime="2025-10-05 21:08:58.54811689 +0000 UTC m=+862.452235600" watchObservedRunningTime="2025-10-05 21:08:58.553182173 +0000 UTC m=+862.457300883" Oct 05 21:08:58 crc kubenswrapper[4754]: I1005 21:08:58.885336 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-fl6cv" Oct 05 21:08:58 crc kubenswrapper[4754]: I1005 21:08:58.885645 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-fl6cv" Oct 05 21:08:59 crc kubenswrapper[4754]: I1005 21:08:59.286897 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-fx25z" Oct 05 21:08:59 crc kubenswrapper[4754]: I1005 21:08:59.289777 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-fx25z" Oct 05 21:08:59 crc kubenswrapper[4754]: I1005 21:08:59.321529 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-6cbc6dd547-zq79b" Oct 05 21:08:59 crc kubenswrapper[4754]: I1005 21:08:59.332809 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-6cbc6dd547-zq79b" Oct 05 21:08:59 crc kubenswrapper[4754]: I1005 21:08:59.382878 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-456m8" podUID="aee8b142-dea8-4ab3-87e9-14f3d3a75543" containerName="registry-server" probeResult="failure" output=< Oct 05 21:08:59 crc kubenswrapper[4754]: timeout: failed to connect service ":50051" within 1s Oct 05 21:08:59 crc kubenswrapper[4754]: > Oct 05 21:08:59 crc kubenswrapper[4754]: I1005 21:08:59.948972 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-fl6cv" podUID="8a49c2cc-d61f-404e-946b-bb74451f895d" containerName="registry-server" probeResult="failure" output=< Oct 05 21:08:59 crc kubenswrapper[4754]: timeout: failed to connect service ":50051" within 1s Oct 05 21:08:59 crc kubenswrapper[4754]: > Oct 05 21:08:59 crc kubenswrapper[4754]: I1005 21:08:59.981824 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665cfq946" Oct 05 21:09:00 crc kubenswrapper[4754]: I1005 21:09:00.550917 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-rz4fg" Oct 05 21:09:00 crc kubenswrapper[4754]: I1005 21:09:00.551074 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-rz4fg" Oct 05 21:09:01 crc kubenswrapper[4754]: I1005 21:09:01.624233 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-marketplace-rz4fg" podUID="92d8d285-7fd2-49a6-80b2-c72270bf1164" containerName="registry-server" probeResult="failure" output=< Oct 05 21:09:01 crc kubenswrapper[4754]: timeout: failed to connect service ":50051" within 1s Oct 05 21:09:01 crc kubenswrapper[4754]: > Oct 05 21:09:07 crc kubenswrapper[4754]: I1005 21:09:07.608373 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-5d4d74dd89-h44bd" event={"ID":"f48f3071-cd94-49dd-bfa0-1ec74f495c72","Type":"ContainerStarted","Data":"db217e94745ff78a64f0f381bd78a66da281f9e8332df2561ec7625624de4742"} Oct 05 21:09:07 crc kubenswrapper[4754]: I1005 21:09:07.609770 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-5d4d74dd89-h44bd" Oct 05 21:09:07 crc kubenswrapper[4754]: I1005 21:09:07.630200 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-5d4d74dd89-h44bd" podStartSLOduration=4.080307605 podStartE2EDuration="49.630170796s" podCreationTimestamp="2025-10-05 21:08:18 +0000 UTC" firstStartedPulling="2025-10-05 21:08:20.968832095 +0000 UTC m=+824.872950805" lastFinishedPulling="2025-10-05 21:09:06.518695256 +0000 UTC m=+870.422813996" observedRunningTime="2025-10-05 21:09:07.625793311 +0000 UTC m=+871.529912031" watchObservedRunningTime="2025-10-05 21:09:07.630170796 +0000 UTC m=+871.534289516" Oct 05 21:09:08 crc kubenswrapper[4754]: I1005 21:09:08.309449 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-65d89cfd9f-msh87" Oct 05 21:09:08 crc kubenswrapper[4754]: I1005 21:09:08.363799 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-6cd6d7bdf5-f5jxx" Oct 05 21:09:08 crc kubenswrapper[4754]: I1005 21:09:08.421039 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-7468f855d8-w8z85" Oct 05 21:09:08 crc kubenswrapper[4754]: I1005 21:09:08.472360 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-7c7fc454ff-4fb5b" Oct 05 21:09:08 crc kubenswrapper[4754]: I1005 21:09:08.479020 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-649675d675-v7285" Oct 05 21:09:08 crc kubenswrapper[4754]: I1005 21:09:08.679440 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-579449c7d5-zcct6" Oct 05 21:09:08 crc kubenswrapper[4754]: I1005 21:09:08.862256 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-54689d9f88-4lrzz" Oct 05 21:09:08 crc kubenswrapper[4754]: I1005 21:09:08.903300 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-658588b8c9-z7tq9" Oct 05 21:09:08 crc kubenswrapper[4754]: I1005 21:09:08.954599 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-fl6cv" Oct 05 21:09:08 crc kubenswrapper[4754]: I1005 21:09:08.984372 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-6859f9b676-hgkv4" Oct 05 21:09:09 crc kubenswrapper[4754]: I1005 21:09:09.023828 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-fl6cv" Oct 05 21:09:09 crc kubenswrapper[4754]: I1005 21:09:09.381755 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-456m8" podUID="aee8b142-dea8-4ab3-87e9-14f3d3a75543" containerName="registry-server" probeResult="failure" output=< Oct 05 21:09:09 crc kubenswrapper[4754]: timeout: failed to connect service ":50051" within 1s Oct 05 21:09:09 crc kubenswrapper[4754]: > Oct 05 21:09:09 crc kubenswrapper[4754]: I1005 21:09:09.629941 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-8d984cc4d-7xzhb" event={"ID":"ff3ba6ea-c2d8-4a3e-b0a0-58fd6df5b196","Type":"ContainerStarted","Data":"5263c731f7a7837dca973fc783e4f027cbd6a41a8f34814cd9e80fa40f6688e8"} Oct 05 21:09:09 crc kubenswrapper[4754]: I1005 21:09:09.630910 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-8d984cc4d-7xzhb" Oct 05 21:09:09 crc kubenswrapper[4754]: I1005 21:09:09.649586 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-8d984cc4d-7xzhb" podStartSLOduration=4.951307932 podStartE2EDuration="52.649570213s" podCreationTimestamp="2025-10-05 21:08:17 +0000 UTC" firstStartedPulling="2025-10-05 21:08:20.662743467 +0000 UTC m=+824.566862167" lastFinishedPulling="2025-10-05 21:09:08.361005728 +0000 UTC m=+872.265124448" observedRunningTime="2025-10-05 21:09:09.645392113 +0000 UTC m=+873.549510833" watchObservedRunningTime="2025-10-05 21:09:09.649570213 +0000 UTC m=+873.553688923" Oct 05 21:09:09 crc kubenswrapper[4754]: I1005 21:09:09.740772 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-fl6cv"] Oct 05 21:09:10 crc kubenswrapper[4754]: I1005 21:09:10.613426 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-rz4fg" Oct 05 21:09:10 crc kubenswrapper[4754]: I1005 21:09:10.646712 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-5f7c849b98-n9ksx" event={"ID":"9a1cc24e-8371-44cb-bd37-5b765853fdff","Type":"ContainerStarted","Data":"19ea200c164d850a741b4fc24c87fdbd57647981742ec467a3d9888909c8fb2d"} Oct 05 21:09:10 crc kubenswrapper[4754]: I1005 21:09:10.646935 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-fl6cv" podUID="8a49c2cc-d61f-404e-946b-bb74451f895d" containerName="registry-server" containerID="cri-o://c36e29b107fa44f342cfdce736365f51bfd3b8d03d11ee8823189be7f3c4b554" gracePeriod=2 Oct 05 21:09:10 crc kubenswrapper[4754]: I1005 21:09:10.647612 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-5f7c849b98-n9ksx" Oct 05 21:09:10 crc kubenswrapper[4754]: I1005 21:09:10.674993 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-rz4fg" Oct 05 21:09:10 crc kubenswrapper[4754]: I1005 21:09:10.698479 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-5f7c849b98-n9ksx" podStartSLOduration=4.168035438 podStartE2EDuration="53.698455676s" podCreationTimestamp="2025-10-05 21:08:17 +0000 UTC" firstStartedPulling="2025-10-05 21:08:19.833662985 +0000 UTC m=+823.737781695" lastFinishedPulling="2025-10-05 21:09:09.364083223 +0000 UTC m=+873.268201933" observedRunningTime="2025-10-05 21:09:10.67183699 +0000 UTC m=+874.575955700" watchObservedRunningTime="2025-10-05 21:09:10.698455676 +0000 UTC m=+874.602574396" Oct 05 21:09:11 crc kubenswrapper[4754]: I1005 21:09:11.132984 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fl6cv" Oct 05 21:09:11 crc kubenswrapper[4754]: I1005 21:09:11.314878 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8a49c2cc-d61f-404e-946b-bb74451f895d-catalog-content\") pod \"8a49c2cc-d61f-404e-946b-bb74451f895d\" (UID: \"8a49c2cc-d61f-404e-946b-bb74451f895d\") " Oct 05 21:09:11 crc kubenswrapper[4754]: I1005 21:09:11.315403 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8a49c2cc-d61f-404e-946b-bb74451f895d-utilities\") pod \"8a49c2cc-d61f-404e-946b-bb74451f895d\" (UID: \"8a49c2cc-d61f-404e-946b-bb74451f895d\") " Oct 05 21:09:11 crc kubenswrapper[4754]: I1005 21:09:11.315468 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2xxjr\" (UniqueName: \"kubernetes.io/projected/8a49c2cc-d61f-404e-946b-bb74451f895d-kube-api-access-2xxjr\") pod \"8a49c2cc-d61f-404e-946b-bb74451f895d\" (UID: \"8a49c2cc-d61f-404e-946b-bb74451f895d\") " Oct 05 21:09:11 crc kubenswrapper[4754]: I1005 21:09:11.316363 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8a49c2cc-d61f-404e-946b-bb74451f895d-utilities" (OuterVolumeSpecName: "utilities") pod "8a49c2cc-d61f-404e-946b-bb74451f895d" (UID: "8a49c2cc-d61f-404e-946b-bb74451f895d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 21:09:11 crc kubenswrapper[4754]: I1005 21:09:11.336113 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a49c2cc-d61f-404e-946b-bb74451f895d-kube-api-access-2xxjr" (OuterVolumeSpecName: "kube-api-access-2xxjr") pod "8a49c2cc-d61f-404e-946b-bb74451f895d" (UID: "8a49c2cc-d61f-404e-946b-bb74451f895d"). InnerVolumeSpecName "kube-api-access-2xxjr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:09:11 crc kubenswrapper[4754]: I1005 21:09:11.365626 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8a49c2cc-d61f-404e-946b-bb74451f895d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8a49c2cc-d61f-404e-946b-bb74451f895d" (UID: "8a49c2cc-d61f-404e-946b-bb74451f895d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 21:09:11 crc kubenswrapper[4754]: I1005 21:09:11.417944 4754 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8a49c2cc-d61f-404e-946b-bb74451f895d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 21:09:11 crc kubenswrapper[4754]: I1005 21:09:11.418005 4754 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8a49c2cc-d61f-404e-946b-bb74451f895d-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 21:09:11 crc kubenswrapper[4754]: I1005 21:09:11.418017 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2xxjr\" (UniqueName: \"kubernetes.io/projected/8a49c2cc-d61f-404e-946b-bb74451f895d-kube-api-access-2xxjr\") on node \"crc\" DevicePath \"\"" Oct 05 21:09:11 crc kubenswrapper[4754]: I1005 21:09:11.661793 4754 generic.go:334] "Generic (PLEG): container finished" podID="8a49c2cc-d61f-404e-946b-bb74451f895d" containerID="c36e29b107fa44f342cfdce736365f51bfd3b8d03d11ee8823189be7f3c4b554" exitCode=0 Oct 05 21:09:11 crc kubenswrapper[4754]: I1005 21:09:11.662819 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fl6cv" event={"ID":"8a49c2cc-d61f-404e-946b-bb74451f895d","Type":"ContainerDied","Data":"c36e29b107fa44f342cfdce736365f51bfd3b8d03d11ee8823189be7f3c4b554"} Oct 05 21:09:11 crc kubenswrapper[4754]: I1005 21:09:11.662876 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fl6cv" event={"ID":"8a49c2cc-d61f-404e-946b-bb74451f895d","Type":"ContainerDied","Data":"72b713d766a0be917aa3e435f4ea8ae4b520050643fd4ac6fca3c1289a10bd08"} Oct 05 21:09:11 crc kubenswrapper[4754]: I1005 21:09:11.662899 4754 scope.go:117] "RemoveContainer" containerID="c36e29b107fa44f342cfdce736365f51bfd3b8d03d11ee8823189be7f3c4b554" Oct 05 21:09:11 crc kubenswrapper[4754]: I1005 21:09:11.662943 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fl6cv" Oct 05 21:09:11 crc kubenswrapper[4754]: I1005 21:09:11.715550 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-fl6cv"] Oct 05 21:09:11 crc kubenswrapper[4754]: I1005 21:09:11.718633 4754 scope.go:117] "RemoveContainer" containerID="605ebedf259ab9bbc3a4491c6c01012f58371d58385324dda12d7075faf0957e" Oct 05 21:09:11 crc kubenswrapper[4754]: I1005 21:09:11.724700 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-fl6cv"] Oct 05 21:09:11 crc kubenswrapper[4754]: I1005 21:09:11.741833 4754 scope.go:117] "RemoveContainer" containerID="1d265c1fd4de76309817aba1e212a9f6666f0ec33471058ab8942dc4eba4f84a" Oct 05 21:09:11 crc kubenswrapper[4754]: I1005 21:09:11.785897 4754 scope.go:117] "RemoveContainer" containerID="c36e29b107fa44f342cfdce736365f51bfd3b8d03d11ee8823189be7f3c4b554" Oct 05 21:09:11 crc kubenswrapper[4754]: E1005 21:09:11.786542 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c36e29b107fa44f342cfdce736365f51bfd3b8d03d11ee8823189be7f3c4b554\": container with ID starting with c36e29b107fa44f342cfdce736365f51bfd3b8d03d11ee8823189be7f3c4b554 not found: ID does not exist" containerID="c36e29b107fa44f342cfdce736365f51bfd3b8d03d11ee8823189be7f3c4b554" Oct 05 21:09:11 crc kubenswrapper[4754]: I1005 21:09:11.786595 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c36e29b107fa44f342cfdce736365f51bfd3b8d03d11ee8823189be7f3c4b554"} err="failed to get container status \"c36e29b107fa44f342cfdce736365f51bfd3b8d03d11ee8823189be7f3c4b554\": rpc error: code = NotFound desc = could not find container \"c36e29b107fa44f342cfdce736365f51bfd3b8d03d11ee8823189be7f3c4b554\": container with ID starting with c36e29b107fa44f342cfdce736365f51bfd3b8d03d11ee8823189be7f3c4b554 not found: ID does not exist" Oct 05 21:09:11 crc kubenswrapper[4754]: I1005 21:09:11.786630 4754 scope.go:117] "RemoveContainer" containerID="605ebedf259ab9bbc3a4491c6c01012f58371d58385324dda12d7075faf0957e" Oct 05 21:09:11 crc kubenswrapper[4754]: E1005 21:09:11.787286 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"605ebedf259ab9bbc3a4491c6c01012f58371d58385324dda12d7075faf0957e\": container with ID starting with 605ebedf259ab9bbc3a4491c6c01012f58371d58385324dda12d7075faf0957e not found: ID does not exist" containerID="605ebedf259ab9bbc3a4491c6c01012f58371d58385324dda12d7075faf0957e" Oct 05 21:09:11 crc kubenswrapper[4754]: I1005 21:09:11.787316 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"605ebedf259ab9bbc3a4491c6c01012f58371d58385324dda12d7075faf0957e"} err="failed to get container status \"605ebedf259ab9bbc3a4491c6c01012f58371d58385324dda12d7075faf0957e\": rpc error: code = NotFound desc = could not find container \"605ebedf259ab9bbc3a4491c6c01012f58371d58385324dda12d7075faf0957e\": container with ID starting with 605ebedf259ab9bbc3a4491c6c01012f58371d58385324dda12d7075faf0957e not found: ID does not exist" Oct 05 21:09:11 crc kubenswrapper[4754]: I1005 21:09:11.787336 4754 scope.go:117] "RemoveContainer" containerID="1d265c1fd4de76309817aba1e212a9f6666f0ec33471058ab8942dc4eba4f84a" Oct 05 21:09:11 crc kubenswrapper[4754]: E1005 21:09:11.787775 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1d265c1fd4de76309817aba1e212a9f6666f0ec33471058ab8942dc4eba4f84a\": container with ID starting with 1d265c1fd4de76309817aba1e212a9f6666f0ec33471058ab8942dc4eba4f84a not found: ID does not exist" containerID="1d265c1fd4de76309817aba1e212a9f6666f0ec33471058ab8942dc4eba4f84a" Oct 05 21:09:11 crc kubenswrapper[4754]: I1005 21:09:11.787796 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d265c1fd4de76309817aba1e212a9f6666f0ec33471058ab8942dc4eba4f84a"} err="failed to get container status \"1d265c1fd4de76309817aba1e212a9f6666f0ec33471058ab8942dc4eba4f84a\": rpc error: code = NotFound desc = could not find container \"1d265c1fd4de76309817aba1e212a9f6666f0ec33471058ab8942dc4eba4f84a\": container with ID starting with 1d265c1fd4de76309817aba1e212a9f6666f0ec33471058ab8942dc4eba4f84a not found: ID does not exist" Oct 05 21:09:12 crc kubenswrapper[4754]: I1005 21:09:12.850693 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8a49c2cc-d61f-404e-946b-bb74451f895d" path="/var/lib/kubelet/pods/8a49c2cc-d61f-404e-946b-bb74451f895d/volumes" Oct 05 21:09:12 crc kubenswrapper[4754]: I1005 21:09:12.944063 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-rz4fg"] Oct 05 21:09:12 crc kubenswrapper[4754]: I1005 21:09:12.944330 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-rz4fg" podUID="92d8d285-7fd2-49a6-80b2-c72270bf1164" containerName="registry-server" containerID="cri-o://9e594042f83cbe456e8700dd35ebc3149b6f47e7ddf92d9f04bad2553b63f7ca" gracePeriod=2 Oct 05 21:09:13 crc kubenswrapper[4754]: I1005 21:09:13.379015 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rz4fg" Oct 05 21:09:13 crc kubenswrapper[4754]: I1005 21:09:13.552644 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92d8d285-7fd2-49a6-80b2-c72270bf1164-utilities\") pod \"92d8d285-7fd2-49a6-80b2-c72270bf1164\" (UID: \"92d8d285-7fd2-49a6-80b2-c72270bf1164\") " Oct 05 21:09:13 crc kubenswrapper[4754]: I1005 21:09:13.552697 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7kf8d\" (UniqueName: \"kubernetes.io/projected/92d8d285-7fd2-49a6-80b2-c72270bf1164-kube-api-access-7kf8d\") pod \"92d8d285-7fd2-49a6-80b2-c72270bf1164\" (UID: \"92d8d285-7fd2-49a6-80b2-c72270bf1164\") " Oct 05 21:09:13 crc kubenswrapper[4754]: I1005 21:09:13.552853 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92d8d285-7fd2-49a6-80b2-c72270bf1164-catalog-content\") pod \"92d8d285-7fd2-49a6-80b2-c72270bf1164\" (UID: \"92d8d285-7fd2-49a6-80b2-c72270bf1164\") " Oct 05 21:09:13 crc kubenswrapper[4754]: I1005 21:09:13.553479 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/92d8d285-7fd2-49a6-80b2-c72270bf1164-utilities" (OuterVolumeSpecName: "utilities") pod "92d8d285-7fd2-49a6-80b2-c72270bf1164" (UID: "92d8d285-7fd2-49a6-80b2-c72270bf1164"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 21:09:13 crc kubenswrapper[4754]: I1005 21:09:13.557931 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/92d8d285-7fd2-49a6-80b2-c72270bf1164-kube-api-access-7kf8d" (OuterVolumeSpecName: "kube-api-access-7kf8d") pod "92d8d285-7fd2-49a6-80b2-c72270bf1164" (UID: "92d8d285-7fd2-49a6-80b2-c72270bf1164"). InnerVolumeSpecName "kube-api-access-7kf8d". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:09:13 crc kubenswrapper[4754]: I1005 21:09:13.570287 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/92d8d285-7fd2-49a6-80b2-c72270bf1164-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "92d8d285-7fd2-49a6-80b2-c72270bf1164" (UID: "92d8d285-7fd2-49a6-80b2-c72270bf1164"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 21:09:13 crc kubenswrapper[4754]: I1005 21:09:13.654196 4754 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92d8d285-7fd2-49a6-80b2-c72270bf1164-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 21:09:13 crc kubenswrapper[4754]: I1005 21:09:13.654240 4754 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92d8d285-7fd2-49a6-80b2-c72270bf1164-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 21:09:13 crc kubenswrapper[4754]: I1005 21:09:13.654253 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7kf8d\" (UniqueName: \"kubernetes.io/projected/92d8d285-7fd2-49a6-80b2-c72270bf1164-kube-api-access-7kf8d\") on node \"crc\" DevicePath \"\"" Oct 05 21:09:13 crc kubenswrapper[4754]: I1005 21:09:13.682158 4754 generic.go:334] "Generic (PLEG): container finished" podID="92d8d285-7fd2-49a6-80b2-c72270bf1164" containerID="9e594042f83cbe456e8700dd35ebc3149b6f47e7ddf92d9f04bad2553b63f7ca" exitCode=0 Oct 05 21:09:13 crc kubenswrapper[4754]: I1005 21:09:13.682214 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rz4fg" event={"ID":"92d8d285-7fd2-49a6-80b2-c72270bf1164","Type":"ContainerDied","Data":"9e594042f83cbe456e8700dd35ebc3149b6f47e7ddf92d9f04bad2553b63f7ca"} Oct 05 21:09:13 crc kubenswrapper[4754]: I1005 21:09:13.682238 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rz4fg" Oct 05 21:09:13 crc kubenswrapper[4754]: I1005 21:09:13.682253 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rz4fg" event={"ID":"92d8d285-7fd2-49a6-80b2-c72270bf1164","Type":"ContainerDied","Data":"27dabd6a42fcc4669fda2e7bcbf264e49859a363597fe148ed4b527c78756a78"} Oct 05 21:09:13 crc kubenswrapper[4754]: I1005 21:09:13.682280 4754 scope.go:117] "RemoveContainer" containerID="9e594042f83cbe456e8700dd35ebc3149b6f47e7ddf92d9f04bad2553b63f7ca" Oct 05 21:09:13 crc kubenswrapper[4754]: I1005 21:09:13.707483 4754 scope.go:117] "RemoveContainer" containerID="319a83a7eccd933fcd2cf1d953c9aeb4c55b86744ba95b122118d8491f28dcf9" Oct 05 21:09:13 crc kubenswrapper[4754]: I1005 21:09:13.715739 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-rz4fg"] Oct 05 21:09:13 crc kubenswrapper[4754]: I1005 21:09:13.734835 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-rz4fg"] Oct 05 21:09:13 crc kubenswrapper[4754]: I1005 21:09:13.751055 4754 scope.go:117] "RemoveContainer" containerID="019301fbc8086a02212eeecc11c6bda9d83447d02cc0057f30170f3b0ef3f430" Oct 05 21:09:13 crc kubenswrapper[4754]: I1005 21:09:13.772175 4754 scope.go:117] "RemoveContainer" containerID="9e594042f83cbe456e8700dd35ebc3149b6f47e7ddf92d9f04bad2553b63f7ca" Oct 05 21:09:13 crc kubenswrapper[4754]: E1005 21:09:13.772785 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9e594042f83cbe456e8700dd35ebc3149b6f47e7ddf92d9f04bad2553b63f7ca\": container with ID starting with 9e594042f83cbe456e8700dd35ebc3149b6f47e7ddf92d9f04bad2553b63f7ca not found: ID does not exist" containerID="9e594042f83cbe456e8700dd35ebc3149b6f47e7ddf92d9f04bad2553b63f7ca" Oct 05 21:09:13 crc kubenswrapper[4754]: I1005 21:09:13.772837 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9e594042f83cbe456e8700dd35ebc3149b6f47e7ddf92d9f04bad2553b63f7ca"} err="failed to get container status \"9e594042f83cbe456e8700dd35ebc3149b6f47e7ddf92d9f04bad2553b63f7ca\": rpc error: code = NotFound desc = could not find container \"9e594042f83cbe456e8700dd35ebc3149b6f47e7ddf92d9f04bad2553b63f7ca\": container with ID starting with 9e594042f83cbe456e8700dd35ebc3149b6f47e7ddf92d9f04bad2553b63f7ca not found: ID does not exist" Oct 05 21:09:13 crc kubenswrapper[4754]: I1005 21:09:13.772879 4754 scope.go:117] "RemoveContainer" containerID="319a83a7eccd933fcd2cf1d953c9aeb4c55b86744ba95b122118d8491f28dcf9" Oct 05 21:09:13 crc kubenswrapper[4754]: E1005 21:09:13.773224 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"319a83a7eccd933fcd2cf1d953c9aeb4c55b86744ba95b122118d8491f28dcf9\": container with ID starting with 319a83a7eccd933fcd2cf1d953c9aeb4c55b86744ba95b122118d8491f28dcf9 not found: ID does not exist" containerID="319a83a7eccd933fcd2cf1d953c9aeb4c55b86744ba95b122118d8491f28dcf9" Oct 05 21:09:13 crc kubenswrapper[4754]: I1005 21:09:13.773288 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"319a83a7eccd933fcd2cf1d953c9aeb4c55b86744ba95b122118d8491f28dcf9"} err="failed to get container status \"319a83a7eccd933fcd2cf1d953c9aeb4c55b86744ba95b122118d8491f28dcf9\": rpc error: code = NotFound desc = could not find container \"319a83a7eccd933fcd2cf1d953c9aeb4c55b86744ba95b122118d8491f28dcf9\": container with ID starting with 319a83a7eccd933fcd2cf1d953c9aeb4c55b86744ba95b122118d8491f28dcf9 not found: ID does not exist" Oct 05 21:09:13 crc kubenswrapper[4754]: I1005 21:09:13.773355 4754 scope.go:117] "RemoveContainer" containerID="019301fbc8086a02212eeecc11c6bda9d83447d02cc0057f30170f3b0ef3f430" Oct 05 21:09:13 crc kubenswrapper[4754]: E1005 21:09:13.773985 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"019301fbc8086a02212eeecc11c6bda9d83447d02cc0057f30170f3b0ef3f430\": container with ID starting with 019301fbc8086a02212eeecc11c6bda9d83447d02cc0057f30170f3b0ef3f430 not found: ID does not exist" containerID="019301fbc8086a02212eeecc11c6bda9d83447d02cc0057f30170f3b0ef3f430" Oct 05 21:09:13 crc kubenswrapper[4754]: I1005 21:09:13.774050 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"019301fbc8086a02212eeecc11c6bda9d83447d02cc0057f30170f3b0ef3f430"} err="failed to get container status \"019301fbc8086a02212eeecc11c6bda9d83447d02cc0057f30170f3b0ef3f430\": rpc error: code = NotFound desc = could not find container \"019301fbc8086a02212eeecc11c6bda9d83447d02cc0057f30170f3b0ef3f430\": container with ID starting with 019301fbc8086a02212eeecc11c6bda9d83447d02cc0057f30170f3b0ef3f430 not found: ID does not exist" Oct 05 21:09:14 crc kubenswrapper[4754]: I1005 21:09:14.852117 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="92d8d285-7fd2-49a6-80b2-c72270bf1164" path="/var/lib/kubelet/pods/92d8d285-7fd2-49a6-80b2-c72270bf1164/volumes" Oct 05 21:09:18 crc kubenswrapper[4754]: I1005 21:09:18.145436 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-5f7c849b98-n9ksx" Oct 05 21:09:18 crc kubenswrapper[4754]: I1005 21:09:18.391958 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-8d984cc4d-7xzhb" Oct 05 21:09:18 crc kubenswrapper[4754]: I1005 21:09:18.959168 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-5d4d74dd89-h44bd" Oct 05 21:09:19 crc kubenswrapper[4754]: I1005 21:09:19.397197 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-456m8" podUID="aee8b142-dea8-4ab3-87e9-14f3d3a75543" containerName="registry-server" probeResult="failure" output=< Oct 05 21:09:19 crc kubenswrapper[4754]: timeout: failed to connect service ":50051" within 1s Oct 05 21:09:19 crc kubenswrapper[4754]: > Oct 05 21:09:28 crc kubenswrapper[4754]: I1005 21:09:28.390537 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-456m8" Oct 05 21:09:28 crc kubenswrapper[4754]: I1005 21:09:28.458244 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-456m8" Oct 05 21:09:33 crc kubenswrapper[4754]: I1005 21:09:33.276422 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-456m8"] Oct 05 21:09:33 crc kubenswrapper[4754]: I1005 21:09:33.277598 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-456m8" podUID="aee8b142-dea8-4ab3-87e9-14f3d3a75543" containerName="registry-server" containerID="cri-o://fb35c6fd17cd23d67f4fbdd6d4e4977d77bbb398d847e6b5ad80461080ecc8a7" gracePeriod=2 Oct 05 21:09:33 crc kubenswrapper[4754]: I1005 21:09:33.651438 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-456m8" Oct 05 21:09:33 crc kubenswrapper[4754]: I1005 21:09:33.804983 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bmvvr\" (UniqueName: \"kubernetes.io/projected/aee8b142-dea8-4ab3-87e9-14f3d3a75543-kube-api-access-bmvvr\") pod \"aee8b142-dea8-4ab3-87e9-14f3d3a75543\" (UID: \"aee8b142-dea8-4ab3-87e9-14f3d3a75543\") " Oct 05 21:09:33 crc kubenswrapper[4754]: I1005 21:09:33.805634 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aee8b142-dea8-4ab3-87e9-14f3d3a75543-catalog-content\") pod \"aee8b142-dea8-4ab3-87e9-14f3d3a75543\" (UID: \"aee8b142-dea8-4ab3-87e9-14f3d3a75543\") " Oct 05 21:09:33 crc kubenswrapper[4754]: I1005 21:09:33.805715 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aee8b142-dea8-4ab3-87e9-14f3d3a75543-utilities\") pod \"aee8b142-dea8-4ab3-87e9-14f3d3a75543\" (UID: \"aee8b142-dea8-4ab3-87e9-14f3d3a75543\") " Oct 05 21:09:33 crc kubenswrapper[4754]: I1005 21:09:33.806231 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aee8b142-dea8-4ab3-87e9-14f3d3a75543-utilities" (OuterVolumeSpecName: "utilities") pod "aee8b142-dea8-4ab3-87e9-14f3d3a75543" (UID: "aee8b142-dea8-4ab3-87e9-14f3d3a75543"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 21:09:33 crc kubenswrapper[4754]: I1005 21:09:33.819247 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aee8b142-dea8-4ab3-87e9-14f3d3a75543-kube-api-access-bmvvr" (OuterVolumeSpecName: "kube-api-access-bmvvr") pod "aee8b142-dea8-4ab3-87e9-14f3d3a75543" (UID: "aee8b142-dea8-4ab3-87e9-14f3d3a75543"). InnerVolumeSpecName "kube-api-access-bmvvr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:09:33 crc kubenswrapper[4754]: I1005 21:09:33.864044 4754 generic.go:334] "Generic (PLEG): container finished" podID="aee8b142-dea8-4ab3-87e9-14f3d3a75543" containerID="fb35c6fd17cd23d67f4fbdd6d4e4977d77bbb398d847e6b5ad80461080ecc8a7" exitCode=0 Oct 05 21:09:33 crc kubenswrapper[4754]: I1005 21:09:33.864342 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-456m8" event={"ID":"aee8b142-dea8-4ab3-87e9-14f3d3a75543","Type":"ContainerDied","Data":"fb35c6fd17cd23d67f4fbdd6d4e4977d77bbb398d847e6b5ad80461080ecc8a7"} Oct 05 21:09:33 crc kubenswrapper[4754]: I1005 21:09:33.864429 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-456m8" event={"ID":"aee8b142-dea8-4ab3-87e9-14f3d3a75543","Type":"ContainerDied","Data":"43d9575fbca631b0b9c86dabf6db03f572663a8c9fc41e9b57e634d93ad5a71a"} Oct 05 21:09:33 crc kubenswrapper[4754]: I1005 21:09:33.864528 4754 scope.go:117] "RemoveContainer" containerID="fb35c6fd17cd23d67f4fbdd6d4e4977d77bbb398d847e6b5ad80461080ecc8a7" Oct 05 21:09:33 crc kubenswrapper[4754]: I1005 21:09:33.864827 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-456m8" Oct 05 21:09:33 crc kubenswrapper[4754]: I1005 21:09:33.893236 4754 scope.go:117] "RemoveContainer" containerID="2ca19477259a6da6812707cc7775fbd87030fa5c33d188c619770772bf12e063" Oct 05 21:09:33 crc kubenswrapper[4754]: I1005 21:09:33.906587 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aee8b142-dea8-4ab3-87e9-14f3d3a75543-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "aee8b142-dea8-4ab3-87e9-14f3d3a75543" (UID: "aee8b142-dea8-4ab3-87e9-14f3d3a75543"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 21:09:33 crc kubenswrapper[4754]: I1005 21:09:33.907462 4754 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aee8b142-dea8-4ab3-87e9-14f3d3a75543-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 21:09:33 crc kubenswrapper[4754]: I1005 21:09:33.907488 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bmvvr\" (UniqueName: \"kubernetes.io/projected/aee8b142-dea8-4ab3-87e9-14f3d3a75543-kube-api-access-bmvvr\") on node \"crc\" DevicePath \"\"" Oct 05 21:09:33 crc kubenswrapper[4754]: I1005 21:09:33.907512 4754 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aee8b142-dea8-4ab3-87e9-14f3d3a75543-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 21:09:33 crc kubenswrapper[4754]: I1005 21:09:33.913702 4754 scope.go:117] "RemoveContainer" containerID="cf9684df762a16a28d139678817b7a635cac7b5268b7c6f9a53d8cfbcf0c147a" Oct 05 21:09:33 crc kubenswrapper[4754]: I1005 21:09:33.954118 4754 scope.go:117] "RemoveContainer" containerID="fb35c6fd17cd23d67f4fbdd6d4e4977d77bbb398d847e6b5ad80461080ecc8a7" Oct 05 21:09:33 crc kubenswrapper[4754]: E1005 21:09:33.955231 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fb35c6fd17cd23d67f4fbdd6d4e4977d77bbb398d847e6b5ad80461080ecc8a7\": container with ID starting with fb35c6fd17cd23d67f4fbdd6d4e4977d77bbb398d847e6b5ad80461080ecc8a7 not found: ID does not exist" containerID="fb35c6fd17cd23d67f4fbdd6d4e4977d77bbb398d847e6b5ad80461080ecc8a7" Oct 05 21:09:33 crc kubenswrapper[4754]: I1005 21:09:33.955325 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fb35c6fd17cd23d67f4fbdd6d4e4977d77bbb398d847e6b5ad80461080ecc8a7"} err="failed to get container status \"fb35c6fd17cd23d67f4fbdd6d4e4977d77bbb398d847e6b5ad80461080ecc8a7\": rpc error: code = NotFound desc = could not find container \"fb35c6fd17cd23d67f4fbdd6d4e4977d77bbb398d847e6b5ad80461080ecc8a7\": container with ID starting with fb35c6fd17cd23d67f4fbdd6d4e4977d77bbb398d847e6b5ad80461080ecc8a7 not found: ID does not exist" Oct 05 21:09:33 crc kubenswrapper[4754]: I1005 21:09:33.955396 4754 scope.go:117] "RemoveContainer" containerID="2ca19477259a6da6812707cc7775fbd87030fa5c33d188c619770772bf12e063" Oct 05 21:09:33 crc kubenswrapper[4754]: E1005 21:09:33.956126 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2ca19477259a6da6812707cc7775fbd87030fa5c33d188c619770772bf12e063\": container with ID starting with 2ca19477259a6da6812707cc7775fbd87030fa5c33d188c619770772bf12e063 not found: ID does not exist" containerID="2ca19477259a6da6812707cc7775fbd87030fa5c33d188c619770772bf12e063" Oct 05 21:09:33 crc kubenswrapper[4754]: I1005 21:09:33.956244 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2ca19477259a6da6812707cc7775fbd87030fa5c33d188c619770772bf12e063"} err="failed to get container status \"2ca19477259a6da6812707cc7775fbd87030fa5c33d188c619770772bf12e063\": rpc error: code = NotFound desc = could not find container \"2ca19477259a6da6812707cc7775fbd87030fa5c33d188c619770772bf12e063\": container with ID starting with 2ca19477259a6da6812707cc7775fbd87030fa5c33d188c619770772bf12e063 not found: ID does not exist" Oct 05 21:09:33 crc kubenswrapper[4754]: I1005 21:09:33.956329 4754 scope.go:117] "RemoveContainer" containerID="cf9684df762a16a28d139678817b7a635cac7b5268b7c6f9a53d8cfbcf0c147a" Oct 05 21:09:33 crc kubenswrapper[4754]: E1005 21:09:33.957877 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cf9684df762a16a28d139678817b7a635cac7b5268b7c6f9a53d8cfbcf0c147a\": container with ID starting with cf9684df762a16a28d139678817b7a635cac7b5268b7c6f9a53d8cfbcf0c147a not found: ID does not exist" containerID="cf9684df762a16a28d139678817b7a635cac7b5268b7c6f9a53d8cfbcf0c147a" Oct 05 21:09:33 crc kubenswrapper[4754]: I1005 21:09:33.957993 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cf9684df762a16a28d139678817b7a635cac7b5268b7c6f9a53d8cfbcf0c147a"} err="failed to get container status \"cf9684df762a16a28d139678817b7a635cac7b5268b7c6f9a53d8cfbcf0c147a\": rpc error: code = NotFound desc = could not find container \"cf9684df762a16a28d139678817b7a635cac7b5268b7c6f9a53d8cfbcf0c147a\": container with ID starting with cf9684df762a16a28d139678817b7a635cac7b5268b7c6f9a53d8cfbcf0c147a not found: ID does not exist" Oct 05 21:09:34 crc kubenswrapper[4754]: I1005 21:09:34.210292 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-456m8"] Oct 05 21:09:34 crc kubenswrapper[4754]: I1005 21:09:34.222820 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-456m8"] Oct 05 21:09:34 crc kubenswrapper[4754]: I1005 21:09:34.846091 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aee8b142-dea8-4ab3-87e9-14f3d3a75543" path="/var/lib/kubelet/pods/aee8b142-dea8-4ab3-87e9-14f3d3a75543/volumes" Oct 05 21:09:35 crc kubenswrapper[4754]: I1005 21:09:35.244645 4754 patch_prober.go:28] interesting pod/machine-config-daemon-b2h9k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 21:09:35 crc kubenswrapper[4754]: I1005 21:09:35.244705 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 21:09:35 crc kubenswrapper[4754]: I1005 21:09:35.529890 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-l4bph"] Oct 05 21:09:35 crc kubenswrapper[4754]: E1005 21:09:35.530296 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a49c2cc-d61f-404e-946b-bb74451f895d" containerName="extract-content" Oct 05 21:09:35 crc kubenswrapper[4754]: I1005 21:09:35.530309 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a49c2cc-d61f-404e-946b-bb74451f895d" containerName="extract-content" Oct 05 21:09:35 crc kubenswrapper[4754]: E1005 21:09:35.530349 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92d8d285-7fd2-49a6-80b2-c72270bf1164" containerName="extract-utilities" Oct 05 21:09:35 crc kubenswrapper[4754]: I1005 21:09:35.530356 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="92d8d285-7fd2-49a6-80b2-c72270bf1164" containerName="extract-utilities" Oct 05 21:09:35 crc kubenswrapper[4754]: E1005 21:09:35.530377 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a49c2cc-d61f-404e-946b-bb74451f895d" containerName="extract-utilities" Oct 05 21:09:35 crc kubenswrapper[4754]: I1005 21:09:35.530384 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a49c2cc-d61f-404e-946b-bb74451f895d" containerName="extract-utilities" Oct 05 21:09:35 crc kubenswrapper[4754]: E1005 21:09:35.530399 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aee8b142-dea8-4ab3-87e9-14f3d3a75543" containerName="registry-server" Oct 05 21:09:35 crc kubenswrapper[4754]: I1005 21:09:35.530406 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="aee8b142-dea8-4ab3-87e9-14f3d3a75543" containerName="registry-server" Oct 05 21:09:35 crc kubenswrapper[4754]: E1005 21:09:35.530415 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92d8d285-7fd2-49a6-80b2-c72270bf1164" containerName="registry-server" Oct 05 21:09:35 crc kubenswrapper[4754]: I1005 21:09:35.530420 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="92d8d285-7fd2-49a6-80b2-c72270bf1164" containerName="registry-server" Oct 05 21:09:35 crc kubenswrapper[4754]: E1005 21:09:35.530436 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a49c2cc-d61f-404e-946b-bb74451f895d" containerName="registry-server" Oct 05 21:09:35 crc kubenswrapper[4754]: I1005 21:09:35.530441 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a49c2cc-d61f-404e-946b-bb74451f895d" containerName="registry-server" Oct 05 21:09:35 crc kubenswrapper[4754]: E1005 21:09:35.530455 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aee8b142-dea8-4ab3-87e9-14f3d3a75543" containerName="extract-content" Oct 05 21:09:35 crc kubenswrapper[4754]: I1005 21:09:35.530462 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="aee8b142-dea8-4ab3-87e9-14f3d3a75543" containerName="extract-content" Oct 05 21:09:35 crc kubenswrapper[4754]: E1005 21:09:35.530472 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aee8b142-dea8-4ab3-87e9-14f3d3a75543" containerName="extract-utilities" Oct 05 21:09:35 crc kubenswrapper[4754]: I1005 21:09:35.530478 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="aee8b142-dea8-4ab3-87e9-14f3d3a75543" containerName="extract-utilities" Oct 05 21:09:35 crc kubenswrapper[4754]: E1005 21:09:35.530486 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92d8d285-7fd2-49a6-80b2-c72270bf1164" containerName="extract-content" Oct 05 21:09:35 crc kubenswrapper[4754]: I1005 21:09:35.530597 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="92d8d285-7fd2-49a6-80b2-c72270bf1164" containerName="extract-content" Oct 05 21:09:35 crc kubenswrapper[4754]: I1005 21:09:35.530773 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a49c2cc-d61f-404e-946b-bb74451f895d" containerName="registry-server" Oct 05 21:09:35 crc kubenswrapper[4754]: I1005 21:09:35.530824 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="aee8b142-dea8-4ab3-87e9-14f3d3a75543" containerName="registry-server" Oct 05 21:09:35 crc kubenswrapper[4754]: I1005 21:09:35.530840 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="92d8d285-7fd2-49a6-80b2-c72270bf1164" containerName="registry-server" Oct 05 21:09:35 crc kubenswrapper[4754]: I1005 21:09:35.531766 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-l4bph" Oct 05 21:09:35 crc kubenswrapper[4754]: I1005 21:09:35.535150 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Oct 05 21:09:35 crc kubenswrapper[4754]: I1005 21:09:35.535478 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-r2wr8" Oct 05 21:09:35 crc kubenswrapper[4754]: I1005 21:09:35.541320 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Oct 05 21:09:35 crc kubenswrapper[4754]: I1005 21:09:35.541851 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Oct 05 21:09:35 crc kubenswrapper[4754]: I1005 21:09:35.601650 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-l4bph"] Oct 05 21:09:35 crc kubenswrapper[4754]: I1005 21:09:35.621181 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-crmqb"] Oct 05 21:09:35 crc kubenswrapper[4754]: I1005 21:09:35.622712 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-crmqb" Oct 05 21:09:35 crc kubenswrapper[4754]: I1005 21:09:35.627941 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Oct 05 21:09:35 crc kubenswrapper[4754]: I1005 21:09:35.633820 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7lg65\" (UniqueName: \"kubernetes.io/projected/de3fa71c-9583-40d3-a47d-48ce71a7793c-kube-api-access-7lg65\") pod \"dnsmasq-dns-675f4bcbfc-l4bph\" (UID: \"de3fa71c-9583-40d3-a47d-48ce71a7793c\") " pod="openstack/dnsmasq-dns-675f4bcbfc-l4bph" Oct 05 21:09:35 crc kubenswrapper[4754]: I1005 21:09:35.633907 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/de3fa71c-9583-40d3-a47d-48ce71a7793c-config\") pod \"dnsmasq-dns-675f4bcbfc-l4bph\" (UID: \"de3fa71c-9583-40d3-a47d-48ce71a7793c\") " pod="openstack/dnsmasq-dns-675f4bcbfc-l4bph" Oct 05 21:09:35 crc kubenswrapper[4754]: I1005 21:09:35.655662 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-crmqb"] Oct 05 21:09:35 crc kubenswrapper[4754]: I1005 21:09:35.736618 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bcac2b2f-d2f3-438a-8bbd-92586d284527-config\") pod \"dnsmasq-dns-78dd6ddcc-crmqb\" (UID: \"bcac2b2f-d2f3-438a-8bbd-92586d284527\") " pod="openstack/dnsmasq-dns-78dd6ddcc-crmqb" Oct 05 21:09:35 crc kubenswrapper[4754]: I1005 21:09:35.736675 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-67tht\" (UniqueName: \"kubernetes.io/projected/bcac2b2f-d2f3-438a-8bbd-92586d284527-kube-api-access-67tht\") pod \"dnsmasq-dns-78dd6ddcc-crmqb\" (UID: \"bcac2b2f-d2f3-438a-8bbd-92586d284527\") " pod="openstack/dnsmasq-dns-78dd6ddcc-crmqb" Oct 05 21:09:35 crc kubenswrapper[4754]: I1005 21:09:35.736716 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7lg65\" (UniqueName: \"kubernetes.io/projected/de3fa71c-9583-40d3-a47d-48ce71a7793c-kube-api-access-7lg65\") pod \"dnsmasq-dns-675f4bcbfc-l4bph\" (UID: \"de3fa71c-9583-40d3-a47d-48ce71a7793c\") " pod="openstack/dnsmasq-dns-675f4bcbfc-l4bph" Oct 05 21:09:35 crc kubenswrapper[4754]: I1005 21:09:35.736745 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bcac2b2f-d2f3-438a-8bbd-92586d284527-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-crmqb\" (UID: \"bcac2b2f-d2f3-438a-8bbd-92586d284527\") " pod="openstack/dnsmasq-dns-78dd6ddcc-crmqb" Oct 05 21:09:35 crc kubenswrapper[4754]: I1005 21:09:35.736771 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/de3fa71c-9583-40d3-a47d-48ce71a7793c-config\") pod \"dnsmasq-dns-675f4bcbfc-l4bph\" (UID: \"de3fa71c-9583-40d3-a47d-48ce71a7793c\") " pod="openstack/dnsmasq-dns-675f4bcbfc-l4bph" Oct 05 21:09:35 crc kubenswrapper[4754]: I1005 21:09:35.739331 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/de3fa71c-9583-40d3-a47d-48ce71a7793c-config\") pod \"dnsmasq-dns-675f4bcbfc-l4bph\" (UID: \"de3fa71c-9583-40d3-a47d-48ce71a7793c\") " pod="openstack/dnsmasq-dns-675f4bcbfc-l4bph" Oct 05 21:09:35 crc kubenswrapper[4754]: I1005 21:09:35.761640 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7lg65\" (UniqueName: \"kubernetes.io/projected/de3fa71c-9583-40d3-a47d-48ce71a7793c-kube-api-access-7lg65\") pod \"dnsmasq-dns-675f4bcbfc-l4bph\" (UID: \"de3fa71c-9583-40d3-a47d-48ce71a7793c\") " pod="openstack/dnsmasq-dns-675f4bcbfc-l4bph" Oct 05 21:09:35 crc kubenswrapper[4754]: I1005 21:09:35.838171 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-67tht\" (UniqueName: \"kubernetes.io/projected/bcac2b2f-d2f3-438a-8bbd-92586d284527-kube-api-access-67tht\") pod \"dnsmasq-dns-78dd6ddcc-crmqb\" (UID: \"bcac2b2f-d2f3-438a-8bbd-92586d284527\") " pod="openstack/dnsmasq-dns-78dd6ddcc-crmqb" Oct 05 21:09:35 crc kubenswrapper[4754]: I1005 21:09:35.838319 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bcac2b2f-d2f3-438a-8bbd-92586d284527-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-crmqb\" (UID: \"bcac2b2f-d2f3-438a-8bbd-92586d284527\") " pod="openstack/dnsmasq-dns-78dd6ddcc-crmqb" Oct 05 21:09:35 crc kubenswrapper[4754]: I1005 21:09:35.838688 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bcac2b2f-d2f3-438a-8bbd-92586d284527-config\") pod \"dnsmasq-dns-78dd6ddcc-crmqb\" (UID: \"bcac2b2f-d2f3-438a-8bbd-92586d284527\") " pod="openstack/dnsmasq-dns-78dd6ddcc-crmqb" Oct 05 21:09:35 crc kubenswrapper[4754]: I1005 21:09:35.839738 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bcac2b2f-d2f3-438a-8bbd-92586d284527-config\") pod \"dnsmasq-dns-78dd6ddcc-crmqb\" (UID: \"bcac2b2f-d2f3-438a-8bbd-92586d284527\") " pod="openstack/dnsmasq-dns-78dd6ddcc-crmqb" Oct 05 21:09:35 crc kubenswrapper[4754]: I1005 21:09:35.839762 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bcac2b2f-d2f3-438a-8bbd-92586d284527-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-crmqb\" (UID: \"bcac2b2f-d2f3-438a-8bbd-92586d284527\") " pod="openstack/dnsmasq-dns-78dd6ddcc-crmqb" Oct 05 21:09:35 crc kubenswrapper[4754]: I1005 21:09:35.847415 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-l4bph" Oct 05 21:09:35 crc kubenswrapper[4754]: I1005 21:09:35.866724 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-67tht\" (UniqueName: \"kubernetes.io/projected/bcac2b2f-d2f3-438a-8bbd-92586d284527-kube-api-access-67tht\") pod \"dnsmasq-dns-78dd6ddcc-crmqb\" (UID: \"bcac2b2f-d2f3-438a-8bbd-92586d284527\") " pod="openstack/dnsmasq-dns-78dd6ddcc-crmqb" Oct 05 21:09:35 crc kubenswrapper[4754]: I1005 21:09:35.938206 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-crmqb" Oct 05 21:09:36 crc kubenswrapper[4754]: I1005 21:09:36.150232 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-l4bph"] Oct 05 21:09:36 crc kubenswrapper[4754]: I1005 21:09:36.161911 4754 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 05 21:09:36 crc kubenswrapper[4754]: I1005 21:09:36.468655 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-crmqb"] Oct 05 21:09:36 crc kubenswrapper[4754]: I1005 21:09:36.898147 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-crmqb" event={"ID":"bcac2b2f-d2f3-438a-8bbd-92586d284527","Type":"ContainerStarted","Data":"d741907e2653480b7e663c8b051a35f64e26d72e09ce813aec3a33cecc7e8918"} Oct 05 21:09:36 crc kubenswrapper[4754]: I1005 21:09:36.900244 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-l4bph" event={"ID":"de3fa71c-9583-40d3-a47d-48ce71a7793c","Type":"ContainerStarted","Data":"08fc02582a25331039f5d399075e8fb74a8e294bb09c5834fab9f7b6ed76551c"} Oct 05 21:09:38 crc kubenswrapper[4754]: I1005 21:09:38.191901 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-l4bph"] Oct 05 21:09:38 crc kubenswrapper[4754]: I1005 21:09:38.253100 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-wfrrs"] Oct 05 21:09:38 crc kubenswrapper[4754]: I1005 21:09:38.254794 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-wfrrs" Oct 05 21:09:38 crc kubenswrapper[4754]: I1005 21:09:38.281429 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-wfrrs"] Oct 05 21:09:38 crc kubenswrapper[4754]: I1005 21:09:38.430653 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tdg46\" (UniqueName: \"kubernetes.io/projected/c21fb2d2-38b1-41a0-8df8-d34aedfee879-kube-api-access-tdg46\") pod \"dnsmasq-dns-666b6646f7-wfrrs\" (UID: \"c21fb2d2-38b1-41a0-8df8-d34aedfee879\") " pod="openstack/dnsmasq-dns-666b6646f7-wfrrs" Oct 05 21:09:38 crc kubenswrapper[4754]: I1005 21:09:38.430779 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c21fb2d2-38b1-41a0-8df8-d34aedfee879-config\") pod \"dnsmasq-dns-666b6646f7-wfrrs\" (UID: \"c21fb2d2-38b1-41a0-8df8-d34aedfee879\") " pod="openstack/dnsmasq-dns-666b6646f7-wfrrs" Oct 05 21:09:38 crc kubenswrapper[4754]: I1005 21:09:38.430813 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c21fb2d2-38b1-41a0-8df8-d34aedfee879-dns-svc\") pod \"dnsmasq-dns-666b6646f7-wfrrs\" (UID: \"c21fb2d2-38b1-41a0-8df8-d34aedfee879\") " pod="openstack/dnsmasq-dns-666b6646f7-wfrrs" Oct 05 21:09:38 crc kubenswrapper[4754]: I1005 21:09:38.538278 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tdg46\" (UniqueName: \"kubernetes.io/projected/c21fb2d2-38b1-41a0-8df8-d34aedfee879-kube-api-access-tdg46\") pod \"dnsmasq-dns-666b6646f7-wfrrs\" (UID: \"c21fb2d2-38b1-41a0-8df8-d34aedfee879\") " pod="openstack/dnsmasq-dns-666b6646f7-wfrrs" Oct 05 21:09:38 crc kubenswrapper[4754]: I1005 21:09:38.538384 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c21fb2d2-38b1-41a0-8df8-d34aedfee879-config\") pod \"dnsmasq-dns-666b6646f7-wfrrs\" (UID: \"c21fb2d2-38b1-41a0-8df8-d34aedfee879\") " pod="openstack/dnsmasq-dns-666b6646f7-wfrrs" Oct 05 21:09:38 crc kubenswrapper[4754]: I1005 21:09:38.538440 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c21fb2d2-38b1-41a0-8df8-d34aedfee879-dns-svc\") pod \"dnsmasq-dns-666b6646f7-wfrrs\" (UID: \"c21fb2d2-38b1-41a0-8df8-d34aedfee879\") " pod="openstack/dnsmasq-dns-666b6646f7-wfrrs" Oct 05 21:09:38 crc kubenswrapper[4754]: I1005 21:09:38.539989 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c21fb2d2-38b1-41a0-8df8-d34aedfee879-config\") pod \"dnsmasq-dns-666b6646f7-wfrrs\" (UID: \"c21fb2d2-38b1-41a0-8df8-d34aedfee879\") " pod="openstack/dnsmasq-dns-666b6646f7-wfrrs" Oct 05 21:09:38 crc kubenswrapper[4754]: I1005 21:09:38.540031 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c21fb2d2-38b1-41a0-8df8-d34aedfee879-dns-svc\") pod \"dnsmasq-dns-666b6646f7-wfrrs\" (UID: \"c21fb2d2-38b1-41a0-8df8-d34aedfee879\") " pod="openstack/dnsmasq-dns-666b6646f7-wfrrs" Oct 05 21:09:38 crc kubenswrapper[4754]: I1005 21:09:38.603026 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tdg46\" (UniqueName: \"kubernetes.io/projected/c21fb2d2-38b1-41a0-8df8-d34aedfee879-kube-api-access-tdg46\") pod \"dnsmasq-dns-666b6646f7-wfrrs\" (UID: \"c21fb2d2-38b1-41a0-8df8-d34aedfee879\") " pod="openstack/dnsmasq-dns-666b6646f7-wfrrs" Oct 05 21:09:38 crc kubenswrapper[4754]: I1005 21:09:38.806340 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-crmqb"] Oct 05 21:09:38 crc kubenswrapper[4754]: I1005 21:09:38.882556 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-d5rsv"] Oct 05 21:09:38 crc kubenswrapper[4754]: I1005 21:09:38.883690 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-d5rsv" Oct 05 21:09:38 crc kubenswrapper[4754]: I1005 21:09:38.891277 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-wfrrs" Oct 05 21:09:38 crc kubenswrapper[4754]: I1005 21:09:38.904000 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-d5rsv"] Oct 05 21:09:39 crc kubenswrapper[4754]: I1005 21:09:39.046209 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22f48eb7-09d6-44bc-bc47-ba7026cf4cc3-config\") pod \"dnsmasq-dns-57d769cc4f-d5rsv\" (UID: \"22f48eb7-09d6-44bc-bc47-ba7026cf4cc3\") " pod="openstack/dnsmasq-dns-57d769cc4f-d5rsv" Oct 05 21:09:39 crc kubenswrapper[4754]: I1005 21:09:39.046272 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/22f48eb7-09d6-44bc-bc47-ba7026cf4cc3-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-d5rsv\" (UID: \"22f48eb7-09d6-44bc-bc47-ba7026cf4cc3\") " pod="openstack/dnsmasq-dns-57d769cc4f-d5rsv" Oct 05 21:09:39 crc kubenswrapper[4754]: I1005 21:09:39.046303 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dts6z\" (UniqueName: \"kubernetes.io/projected/22f48eb7-09d6-44bc-bc47-ba7026cf4cc3-kube-api-access-dts6z\") pod \"dnsmasq-dns-57d769cc4f-d5rsv\" (UID: \"22f48eb7-09d6-44bc-bc47-ba7026cf4cc3\") " pod="openstack/dnsmasq-dns-57d769cc4f-d5rsv" Oct 05 21:09:39 crc kubenswrapper[4754]: I1005 21:09:39.147549 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22f48eb7-09d6-44bc-bc47-ba7026cf4cc3-config\") pod \"dnsmasq-dns-57d769cc4f-d5rsv\" (UID: \"22f48eb7-09d6-44bc-bc47-ba7026cf4cc3\") " pod="openstack/dnsmasq-dns-57d769cc4f-d5rsv" Oct 05 21:09:39 crc kubenswrapper[4754]: I1005 21:09:39.147608 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/22f48eb7-09d6-44bc-bc47-ba7026cf4cc3-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-d5rsv\" (UID: \"22f48eb7-09d6-44bc-bc47-ba7026cf4cc3\") " pod="openstack/dnsmasq-dns-57d769cc4f-d5rsv" Oct 05 21:09:39 crc kubenswrapper[4754]: I1005 21:09:39.147635 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dts6z\" (UniqueName: \"kubernetes.io/projected/22f48eb7-09d6-44bc-bc47-ba7026cf4cc3-kube-api-access-dts6z\") pod \"dnsmasq-dns-57d769cc4f-d5rsv\" (UID: \"22f48eb7-09d6-44bc-bc47-ba7026cf4cc3\") " pod="openstack/dnsmasq-dns-57d769cc4f-d5rsv" Oct 05 21:09:39 crc kubenswrapper[4754]: I1005 21:09:39.148782 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22f48eb7-09d6-44bc-bc47-ba7026cf4cc3-config\") pod \"dnsmasq-dns-57d769cc4f-d5rsv\" (UID: \"22f48eb7-09d6-44bc-bc47-ba7026cf4cc3\") " pod="openstack/dnsmasq-dns-57d769cc4f-d5rsv" Oct 05 21:09:39 crc kubenswrapper[4754]: I1005 21:09:39.149283 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/22f48eb7-09d6-44bc-bc47-ba7026cf4cc3-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-d5rsv\" (UID: \"22f48eb7-09d6-44bc-bc47-ba7026cf4cc3\") " pod="openstack/dnsmasq-dns-57d769cc4f-d5rsv" Oct 05 21:09:39 crc kubenswrapper[4754]: I1005 21:09:39.169457 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dts6z\" (UniqueName: \"kubernetes.io/projected/22f48eb7-09d6-44bc-bc47-ba7026cf4cc3-kube-api-access-dts6z\") pod \"dnsmasq-dns-57d769cc4f-d5rsv\" (UID: \"22f48eb7-09d6-44bc-bc47-ba7026cf4cc3\") " pod="openstack/dnsmasq-dns-57d769cc4f-d5rsv" Oct 05 21:09:39 crc kubenswrapper[4754]: I1005 21:09:39.211667 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-d5rsv" Oct 05 21:09:39 crc kubenswrapper[4754]: I1005 21:09:39.469105 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Oct 05 21:09:39 crc kubenswrapper[4754]: I1005 21:09:39.472284 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 05 21:09:39 crc kubenswrapper[4754]: I1005 21:09:39.476420 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Oct 05 21:09:39 crc kubenswrapper[4754]: I1005 21:09:39.476768 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Oct 05 21:09:39 crc kubenswrapper[4754]: I1005 21:09:39.476889 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Oct 05 21:09:39 crc kubenswrapper[4754]: I1005 21:09:39.476940 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Oct 05 21:09:39 crc kubenswrapper[4754]: I1005 21:09:39.478119 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Oct 05 21:09:39 crc kubenswrapper[4754]: I1005 21:09:39.478264 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-dm69s" Oct 05 21:09:39 crc kubenswrapper[4754]: I1005 21:09:39.478487 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Oct 05 21:09:39 crc kubenswrapper[4754]: I1005 21:09:39.521576 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 05 21:09:39 crc kubenswrapper[4754]: I1005 21:09:39.556901 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/39aff8a0-2905-4ad3-ad2d-5bf6b5de3858-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"39aff8a0-2905-4ad3-ad2d-5bf6b5de3858\") " pod="openstack/rabbitmq-server-0" Oct 05 21:09:39 crc kubenswrapper[4754]: I1005 21:09:39.557035 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/39aff8a0-2905-4ad3-ad2d-5bf6b5de3858-config-data\") pod \"rabbitmq-server-0\" (UID: \"39aff8a0-2905-4ad3-ad2d-5bf6b5de3858\") " pod="openstack/rabbitmq-server-0" Oct 05 21:09:39 crc kubenswrapper[4754]: I1005 21:09:39.557118 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jkfvq\" (UniqueName: \"kubernetes.io/projected/39aff8a0-2905-4ad3-ad2d-5bf6b5de3858-kube-api-access-jkfvq\") pod \"rabbitmq-server-0\" (UID: \"39aff8a0-2905-4ad3-ad2d-5bf6b5de3858\") " pod="openstack/rabbitmq-server-0" Oct 05 21:09:39 crc kubenswrapper[4754]: I1005 21:09:39.557144 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-server-0\" (UID: \"39aff8a0-2905-4ad3-ad2d-5bf6b5de3858\") " pod="openstack/rabbitmq-server-0" Oct 05 21:09:39 crc kubenswrapper[4754]: I1005 21:09:39.557169 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/39aff8a0-2905-4ad3-ad2d-5bf6b5de3858-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"39aff8a0-2905-4ad3-ad2d-5bf6b5de3858\") " pod="openstack/rabbitmq-server-0" Oct 05 21:09:39 crc kubenswrapper[4754]: I1005 21:09:39.557187 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/39aff8a0-2905-4ad3-ad2d-5bf6b5de3858-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"39aff8a0-2905-4ad3-ad2d-5bf6b5de3858\") " pod="openstack/rabbitmq-server-0" Oct 05 21:09:39 crc kubenswrapper[4754]: I1005 21:09:39.557246 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/39aff8a0-2905-4ad3-ad2d-5bf6b5de3858-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"39aff8a0-2905-4ad3-ad2d-5bf6b5de3858\") " pod="openstack/rabbitmq-server-0" Oct 05 21:09:39 crc kubenswrapper[4754]: I1005 21:09:39.557273 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/39aff8a0-2905-4ad3-ad2d-5bf6b5de3858-server-conf\") pod \"rabbitmq-server-0\" (UID: \"39aff8a0-2905-4ad3-ad2d-5bf6b5de3858\") " pod="openstack/rabbitmq-server-0" Oct 05 21:09:39 crc kubenswrapper[4754]: I1005 21:09:39.557324 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/39aff8a0-2905-4ad3-ad2d-5bf6b5de3858-pod-info\") pod \"rabbitmq-server-0\" (UID: \"39aff8a0-2905-4ad3-ad2d-5bf6b5de3858\") " pod="openstack/rabbitmq-server-0" Oct 05 21:09:39 crc kubenswrapper[4754]: I1005 21:09:39.557352 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/39aff8a0-2905-4ad3-ad2d-5bf6b5de3858-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"39aff8a0-2905-4ad3-ad2d-5bf6b5de3858\") " pod="openstack/rabbitmq-server-0" Oct 05 21:09:39 crc kubenswrapper[4754]: I1005 21:09:39.557367 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/39aff8a0-2905-4ad3-ad2d-5bf6b5de3858-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"39aff8a0-2905-4ad3-ad2d-5bf6b5de3858\") " pod="openstack/rabbitmq-server-0" Oct 05 21:09:39 crc kubenswrapper[4754]: I1005 21:09:39.659335 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/39aff8a0-2905-4ad3-ad2d-5bf6b5de3858-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"39aff8a0-2905-4ad3-ad2d-5bf6b5de3858\") " pod="openstack/rabbitmq-server-0" Oct 05 21:09:39 crc kubenswrapper[4754]: I1005 21:09:39.659402 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/39aff8a0-2905-4ad3-ad2d-5bf6b5de3858-server-conf\") pod \"rabbitmq-server-0\" (UID: \"39aff8a0-2905-4ad3-ad2d-5bf6b5de3858\") " pod="openstack/rabbitmq-server-0" Oct 05 21:09:39 crc kubenswrapper[4754]: I1005 21:09:39.659441 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/39aff8a0-2905-4ad3-ad2d-5bf6b5de3858-pod-info\") pod \"rabbitmq-server-0\" (UID: \"39aff8a0-2905-4ad3-ad2d-5bf6b5de3858\") " pod="openstack/rabbitmq-server-0" Oct 05 21:09:39 crc kubenswrapper[4754]: I1005 21:09:39.659502 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/39aff8a0-2905-4ad3-ad2d-5bf6b5de3858-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"39aff8a0-2905-4ad3-ad2d-5bf6b5de3858\") " pod="openstack/rabbitmq-server-0" Oct 05 21:09:39 crc kubenswrapper[4754]: I1005 21:09:39.659524 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/39aff8a0-2905-4ad3-ad2d-5bf6b5de3858-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"39aff8a0-2905-4ad3-ad2d-5bf6b5de3858\") " pod="openstack/rabbitmq-server-0" Oct 05 21:09:39 crc kubenswrapper[4754]: I1005 21:09:39.659568 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/39aff8a0-2905-4ad3-ad2d-5bf6b5de3858-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"39aff8a0-2905-4ad3-ad2d-5bf6b5de3858\") " pod="openstack/rabbitmq-server-0" Oct 05 21:09:39 crc kubenswrapper[4754]: I1005 21:09:39.659608 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/39aff8a0-2905-4ad3-ad2d-5bf6b5de3858-config-data\") pod \"rabbitmq-server-0\" (UID: \"39aff8a0-2905-4ad3-ad2d-5bf6b5de3858\") " pod="openstack/rabbitmq-server-0" Oct 05 21:09:39 crc kubenswrapper[4754]: I1005 21:09:39.659630 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jkfvq\" (UniqueName: \"kubernetes.io/projected/39aff8a0-2905-4ad3-ad2d-5bf6b5de3858-kube-api-access-jkfvq\") pod \"rabbitmq-server-0\" (UID: \"39aff8a0-2905-4ad3-ad2d-5bf6b5de3858\") " pod="openstack/rabbitmq-server-0" Oct 05 21:09:39 crc kubenswrapper[4754]: I1005 21:09:39.659655 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-server-0\" (UID: \"39aff8a0-2905-4ad3-ad2d-5bf6b5de3858\") " pod="openstack/rabbitmq-server-0" Oct 05 21:09:39 crc kubenswrapper[4754]: I1005 21:09:39.659689 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/39aff8a0-2905-4ad3-ad2d-5bf6b5de3858-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"39aff8a0-2905-4ad3-ad2d-5bf6b5de3858\") " pod="openstack/rabbitmq-server-0" Oct 05 21:09:39 crc kubenswrapper[4754]: I1005 21:09:39.661209 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/39aff8a0-2905-4ad3-ad2d-5bf6b5de3858-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"39aff8a0-2905-4ad3-ad2d-5bf6b5de3858\") " pod="openstack/rabbitmq-server-0" Oct 05 21:09:39 crc kubenswrapper[4754]: I1005 21:09:39.661472 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/39aff8a0-2905-4ad3-ad2d-5bf6b5de3858-server-conf\") pod \"rabbitmq-server-0\" (UID: \"39aff8a0-2905-4ad3-ad2d-5bf6b5de3858\") " pod="openstack/rabbitmq-server-0" Oct 05 21:09:39 crc kubenswrapper[4754]: I1005 21:09:39.662550 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/39aff8a0-2905-4ad3-ad2d-5bf6b5de3858-config-data\") pod \"rabbitmq-server-0\" (UID: \"39aff8a0-2905-4ad3-ad2d-5bf6b5de3858\") " pod="openstack/rabbitmq-server-0" Oct 05 21:09:39 crc kubenswrapper[4754]: I1005 21:09:39.662871 4754 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-server-0\" (UID: \"39aff8a0-2905-4ad3-ad2d-5bf6b5de3858\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/rabbitmq-server-0" Oct 05 21:09:39 crc kubenswrapper[4754]: I1005 21:09:39.665795 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/39aff8a0-2905-4ad3-ad2d-5bf6b5de3858-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"39aff8a0-2905-4ad3-ad2d-5bf6b5de3858\") " pod="openstack/rabbitmq-server-0" Oct 05 21:09:39 crc kubenswrapper[4754]: I1005 21:09:39.666659 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/39aff8a0-2905-4ad3-ad2d-5bf6b5de3858-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"39aff8a0-2905-4ad3-ad2d-5bf6b5de3858\") " pod="openstack/rabbitmq-server-0" Oct 05 21:09:39 crc kubenswrapper[4754]: I1005 21:09:39.667228 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/39aff8a0-2905-4ad3-ad2d-5bf6b5de3858-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"39aff8a0-2905-4ad3-ad2d-5bf6b5de3858\") " pod="openstack/rabbitmq-server-0" Oct 05 21:09:39 crc kubenswrapper[4754]: I1005 21:09:39.672823 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/39aff8a0-2905-4ad3-ad2d-5bf6b5de3858-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"39aff8a0-2905-4ad3-ad2d-5bf6b5de3858\") " pod="openstack/rabbitmq-server-0" Oct 05 21:09:39 crc kubenswrapper[4754]: I1005 21:09:39.676315 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/39aff8a0-2905-4ad3-ad2d-5bf6b5de3858-pod-info\") pod \"rabbitmq-server-0\" (UID: \"39aff8a0-2905-4ad3-ad2d-5bf6b5de3858\") " pod="openstack/rabbitmq-server-0" Oct 05 21:09:39 crc kubenswrapper[4754]: I1005 21:09:39.681665 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/39aff8a0-2905-4ad3-ad2d-5bf6b5de3858-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"39aff8a0-2905-4ad3-ad2d-5bf6b5de3858\") " pod="openstack/rabbitmq-server-0" Oct 05 21:09:39 crc kubenswrapper[4754]: I1005 21:09:39.684997 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/39aff8a0-2905-4ad3-ad2d-5bf6b5de3858-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"39aff8a0-2905-4ad3-ad2d-5bf6b5de3858\") " pod="openstack/rabbitmq-server-0" Oct 05 21:09:39 crc kubenswrapper[4754]: I1005 21:09:39.688299 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jkfvq\" (UniqueName: \"kubernetes.io/projected/39aff8a0-2905-4ad3-ad2d-5bf6b5de3858-kube-api-access-jkfvq\") pod \"rabbitmq-server-0\" (UID: \"39aff8a0-2905-4ad3-ad2d-5bf6b5de3858\") " pod="openstack/rabbitmq-server-0" Oct 05 21:09:39 crc kubenswrapper[4754]: I1005 21:09:39.755573 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-server-0\" (UID: \"39aff8a0-2905-4ad3-ad2d-5bf6b5de3858\") " pod="openstack/rabbitmq-server-0" Oct 05 21:09:39 crc kubenswrapper[4754]: I1005 21:09:39.823173 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 05 21:09:39 crc kubenswrapper[4754]: I1005 21:09:39.825967 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-wfrrs"] Oct 05 21:09:39 crc kubenswrapper[4754]: W1005 21:09:39.842947 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc21fb2d2_38b1_41a0_8df8_d34aedfee879.slice/crio-4fa73152575258c410743c0061bcd9f156a2333f04bcf31056053bdc8b58743b WatchSource:0}: Error finding container 4fa73152575258c410743c0061bcd9f156a2333f04bcf31056053bdc8b58743b: Status 404 returned error can't find the container with id 4fa73152575258c410743c0061bcd9f156a2333f04bcf31056053bdc8b58743b Oct 05 21:09:39 crc kubenswrapper[4754]: I1005 21:09:39.957548 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-d5rsv"] Oct 05 21:09:39 crc kubenswrapper[4754]: I1005 21:09:39.978864 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-wfrrs" event={"ID":"c21fb2d2-38b1-41a0-8df8-d34aedfee879","Type":"ContainerStarted","Data":"4fa73152575258c410743c0061bcd9f156a2333f04bcf31056053bdc8b58743b"} Oct 05 21:09:40 crc kubenswrapper[4754]: W1005 21:09:40.030167 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod22f48eb7_09d6_44bc_bc47_ba7026cf4cc3.slice/crio-d36c72ea046cc25ee3cebcab4fdf3d09ac2f457fe2bf71446d7f8e45626f3854 WatchSource:0}: Error finding container d36c72ea046cc25ee3cebcab4fdf3d09ac2f457fe2bf71446d7f8e45626f3854: Status 404 returned error can't find the container with id d36c72ea046cc25ee3cebcab4fdf3d09ac2f457fe2bf71446d7f8e45626f3854 Oct 05 21:09:40 crc kubenswrapper[4754]: I1005 21:09:40.031471 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 05 21:09:40 crc kubenswrapper[4754]: I1005 21:09:40.033075 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 05 21:09:40 crc kubenswrapper[4754]: I1005 21:09:40.045266 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 05 21:09:40 crc kubenswrapper[4754]: I1005 21:09:40.046543 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Oct 05 21:09:40 crc kubenswrapper[4754]: I1005 21:09:40.046793 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Oct 05 21:09:40 crc kubenswrapper[4754]: I1005 21:09:40.047000 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Oct 05 21:09:40 crc kubenswrapper[4754]: I1005 21:09:40.051506 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Oct 05 21:09:40 crc kubenswrapper[4754]: I1005 21:09:40.051728 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Oct 05 21:09:40 crc kubenswrapper[4754]: I1005 21:09:40.051869 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-bkl2n" Oct 05 21:09:40 crc kubenswrapper[4754]: I1005 21:09:40.055249 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Oct 05 21:09:40 crc kubenswrapper[4754]: I1005 21:09:40.071811 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/1d96e280-8c26-41ea-ab29-64ce5b9c6a2c-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"1d96e280-8c26-41ea-ab29-64ce5b9c6a2c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 21:09:40 crc kubenswrapper[4754]: I1005 21:09:40.071883 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6ltd4\" (UniqueName: \"kubernetes.io/projected/1d96e280-8c26-41ea-ab29-64ce5b9c6a2c-kube-api-access-6ltd4\") pod \"rabbitmq-cell1-server-0\" (UID: \"1d96e280-8c26-41ea-ab29-64ce5b9c6a2c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 21:09:40 crc kubenswrapper[4754]: I1005 21:09:40.071918 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/1d96e280-8c26-41ea-ab29-64ce5b9c6a2c-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"1d96e280-8c26-41ea-ab29-64ce5b9c6a2c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 21:09:40 crc kubenswrapper[4754]: I1005 21:09:40.071972 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/1d96e280-8c26-41ea-ab29-64ce5b9c6a2c-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"1d96e280-8c26-41ea-ab29-64ce5b9c6a2c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 21:09:40 crc kubenswrapper[4754]: I1005 21:09:40.072015 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/1d96e280-8c26-41ea-ab29-64ce5b9c6a2c-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"1d96e280-8c26-41ea-ab29-64ce5b9c6a2c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 21:09:40 crc kubenswrapper[4754]: I1005 21:09:40.072044 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/1d96e280-8c26-41ea-ab29-64ce5b9c6a2c-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"1d96e280-8c26-41ea-ab29-64ce5b9c6a2c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 21:09:40 crc kubenswrapper[4754]: I1005 21:09:40.072080 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"1d96e280-8c26-41ea-ab29-64ce5b9c6a2c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 21:09:40 crc kubenswrapper[4754]: I1005 21:09:40.072104 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1d96e280-8c26-41ea-ab29-64ce5b9c6a2c-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"1d96e280-8c26-41ea-ab29-64ce5b9c6a2c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 21:09:40 crc kubenswrapper[4754]: I1005 21:09:40.072130 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/1d96e280-8c26-41ea-ab29-64ce5b9c6a2c-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"1d96e280-8c26-41ea-ab29-64ce5b9c6a2c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 21:09:40 crc kubenswrapper[4754]: I1005 21:09:40.072202 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/1d96e280-8c26-41ea-ab29-64ce5b9c6a2c-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"1d96e280-8c26-41ea-ab29-64ce5b9c6a2c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 21:09:40 crc kubenswrapper[4754]: I1005 21:09:40.072228 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/1d96e280-8c26-41ea-ab29-64ce5b9c6a2c-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"1d96e280-8c26-41ea-ab29-64ce5b9c6a2c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 21:09:40 crc kubenswrapper[4754]: I1005 21:09:40.173965 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6ltd4\" (UniqueName: \"kubernetes.io/projected/1d96e280-8c26-41ea-ab29-64ce5b9c6a2c-kube-api-access-6ltd4\") pod \"rabbitmq-cell1-server-0\" (UID: \"1d96e280-8c26-41ea-ab29-64ce5b9c6a2c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 21:09:40 crc kubenswrapper[4754]: I1005 21:09:40.174019 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/1d96e280-8c26-41ea-ab29-64ce5b9c6a2c-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"1d96e280-8c26-41ea-ab29-64ce5b9c6a2c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 21:09:40 crc kubenswrapper[4754]: I1005 21:09:40.174063 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/1d96e280-8c26-41ea-ab29-64ce5b9c6a2c-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"1d96e280-8c26-41ea-ab29-64ce5b9c6a2c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 21:09:40 crc kubenswrapper[4754]: I1005 21:09:40.174090 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/1d96e280-8c26-41ea-ab29-64ce5b9c6a2c-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"1d96e280-8c26-41ea-ab29-64ce5b9c6a2c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 21:09:40 crc kubenswrapper[4754]: I1005 21:09:40.174116 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/1d96e280-8c26-41ea-ab29-64ce5b9c6a2c-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"1d96e280-8c26-41ea-ab29-64ce5b9c6a2c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 21:09:40 crc kubenswrapper[4754]: I1005 21:09:40.174152 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"1d96e280-8c26-41ea-ab29-64ce5b9c6a2c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 21:09:40 crc kubenswrapper[4754]: I1005 21:09:40.174178 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1d96e280-8c26-41ea-ab29-64ce5b9c6a2c-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"1d96e280-8c26-41ea-ab29-64ce5b9c6a2c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 21:09:40 crc kubenswrapper[4754]: I1005 21:09:40.174198 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/1d96e280-8c26-41ea-ab29-64ce5b9c6a2c-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"1d96e280-8c26-41ea-ab29-64ce5b9c6a2c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 21:09:40 crc kubenswrapper[4754]: I1005 21:09:40.174243 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/1d96e280-8c26-41ea-ab29-64ce5b9c6a2c-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"1d96e280-8c26-41ea-ab29-64ce5b9c6a2c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 21:09:40 crc kubenswrapper[4754]: I1005 21:09:40.174269 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/1d96e280-8c26-41ea-ab29-64ce5b9c6a2c-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"1d96e280-8c26-41ea-ab29-64ce5b9c6a2c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 21:09:40 crc kubenswrapper[4754]: I1005 21:09:40.174298 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/1d96e280-8c26-41ea-ab29-64ce5b9c6a2c-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"1d96e280-8c26-41ea-ab29-64ce5b9c6a2c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 21:09:40 crc kubenswrapper[4754]: I1005 21:09:40.174857 4754 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"1d96e280-8c26-41ea-ab29-64ce5b9c6a2c\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/rabbitmq-cell1-server-0" Oct 05 21:09:40 crc kubenswrapper[4754]: I1005 21:09:40.180141 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1d96e280-8c26-41ea-ab29-64ce5b9c6a2c-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"1d96e280-8c26-41ea-ab29-64ce5b9c6a2c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 21:09:40 crc kubenswrapper[4754]: I1005 21:09:40.181818 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/1d96e280-8c26-41ea-ab29-64ce5b9c6a2c-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"1d96e280-8c26-41ea-ab29-64ce5b9c6a2c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 21:09:40 crc kubenswrapper[4754]: I1005 21:09:40.181955 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/1d96e280-8c26-41ea-ab29-64ce5b9c6a2c-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"1d96e280-8c26-41ea-ab29-64ce5b9c6a2c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 21:09:40 crc kubenswrapper[4754]: I1005 21:09:40.182908 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/1d96e280-8c26-41ea-ab29-64ce5b9c6a2c-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"1d96e280-8c26-41ea-ab29-64ce5b9c6a2c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 21:09:40 crc kubenswrapper[4754]: I1005 21:09:40.185523 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/1d96e280-8c26-41ea-ab29-64ce5b9c6a2c-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"1d96e280-8c26-41ea-ab29-64ce5b9c6a2c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 21:09:40 crc kubenswrapper[4754]: I1005 21:09:40.186622 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/1d96e280-8c26-41ea-ab29-64ce5b9c6a2c-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"1d96e280-8c26-41ea-ab29-64ce5b9c6a2c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 21:09:40 crc kubenswrapper[4754]: I1005 21:09:40.189605 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/1d96e280-8c26-41ea-ab29-64ce5b9c6a2c-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"1d96e280-8c26-41ea-ab29-64ce5b9c6a2c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 21:09:40 crc kubenswrapper[4754]: I1005 21:09:40.189809 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/1d96e280-8c26-41ea-ab29-64ce5b9c6a2c-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"1d96e280-8c26-41ea-ab29-64ce5b9c6a2c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 21:09:40 crc kubenswrapper[4754]: I1005 21:09:40.191942 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/1d96e280-8c26-41ea-ab29-64ce5b9c6a2c-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"1d96e280-8c26-41ea-ab29-64ce5b9c6a2c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 21:09:40 crc kubenswrapper[4754]: I1005 21:09:40.213605 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6ltd4\" (UniqueName: \"kubernetes.io/projected/1d96e280-8c26-41ea-ab29-64ce5b9c6a2c-kube-api-access-6ltd4\") pod \"rabbitmq-cell1-server-0\" (UID: \"1d96e280-8c26-41ea-ab29-64ce5b9c6a2c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 21:09:40 crc kubenswrapper[4754]: I1005 21:09:40.229619 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"1d96e280-8c26-41ea-ab29-64ce5b9c6a2c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 21:09:40 crc kubenswrapper[4754]: I1005 21:09:40.370181 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 05 21:09:40 crc kubenswrapper[4754]: I1005 21:09:40.426030 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 05 21:09:40 crc kubenswrapper[4754]: W1005 21:09:40.499128 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod39aff8a0_2905_4ad3_ad2d_5bf6b5de3858.slice/crio-2e640693fa23c07efabf6eb4ac474cd2b47ab78643ef4513ff944ad3ce0b5dc9 WatchSource:0}: Error finding container 2e640693fa23c07efabf6eb4ac474cd2b47ab78643ef4513ff944ad3ce0b5dc9: Status 404 returned error can't find the container with id 2e640693fa23c07efabf6eb4ac474cd2b47ab78643ef4513ff944ad3ce0b5dc9 Oct 05 21:09:41 crc kubenswrapper[4754]: I1005 21:09:41.008332 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"39aff8a0-2905-4ad3-ad2d-5bf6b5de3858","Type":"ContainerStarted","Data":"2e640693fa23c07efabf6eb4ac474cd2b47ab78643ef4513ff944ad3ce0b5dc9"} Oct 05 21:09:41 crc kubenswrapper[4754]: I1005 21:09:41.013664 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-d5rsv" event={"ID":"22f48eb7-09d6-44bc-bc47-ba7026cf4cc3","Type":"ContainerStarted","Data":"d36c72ea046cc25ee3cebcab4fdf3d09ac2f457fe2bf71446d7f8e45626f3854"} Oct 05 21:09:41 crc kubenswrapper[4754]: I1005 21:09:41.074412 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 05 21:09:41 crc kubenswrapper[4754]: I1005 21:09:41.286547 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Oct 05 21:09:41 crc kubenswrapper[4754]: I1005 21:09:41.290213 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 05 21:09:41 crc kubenswrapper[4754]: I1005 21:09:41.357226 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Oct 05 21:09:41 crc kubenswrapper[4754]: I1005 21:09:41.357464 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-5sbzf" Oct 05 21:09:41 crc kubenswrapper[4754]: I1005 21:09:41.357626 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Oct 05 21:09:41 crc kubenswrapper[4754]: I1005 21:09:41.357746 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Oct 05 21:09:41 crc kubenswrapper[4754]: I1005 21:09:41.357852 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Oct 05 21:09:41 crc kubenswrapper[4754]: I1005 21:09:41.372741 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Oct 05 21:09:41 crc kubenswrapper[4754]: I1005 21:09:41.392278 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Oct 05 21:09:41 crc kubenswrapper[4754]: I1005 21:09:41.453601 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-galera-0\" (UID: \"d311946e-91c6-4be6-9d48-1b823ee77607\") " pod="openstack/openstack-galera-0" Oct 05 21:09:41 crc kubenswrapper[4754]: I1005 21:09:41.453744 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/d311946e-91c6-4be6-9d48-1b823ee77607-config-data-default\") pod \"openstack-galera-0\" (UID: \"d311946e-91c6-4be6-9d48-1b823ee77607\") " pod="openstack/openstack-galera-0" Oct 05 21:09:41 crc kubenswrapper[4754]: I1005 21:09:41.453772 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d311946e-91c6-4be6-9d48-1b823ee77607-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"d311946e-91c6-4be6-9d48-1b823ee77607\") " pod="openstack/openstack-galera-0" Oct 05 21:09:41 crc kubenswrapper[4754]: I1005 21:09:41.453839 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/d311946e-91c6-4be6-9d48-1b823ee77607-config-data-generated\") pod \"openstack-galera-0\" (UID: \"d311946e-91c6-4be6-9d48-1b823ee77607\") " pod="openstack/openstack-galera-0" Oct 05 21:09:41 crc kubenswrapper[4754]: I1005 21:09:41.453863 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/d311946e-91c6-4be6-9d48-1b823ee77607-secrets\") pod \"openstack-galera-0\" (UID: \"d311946e-91c6-4be6-9d48-1b823ee77607\") " pod="openstack/openstack-galera-0" Oct 05 21:09:41 crc kubenswrapper[4754]: I1005 21:09:41.453911 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j44cz\" (UniqueName: \"kubernetes.io/projected/d311946e-91c6-4be6-9d48-1b823ee77607-kube-api-access-j44cz\") pod \"openstack-galera-0\" (UID: \"d311946e-91c6-4be6-9d48-1b823ee77607\") " pod="openstack/openstack-galera-0" Oct 05 21:09:41 crc kubenswrapper[4754]: I1005 21:09:41.453939 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/d311946e-91c6-4be6-9d48-1b823ee77607-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"d311946e-91c6-4be6-9d48-1b823ee77607\") " pod="openstack/openstack-galera-0" Oct 05 21:09:41 crc kubenswrapper[4754]: I1005 21:09:41.453957 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d311946e-91c6-4be6-9d48-1b823ee77607-operator-scripts\") pod \"openstack-galera-0\" (UID: \"d311946e-91c6-4be6-9d48-1b823ee77607\") " pod="openstack/openstack-galera-0" Oct 05 21:09:41 crc kubenswrapper[4754]: I1005 21:09:41.454007 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/d311946e-91c6-4be6-9d48-1b823ee77607-kolla-config\") pod \"openstack-galera-0\" (UID: \"d311946e-91c6-4be6-9d48-1b823ee77607\") " pod="openstack/openstack-galera-0" Oct 05 21:09:41 crc kubenswrapper[4754]: I1005 21:09:41.555849 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/d311946e-91c6-4be6-9d48-1b823ee77607-config-data-generated\") pod \"openstack-galera-0\" (UID: \"d311946e-91c6-4be6-9d48-1b823ee77607\") " pod="openstack/openstack-galera-0" Oct 05 21:09:41 crc kubenswrapper[4754]: I1005 21:09:41.555954 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/d311946e-91c6-4be6-9d48-1b823ee77607-secrets\") pod \"openstack-galera-0\" (UID: \"d311946e-91c6-4be6-9d48-1b823ee77607\") " pod="openstack/openstack-galera-0" Oct 05 21:09:41 crc kubenswrapper[4754]: I1005 21:09:41.556010 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j44cz\" (UniqueName: \"kubernetes.io/projected/d311946e-91c6-4be6-9d48-1b823ee77607-kube-api-access-j44cz\") pod \"openstack-galera-0\" (UID: \"d311946e-91c6-4be6-9d48-1b823ee77607\") " pod="openstack/openstack-galera-0" Oct 05 21:09:41 crc kubenswrapper[4754]: I1005 21:09:41.556040 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/d311946e-91c6-4be6-9d48-1b823ee77607-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"d311946e-91c6-4be6-9d48-1b823ee77607\") " pod="openstack/openstack-galera-0" Oct 05 21:09:41 crc kubenswrapper[4754]: I1005 21:09:41.556064 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d311946e-91c6-4be6-9d48-1b823ee77607-operator-scripts\") pod \"openstack-galera-0\" (UID: \"d311946e-91c6-4be6-9d48-1b823ee77607\") " pod="openstack/openstack-galera-0" Oct 05 21:09:41 crc kubenswrapper[4754]: I1005 21:09:41.556100 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/d311946e-91c6-4be6-9d48-1b823ee77607-kolla-config\") pod \"openstack-galera-0\" (UID: \"d311946e-91c6-4be6-9d48-1b823ee77607\") " pod="openstack/openstack-galera-0" Oct 05 21:09:41 crc kubenswrapper[4754]: I1005 21:09:41.556144 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-galera-0\" (UID: \"d311946e-91c6-4be6-9d48-1b823ee77607\") " pod="openstack/openstack-galera-0" Oct 05 21:09:41 crc kubenswrapper[4754]: I1005 21:09:41.556171 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/d311946e-91c6-4be6-9d48-1b823ee77607-config-data-default\") pod \"openstack-galera-0\" (UID: \"d311946e-91c6-4be6-9d48-1b823ee77607\") " pod="openstack/openstack-galera-0" Oct 05 21:09:41 crc kubenswrapper[4754]: I1005 21:09:41.556190 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d311946e-91c6-4be6-9d48-1b823ee77607-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"d311946e-91c6-4be6-9d48-1b823ee77607\") " pod="openstack/openstack-galera-0" Oct 05 21:09:41 crc kubenswrapper[4754]: I1005 21:09:41.557479 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/d311946e-91c6-4be6-9d48-1b823ee77607-config-data-generated\") pod \"openstack-galera-0\" (UID: \"d311946e-91c6-4be6-9d48-1b823ee77607\") " pod="openstack/openstack-galera-0" Oct 05 21:09:41 crc kubenswrapper[4754]: I1005 21:09:41.558007 4754 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-galera-0\" (UID: \"d311946e-91c6-4be6-9d48-1b823ee77607\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/openstack-galera-0" Oct 05 21:09:41 crc kubenswrapper[4754]: I1005 21:09:41.558874 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d311946e-91c6-4be6-9d48-1b823ee77607-operator-scripts\") pod \"openstack-galera-0\" (UID: \"d311946e-91c6-4be6-9d48-1b823ee77607\") " pod="openstack/openstack-galera-0" Oct 05 21:09:41 crc kubenswrapper[4754]: I1005 21:09:41.562791 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/d311946e-91c6-4be6-9d48-1b823ee77607-config-data-default\") pod \"openstack-galera-0\" (UID: \"d311946e-91c6-4be6-9d48-1b823ee77607\") " pod="openstack/openstack-galera-0" Oct 05 21:09:41 crc kubenswrapper[4754]: I1005 21:09:41.563325 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/d311946e-91c6-4be6-9d48-1b823ee77607-kolla-config\") pod \"openstack-galera-0\" (UID: \"d311946e-91c6-4be6-9d48-1b823ee77607\") " pod="openstack/openstack-galera-0" Oct 05 21:09:41 crc kubenswrapper[4754]: I1005 21:09:41.574106 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d311946e-91c6-4be6-9d48-1b823ee77607-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"d311946e-91c6-4be6-9d48-1b823ee77607\") " pod="openstack/openstack-galera-0" Oct 05 21:09:41 crc kubenswrapper[4754]: I1005 21:09:41.577027 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/d311946e-91c6-4be6-9d48-1b823ee77607-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"d311946e-91c6-4be6-9d48-1b823ee77607\") " pod="openstack/openstack-galera-0" Oct 05 21:09:41 crc kubenswrapper[4754]: I1005 21:09:41.580202 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j44cz\" (UniqueName: \"kubernetes.io/projected/d311946e-91c6-4be6-9d48-1b823ee77607-kube-api-access-j44cz\") pod \"openstack-galera-0\" (UID: \"d311946e-91c6-4be6-9d48-1b823ee77607\") " pod="openstack/openstack-galera-0" Oct 05 21:09:41 crc kubenswrapper[4754]: I1005 21:09:41.580821 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/d311946e-91c6-4be6-9d48-1b823ee77607-secrets\") pod \"openstack-galera-0\" (UID: \"d311946e-91c6-4be6-9d48-1b823ee77607\") " pod="openstack/openstack-galera-0" Oct 05 21:09:41 crc kubenswrapper[4754]: I1005 21:09:41.597209 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-galera-0\" (UID: \"d311946e-91c6-4be6-9d48-1b823ee77607\") " pod="openstack/openstack-galera-0" Oct 05 21:09:41 crc kubenswrapper[4754]: I1005 21:09:41.711252 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 05 21:09:42 crc kubenswrapper[4754]: I1005 21:09:42.035760 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"1d96e280-8c26-41ea-ab29-64ce5b9c6a2c","Type":"ContainerStarted","Data":"5709134c4abd234ef35cfa828d09fdbab849e2251253f921d5bf35df040753b0"} Oct 05 21:09:42 crc kubenswrapper[4754]: I1005 21:09:42.483341 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 05 21:09:42 crc kubenswrapper[4754]: I1005 21:09:42.484969 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 05 21:09:42 crc kubenswrapper[4754]: I1005 21:09:42.490913 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-x9d2n" Oct 05 21:09:42 crc kubenswrapper[4754]: I1005 21:09:42.491197 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Oct 05 21:09:42 crc kubenswrapper[4754]: I1005 21:09:42.491439 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Oct 05 21:09:42 crc kubenswrapper[4754]: I1005 21:09:42.495812 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Oct 05 21:09:42 crc kubenswrapper[4754]: I1005 21:09:42.503326 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 05 21:09:42 crc kubenswrapper[4754]: I1005 21:09:42.586606 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/d97f6b34-ef35-4b21-81eb-9e7009945fa6-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"d97f6b34-ef35-4b21-81eb-9e7009945fa6\") " pod="openstack/openstack-cell1-galera-0" Oct 05 21:09:42 crc kubenswrapper[4754]: I1005 21:09:42.586686 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d97f6b34-ef35-4b21-81eb-9e7009945fa6-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"d97f6b34-ef35-4b21-81eb-9e7009945fa6\") " pod="openstack/openstack-cell1-galera-0" Oct 05 21:09:42 crc kubenswrapper[4754]: I1005 21:09:42.586706 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d97f6b34-ef35-4b21-81eb-9e7009945fa6-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"d97f6b34-ef35-4b21-81eb-9e7009945fa6\") " pod="openstack/openstack-cell1-galera-0" Oct 05 21:09:42 crc kubenswrapper[4754]: I1005 21:09:42.586726 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-cell1-galera-0\" (UID: \"d97f6b34-ef35-4b21-81eb-9e7009945fa6\") " pod="openstack/openstack-cell1-galera-0" Oct 05 21:09:42 crc kubenswrapper[4754]: I1005 21:09:42.586758 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/d97f6b34-ef35-4b21-81eb-9e7009945fa6-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"d97f6b34-ef35-4b21-81eb-9e7009945fa6\") " pod="openstack/openstack-cell1-galera-0" Oct 05 21:09:42 crc kubenswrapper[4754]: I1005 21:09:42.586784 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/d97f6b34-ef35-4b21-81eb-9e7009945fa6-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"d97f6b34-ef35-4b21-81eb-9e7009945fa6\") " pod="openstack/openstack-cell1-galera-0" Oct 05 21:09:42 crc kubenswrapper[4754]: I1005 21:09:42.586829 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/d97f6b34-ef35-4b21-81eb-9e7009945fa6-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"d97f6b34-ef35-4b21-81eb-9e7009945fa6\") " pod="openstack/openstack-cell1-galera-0" Oct 05 21:09:42 crc kubenswrapper[4754]: I1005 21:09:42.586870 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7jn87\" (UniqueName: \"kubernetes.io/projected/d97f6b34-ef35-4b21-81eb-9e7009945fa6-kube-api-access-7jn87\") pod \"openstack-cell1-galera-0\" (UID: \"d97f6b34-ef35-4b21-81eb-9e7009945fa6\") " pod="openstack/openstack-cell1-galera-0" Oct 05 21:09:42 crc kubenswrapper[4754]: I1005 21:09:42.586906 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/d97f6b34-ef35-4b21-81eb-9e7009945fa6-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"d97f6b34-ef35-4b21-81eb-9e7009945fa6\") " pod="openstack/openstack-cell1-galera-0" Oct 05 21:09:42 crc kubenswrapper[4754]: I1005 21:09:42.608312 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Oct 05 21:09:42 crc kubenswrapper[4754]: I1005 21:09:42.691042 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7jn87\" (UniqueName: \"kubernetes.io/projected/d97f6b34-ef35-4b21-81eb-9e7009945fa6-kube-api-access-7jn87\") pod \"openstack-cell1-galera-0\" (UID: \"d97f6b34-ef35-4b21-81eb-9e7009945fa6\") " pod="openstack/openstack-cell1-galera-0" Oct 05 21:09:42 crc kubenswrapper[4754]: I1005 21:09:42.691123 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/d97f6b34-ef35-4b21-81eb-9e7009945fa6-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"d97f6b34-ef35-4b21-81eb-9e7009945fa6\") " pod="openstack/openstack-cell1-galera-0" Oct 05 21:09:42 crc kubenswrapper[4754]: I1005 21:09:42.691236 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/d97f6b34-ef35-4b21-81eb-9e7009945fa6-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"d97f6b34-ef35-4b21-81eb-9e7009945fa6\") " pod="openstack/openstack-cell1-galera-0" Oct 05 21:09:42 crc kubenswrapper[4754]: I1005 21:09:42.691316 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d97f6b34-ef35-4b21-81eb-9e7009945fa6-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"d97f6b34-ef35-4b21-81eb-9e7009945fa6\") " pod="openstack/openstack-cell1-galera-0" Oct 05 21:09:42 crc kubenswrapper[4754]: I1005 21:09:42.691340 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d97f6b34-ef35-4b21-81eb-9e7009945fa6-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"d97f6b34-ef35-4b21-81eb-9e7009945fa6\") " pod="openstack/openstack-cell1-galera-0" Oct 05 21:09:42 crc kubenswrapper[4754]: I1005 21:09:42.691375 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-cell1-galera-0\" (UID: \"d97f6b34-ef35-4b21-81eb-9e7009945fa6\") " pod="openstack/openstack-cell1-galera-0" Oct 05 21:09:42 crc kubenswrapper[4754]: I1005 21:09:42.691407 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/d97f6b34-ef35-4b21-81eb-9e7009945fa6-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"d97f6b34-ef35-4b21-81eb-9e7009945fa6\") " pod="openstack/openstack-cell1-galera-0" Oct 05 21:09:42 crc kubenswrapper[4754]: I1005 21:09:42.691450 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/d97f6b34-ef35-4b21-81eb-9e7009945fa6-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"d97f6b34-ef35-4b21-81eb-9e7009945fa6\") " pod="openstack/openstack-cell1-galera-0" Oct 05 21:09:42 crc kubenswrapper[4754]: I1005 21:09:42.691519 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/d97f6b34-ef35-4b21-81eb-9e7009945fa6-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"d97f6b34-ef35-4b21-81eb-9e7009945fa6\") " pod="openstack/openstack-cell1-galera-0" Oct 05 21:09:42 crc kubenswrapper[4754]: I1005 21:09:42.692510 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/d97f6b34-ef35-4b21-81eb-9e7009945fa6-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"d97f6b34-ef35-4b21-81eb-9e7009945fa6\") " pod="openstack/openstack-cell1-galera-0" Oct 05 21:09:42 crc kubenswrapper[4754]: I1005 21:09:42.696863 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d97f6b34-ef35-4b21-81eb-9e7009945fa6-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"d97f6b34-ef35-4b21-81eb-9e7009945fa6\") " pod="openstack/openstack-cell1-galera-0" Oct 05 21:09:42 crc kubenswrapper[4754]: I1005 21:09:42.703286 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/d97f6b34-ef35-4b21-81eb-9e7009945fa6-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"d97f6b34-ef35-4b21-81eb-9e7009945fa6\") " pod="openstack/openstack-cell1-galera-0" Oct 05 21:09:42 crc kubenswrapper[4754]: I1005 21:09:42.703654 4754 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-cell1-galera-0\" (UID: \"d97f6b34-ef35-4b21-81eb-9e7009945fa6\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/openstack-cell1-galera-0" Oct 05 21:09:42 crc kubenswrapper[4754]: I1005 21:09:42.708127 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/d97f6b34-ef35-4b21-81eb-9e7009945fa6-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"d97f6b34-ef35-4b21-81eb-9e7009945fa6\") " pod="openstack/openstack-cell1-galera-0" Oct 05 21:09:42 crc kubenswrapper[4754]: I1005 21:09:42.724677 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/d97f6b34-ef35-4b21-81eb-9e7009945fa6-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"d97f6b34-ef35-4b21-81eb-9e7009945fa6\") " pod="openstack/openstack-cell1-galera-0" Oct 05 21:09:42 crc kubenswrapper[4754]: I1005 21:09:42.733354 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d97f6b34-ef35-4b21-81eb-9e7009945fa6-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"d97f6b34-ef35-4b21-81eb-9e7009945fa6\") " pod="openstack/openstack-cell1-galera-0" Oct 05 21:09:42 crc kubenswrapper[4754]: I1005 21:09:42.734207 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7jn87\" (UniqueName: \"kubernetes.io/projected/d97f6b34-ef35-4b21-81eb-9e7009945fa6-kube-api-access-7jn87\") pod \"openstack-cell1-galera-0\" (UID: \"d97f6b34-ef35-4b21-81eb-9e7009945fa6\") " pod="openstack/openstack-cell1-galera-0" Oct 05 21:09:42 crc kubenswrapper[4754]: I1005 21:09:42.746891 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-cell1-galera-0\" (UID: \"d97f6b34-ef35-4b21-81eb-9e7009945fa6\") " pod="openstack/openstack-cell1-galera-0" Oct 05 21:09:42 crc kubenswrapper[4754]: I1005 21:09:42.751149 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/d97f6b34-ef35-4b21-81eb-9e7009945fa6-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"d97f6b34-ef35-4b21-81eb-9e7009945fa6\") " pod="openstack/openstack-cell1-galera-0" Oct 05 21:09:42 crc kubenswrapper[4754]: I1005 21:09:42.879293 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 05 21:09:42 crc kubenswrapper[4754]: I1005 21:09:42.989020 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Oct 05 21:09:42 crc kubenswrapper[4754]: I1005 21:09:42.992505 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 05 21:09:43 crc kubenswrapper[4754]: I1005 21:09:43.009205 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Oct 05 21:09:43 crc kubenswrapper[4754]: I1005 21:09:43.017457 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-5pfj6" Oct 05 21:09:43 crc kubenswrapper[4754]: I1005 21:09:43.017766 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Oct 05 21:09:43 crc kubenswrapper[4754]: I1005 21:09:43.116586 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/bdbca489-8f9b-420e-a306-65575175af99-memcached-tls-certs\") pod \"memcached-0\" (UID: \"bdbca489-8f9b-420e-a306-65575175af99\") " pod="openstack/memcached-0" Oct 05 21:09:43 crc kubenswrapper[4754]: I1005 21:09:43.116637 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bdbca489-8f9b-420e-a306-65575175af99-combined-ca-bundle\") pod \"memcached-0\" (UID: \"bdbca489-8f9b-420e-a306-65575175af99\") " pod="openstack/memcached-0" Oct 05 21:09:43 crc kubenswrapper[4754]: I1005 21:09:43.116711 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jnlj8\" (UniqueName: \"kubernetes.io/projected/bdbca489-8f9b-420e-a306-65575175af99-kube-api-access-jnlj8\") pod \"memcached-0\" (UID: \"bdbca489-8f9b-420e-a306-65575175af99\") " pod="openstack/memcached-0" Oct 05 21:09:43 crc kubenswrapper[4754]: I1005 21:09:43.116742 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bdbca489-8f9b-420e-a306-65575175af99-config-data\") pod \"memcached-0\" (UID: \"bdbca489-8f9b-420e-a306-65575175af99\") " pod="openstack/memcached-0" Oct 05 21:09:43 crc kubenswrapper[4754]: I1005 21:09:43.116783 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/bdbca489-8f9b-420e-a306-65575175af99-kolla-config\") pod \"memcached-0\" (UID: \"bdbca489-8f9b-420e-a306-65575175af99\") " pod="openstack/memcached-0" Oct 05 21:09:43 crc kubenswrapper[4754]: I1005 21:09:43.172588 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Oct 05 21:09:43 crc kubenswrapper[4754]: I1005 21:09:43.209694 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"d311946e-91c6-4be6-9d48-1b823ee77607","Type":"ContainerStarted","Data":"c2975dedd29388f0dd47a06ffbb4977a675d6d55e5ca64bd3cf2ed79787b22ab"} Oct 05 21:09:43 crc kubenswrapper[4754]: I1005 21:09:43.219173 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jnlj8\" (UniqueName: \"kubernetes.io/projected/bdbca489-8f9b-420e-a306-65575175af99-kube-api-access-jnlj8\") pod \"memcached-0\" (UID: \"bdbca489-8f9b-420e-a306-65575175af99\") " pod="openstack/memcached-0" Oct 05 21:09:43 crc kubenswrapper[4754]: I1005 21:09:43.219243 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bdbca489-8f9b-420e-a306-65575175af99-config-data\") pod \"memcached-0\" (UID: \"bdbca489-8f9b-420e-a306-65575175af99\") " pod="openstack/memcached-0" Oct 05 21:09:43 crc kubenswrapper[4754]: I1005 21:09:43.219272 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/bdbca489-8f9b-420e-a306-65575175af99-kolla-config\") pod \"memcached-0\" (UID: \"bdbca489-8f9b-420e-a306-65575175af99\") " pod="openstack/memcached-0" Oct 05 21:09:43 crc kubenswrapper[4754]: I1005 21:09:43.219336 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/bdbca489-8f9b-420e-a306-65575175af99-memcached-tls-certs\") pod \"memcached-0\" (UID: \"bdbca489-8f9b-420e-a306-65575175af99\") " pod="openstack/memcached-0" Oct 05 21:09:43 crc kubenswrapper[4754]: I1005 21:09:43.219358 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bdbca489-8f9b-420e-a306-65575175af99-combined-ca-bundle\") pod \"memcached-0\" (UID: \"bdbca489-8f9b-420e-a306-65575175af99\") " pod="openstack/memcached-0" Oct 05 21:09:43 crc kubenswrapper[4754]: I1005 21:09:43.225397 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bdbca489-8f9b-420e-a306-65575175af99-config-data\") pod \"memcached-0\" (UID: \"bdbca489-8f9b-420e-a306-65575175af99\") " pod="openstack/memcached-0" Oct 05 21:09:43 crc kubenswrapper[4754]: I1005 21:09:43.225738 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/bdbca489-8f9b-420e-a306-65575175af99-kolla-config\") pod \"memcached-0\" (UID: \"bdbca489-8f9b-420e-a306-65575175af99\") " pod="openstack/memcached-0" Oct 05 21:09:43 crc kubenswrapper[4754]: I1005 21:09:43.230297 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bdbca489-8f9b-420e-a306-65575175af99-combined-ca-bundle\") pod \"memcached-0\" (UID: \"bdbca489-8f9b-420e-a306-65575175af99\") " pod="openstack/memcached-0" Oct 05 21:09:43 crc kubenswrapper[4754]: I1005 21:09:43.235677 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/bdbca489-8f9b-420e-a306-65575175af99-memcached-tls-certs\") pod \"memcached-0\" (UID: \"bdbca489-8f9b-420e-a306-65575175af99\") " pod="openstack/memcached-0" Oct 05 21:09:43 crc kubenswrapper[4754]: I1005 21:09:43.253526 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jnlj8\" (UniqueName: \"kubernetes.io/projected/bdbca489-8f9b-420e-a306-65575175af99-kube-api-access-jnlj8\") pod \"memcached-0\" (UID: \"bdbca489-8f9b-420e-a306-65575175af99\") " pod="openstack/memcached-0" Oct 05 21:09:43 crc kubenswrapper[4754]: I1005 21:09:43.418210 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 05 21:09:44 crc kubenswrapper[4754]: I1005 21:09:43.999909 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 05 21:09:44 crc kubenswrapper[4754]: I1005 21:09:44.305431 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"d97f6b34-ef35-4b21-81eb-9e7009945fa6","Type":"ContainerStarted","Data":"661dcdbdd2c6712a8412e427b8f5b3a4c060cda09e2f83bf6bf4e2b4a1297dc0"} Oct 05 21:09:44 crc kubenswrapper[4754]: I1005 21:09:44.519848 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Oct 05 21:09:44 crc kubenswrapper[4754]: I1005 21:09:44.523850 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 05 21:09:44 crc kubenswrapper[4754]: I1005 21:09:44.527183 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-2w7df" Oct 05 21:09:44 crc kubenswrapper[4754]: I1005 21:09:44.532686 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 05 21:09:44 crc kubenswrapper[4754]: I1005 21:09:44.607631 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Oct 05 21:09:44 crc kubenswrapper[4754]: I1005 21:09:44.658478 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qh7bz\" (UniqueName: \"kubernetes.io/projected/280f4e00-063d-486e-bd58-f4b246f54814-kube-api-access-qh7bz\") pod \"kube-state-metrics-0\" (UID: \"280f4e00-063d-486e-bd58-f4b246f54814\") " pod="openstack/kube-state-metrics-0" Oct 05 21:09:44 crc kubenswrapper[4754]: I1005 21:09:44.761657 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qh7bz\" (UniqueName: \"kubernetes.io/projected/280f4e00-063d-486e-bd58-f4b246f54814-kube-api-access-qh7bz\") pod \"kube-state-metrics-0\" (UID: \"280f4e00-063d-486e-bd58-f4b246f54814\") " pod="openstack/kube-state-metrics-0" Oct 05 21:09:44 crc kubenswrapper[4754]: I1005 21:09:44.787025 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qh7bz\" (UniqueName: \"kubernetes.io/projected/280f4e00-063d-486e-bd58-f4b246f54814-kube-api-access-qh7bz\") pod \"kube-state-metrics-0\" (UID: \"280f4e00-063d-486e-bd58-f4b246f54814\") " pod="openstack/kube-state-metrics-0" Oct 05 21:09:44 crc kubenswrapper[4754]: I1005 21:09:44.877488 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 05 21:09:45 crc kubenswrapper[4754]: I1005 21:09:45.371894 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"bdbca489-8f9b-420e-a306-65575175af99","Type":"ContainerStarted","Data":"e508969967ff7d575250685355915a79f05066f19e045b50ff90cc31fa659dba"} Oct 05 21:09:45 crc kubenswrapper[4754]: I1005 21:09:45.657892 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 05 21:09:45 crc kubenswrapper[4754]: W1005 21:09:45.672641 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod280f4e00_063d_486e_bd58_f4b246f54814.slice/crio-9120a0bee39fd16a94bd3da4f4382072be682ad13213dc94566d9b585afb4750 WatchSource:0}: Error finding container 9120a0bee39fd16a94bd3da4f4382072be682ad13213dc94566d9b585afb4750: Status 404 returned error can't find the container with id 9120a0bee39fd16a94bd3da4f4382072be682ad13213dc94566d9b585afb4750 Oct 05 21:09:46 crc kubenswrapper[4754]: I1005 21:09:46.422484 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"280f4e00-063d-486e-bd58-f4b246f54814","Type":"ContainerStarted","Data":"9120a0bee39fd16a94bd3da4f4382072be682ad13213dc94566d9b585afb4750"} Oct 05 21:09:48 crc kubenswrapper[4754]: I1005 21:09:48.357529 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-pp728"] Oct 05 21:09:48 crc kubenswrapper[4754]: I1005 21:09:48.358823 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-pp728" Oct 05 21:09:48 crc kubenswrapper[4754]: I1005 21:09:48.365513 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-jxcsx" Oct 05 21:09:48 crc kubenswrapper[4754]: I1005 21:09:48.365743 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Oct 05 21:09:48 crc kubenswrapper[4754]: I1005 21:09:48.365879 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Oct 05 21:09:48 crc kubenswrapper[4754]: I1005 21:09:48.399315 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-2krf2"] Oct 05 21:09:48 crc kubenswrapper[4754]: I1005 21:09:48.410227 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-pp728"] Oct 05 21:09:48 crc kubenswrapper[4754]: I1005 21:09:48.410370 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-2krf2" Oct 05 21:09:48 crc kubenswrapper[4754]: I1005 21:09:48.453842 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-2krf2"] Oct 05 21:09:48 crc kubenswrapper[4754]: I1005 21:09:48.482626 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/e3785031-57bd-44ab-b30e-b609cca6dcaf-ovn-controller-tls-certs\") pod \"ovn-controller-pp728\" (UID: \"e3785031-57bd-44ab-b30e-b609cca6dcaf\") " pod="openstack/ovn-controller-pp728" Oct 05 21:09:48 crc kubenswrapper[4754]: I1005 21:09:48.482692 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m2glw\" (UniqueName: \"kubernetes.io/projected/e3785031-57bd-44ab-b30e-b609cca6dcaf-kube-api-access-m2glw\") pod \"ovn-controller-pp728\" (UID: \"e3785031-57bd-44ab-b30e-b609cca6dcaf\") " pod="openstack/ovn-controller-pp728" Oct 05 21:09:48 crc kubenswrapper[4754]: I1005 21:09:48.482731 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e3785031-57bd-44ab-b30e-b609cca6dcaf-var-run\") pod \"ovn-controller-pp728\" (UID: \"e3785031-57bd-44ab-b30e-b609cca6dcaf\") " pod="openstack/ovn-controller-pp728" Oct 05 21:09:48 crc kubenswrapper[4754]: I1005 21:09:48.482781 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e3785031-57bd-44ab-b30e-b609cca6dcaf-scripts\") pod \"ovn-controller-pp728\" (UID: \"e3785031-57bd-44ab-b30e-b609cca6dcaf\") " pod="openstack/ovn-controller-pp728" Oct 05 21:09:48 crc kubenswrapper[4754]: I1005 21:09:48.482807 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/e3785031-57bd-44ab-b30e-b609cca6dcaf-var-run-ovn\") pod \"ovn-controller-pp728\" (UID: \"e3785031-57bd-44ab-b30e-b609cca6dcaf\") " pod="openstack/ovn-controller-pp728" Oct 05 21:09:48 crc kubenswrapper[4754]: I1005 21:09:48.482879 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/e3785031-57bd-44ab-b30e-b609cca6dcaf-var-log-ovn\") pod \"ovn-controller-pp728\" (UID: \"e3785031-57bd-44ab-b30e-b609cca6dcaf\") " pod="openstack/ovn-controller-pp728" Oct 05 21:09:48 crc kubenswrapper[4754]: I1005 21:09:48.482900 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3785031-57bd-44ab-b30e-b609cca6dcaf-combined-ca-bundle\") pod \"ovn-controller-pp728\" (UID: \"e3785031-57bd-44ab-b30e-b609cca6dcaf\") " pod="openstack/ovn-controller-pp728" Oct 05 21:09:48 crc kubenswrapper[4754]: I1005 21:09:48.596234 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e3785031-57bd-44ab-b30e-b609cca6dcaf-scripts\") pod \"ovn-controller-pp728\" (UID: \"e3785031-57bd-44ab-b30e-b609cca6dcaf\") " pod="openstack/ovn-controller-pp728" Oct 05 21:09:48 crc kubenswrapper[4754]: I1005 21:09:48.596290 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/1283a11a-7581-4146-87db-b62d6d5d6dc9-var-log\") pod \"ovn-controller-ovs-2krf2\" (UID: \"1283a11a-7581-4146-87db-b62d6d5d6dc9\") " pod="openstack/ovn-controller-ovs-2krf2" Oct 05 21:09:48 crc kubenswrapper[4754]: I1005 21:09:48.596327 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/e3785031-57bd-44ab-b30e-b609cca6dcaf-var-run-ovn\") pod \"ovn-controller-pp728\" (UID: \"e3785031-57bd-44ab-b30e-b609cca6dcaf\") " pod="openstack/ovn-controller-pp728" Oct 05 21:09:48 crc kubenswrapper[4754]: I1005 21:09:48.596377 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5h75z\" (UniqueName: \"kubernetes.io/projected/1283a11a-7581-4146-87db-b62d6d5d6dc9-kube-api-access-5h75z\") pod \"ovn-controller-ovs-2krf2\" (UID: \"1283a11a-7581-4146-87db-b62d6d5d6dc9\") " pod="openstack/ovn-controller-ovs-2krf2" Oct 05 21:09:48 crc kubenswrapper[4754]: I1005 21:09:48.596433 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/1283a11a-7581-4146-87db-b62d6d5d6dc9-var-lib\") pod \"ovn-controller-ovs-2krf2\" (UID: \"1283a11a-7581-4146-87db-b62d6d5d6dc9\") " pod="openstack/ovn-controller-ovs-2krf2" Oct 05 21:09:48 crc kubenswrapper[4754]: I1005 21:09:48.596471 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/e3785031-57bd-44ab-b30e-b609cca6dcaf-var-log-ovn\") pod \"ovn-controller-pp728\" (UID: \"e3785031-57bd-44ab-b30e-b609cca6dcaf\") " pod="openstack/ovn-controller-pp728" Oct 05 21:09:48 crc kubenswrapper[4754]: I1005 21:09:48.596503 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/1283a11a-7581-4146-87db-b62d6d5d6dc9-var-run\") pod \"ovn-controller-ovs-2krf2\" (UID: \"1283a11a-7581-4146-87db-b62d6d5d6dc9\") " pod="openstack/ovn-controller-ovs-2krf2" Oct 05 21:09:48 crc kubenswrapper[4754]: I1005 21:09:48.596521 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3785031-57bd-44ab-b30e-b609cca6dcaf-combined-ca-bundle\") pod \"ovn-controller-pp728\" (UID: \"e3785031-57bd-44ab-b30e-b609cca6dcaf\") " pod="openstack/ovn-controller-pp728" Oct 05 21:09:48 crc kubenswrapper[4754]: I1005 21:09:48.596550 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1283a11a-7581-4146-87db-b62d6d5d6dc9-scripts\") pod \"ovn-controller-ovs-2krf2\" (UID: \"1283a11a-7581-4146-87db-b62d6d5d6dc9\") " pod="openstack/ovn-controller-ovs-2krf2" Oct 05 21:09:48 crc kubenswrapper[4754]: I1005 21:09:48.596574 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/e3785031-57bd-44ab-b30e-b609cca6dcaf-ovn-controller-tls-certs\") pod \"ovn-controller-pp728\" (UID: \"e3785031-57bd-44ab-b30e-b609cca6dcaf\") " pod="openstack/ovn-controller-pp728" Oct 05 21:09:48 crc kubenswrapper[4754]: I1005 21:09:48.596624 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m2glw\" (UniqueName: \"kubernetes.io/projected/e3785031-57bd-44ab-b30e-b609cca6dcaf-kube-api-access-m2glw\") pod \"ovn-controller-pp728\" (UID: \"e3785031-57bd-44ab-b30e-b609cca6dcaf\") " pod="openstack/ovn-controller-pp728" Oct 05 21:09:48 crc kubenswrapper[4754]: I1005 21:09:48.596659 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e3785031-57bd-44ab-b30e-b609cca6dcaf-var-run\") pod \"ovn-controller-pp728\" (UID: \"e3785031-57bd-44ab-b30e-b609cca6dcaf\") " pod="openstack/ovn-controller-pp728" Oct 05 21:09:48 crc kubenswrapper[4754]: I1005 21:09:48.596707 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/1283a11a-7581-4146-87db-b62d6d5d6dc9-etc-ovs\") pod \"ovn-controller-ovs-2krf2\" (UID: \"1283a11a-7581-4146-87db-b62d6d5d6dc9\") " pod="openstack/ovn-controller-ovs-2krf2" Oct 05 21:09:48 crc kubenswrapper[4754]: I1005 21:09:48.599252 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e3785031-57bd-44ab-b30e-b609cca6dcaf-scripts\") pod \"ovn-controller-pp728\" (UID: \"e3785031-57bd-44ab-b30e-b609cca6dcaf\") " pod="openstack/ovn-controller-pp728" Oct 05 21:09:48 crc kubenswrapper[4754]: I1005 21:09:48.599796 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/e3785031-57bd-44ab-b30e-b609cca6dcaf-var-run-ovn\") pod \"ovn-controller-pp728\" (UID: \"e3785031-57bd-44ab-b30e-b609cca6dcaf\") " pod="openstack/ovn-controller-pp728" Oct 05 21:09:48 crc kubenswrapper[4754]: I1005 21:09:48.600085 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/e3785031-57bd-44ab-b30e-b609cca6dcaf-var-log-ovn\") pod \"ovn-controller-pp728\" (UID: \"e3785031-57bd-44ab-b30e-b609cca6dcaf\") " pod="openstack/ovn-controller-pp728" Oct 05 21:09:48 crc kubenswrapper[4754]: I1005 21:09:48.600406 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e3785031-57bd-44ab-b30e-b609cca6dcaf-var-run\") pod \"ovn-controller-pp728\" (UID: \"e3785031-57bd-44ab-b30e-b609cca6dcaf\") " pod="openstack/ovn-controller-pp728" Oct 05 21:09:48 crc kubenswrapper[4754]: I1005 21:09:48.605863 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/e3785031-57bd-44ab-b30e-b609cca6dcaf-ovn-controller-tls-certs\") pod \"ovn-controller-pp728\" (UID: \"e3785031-57bd-44ab-b30e-b609cca6dcaf\") " pod="openstack/ovn-controller-pp728" Oct 05 21:09:48 crc kubenswrapper[4754]: I1005 21:09:48.618627 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3785031-57bd-44ab-b30e-b609cca6dcaf-combined-ca-bundle\") pod \"ovn-controller-pp728\" (UID: \"e3785031-57bd-44ab-b30e-b609cca6dcaf\") " pod="openstack/ovn-controller-pp728" Oct 05 21:09:48 crc kubenswrapper[4754]: I1005 21:09:48.634429 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m2glw\" (UniqueName: \"kubernetes.io/projected/e3785031-57bd-44ab-b30e-b609cca6dcaf-kube-api-access-m2glw\") pod \"ovn-controller-pp728\" (UID: \"e3785031-57bd-44ab-b30e-b609cca6dcaf\") " pod="openstack/ovn-controller-pp728" Oct 05 21:09:48 crc kubenswrapper[4754]: I1005 21:09:48.662575 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 05 21:09:48 crc kubenswrapper[4754]: I1005 21:09:48.664278 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 05 21:09:48 crc kubenswrapper[4754]: I1005 21:09:48.675048 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Oct 05 21:09:48 crc kubenswrapper[4754]: I1005 21:09:48.675377 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Oct 05 21:09:48 crc kubenswrapper[4754]: I1005 21:09:48.675522 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Oct 05 21:09:48 crc kubenswrapper[4754]: I1005 21:09:48.676597 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-snwpm" Oct 05 21:09:48 crc kubenswrapper[4754]: I1005 21:09:48.677218 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Oct 05 21:09:48 crc kubenswrapper[4754]: I1005 21:09:48.696596 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 05 21:09:48 crc kubenswrapper[4754]: I1005 21:09:48.698353 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/1283a11a-7581-4146-87db-b62d6d5d6dc9-etc-ovs\") pod \"ovn-controller-ovs-2krf2\" (UID: \"1283a11a-7581-4146-87db-b62d6d5d6dc9\") " pod="openstack/ovn-controller-ovs-2krf2" Oct 05 21:09:48 crc kubenswrapper[4754]: I1005 21:09:48.698410 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/1283a11a-7581-4146-87db-b62d6d5d6dc9-var-log\") pod \"ovn-controller-ovs-2krf2\" (UID: \"1283a11a-7581-4146-87db-b62d6d5d6dc9\") " pod="openstack/ovn-controller-ovs-2krf2" Oct 05 21:09:48 crc kubenswrapper[4754]: I1005 21:09:48.698450 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5h75z\" (UniqueName: \"kubernetes.io/projected/1283a11a-7581-4146-87db-b62d6d5d6dc9-kube-api-access-5h75z\") pod \"ovn-controller-ovs-2krf2\" (UID: \"1283a11a-7581-4146-87db-b62d6d5d6dc9\") " pod="openstack/ovn-controller-ovs-2krf2" Oct 05 21:09:48 crc kubenswrapper[4754]: I1005 21:09:48.698501 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/1283a11a-7581-4146-87db-b62d6d5d6dc9-var-lib\") pod \"ovn-controller-ovs-2krf2\" (UID: \"1283a11a-7581-4146-87db-b62d6d5d6dc9\") " pod="openstack/ovn-controller-ovs-2krf2" Oct 05 21:09:48 crc kubenswrapper[4754]: I1005 21:09:48.698528 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/1283a11a-7581-4146-87db-b62d6d5d6dc9-var-run\") pod \"ovn-controller-ovs-2krf2\" (UID: \"1283a11a-7581-4146-87db-b62d6d5d6dc9\") " pod="openstack/ovn-controller-ovs-2krf2" Oct 05 21:09:48 crc kubenswrapper[4754]: I1005 21:09:48.698554 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1283a11a-7581-4146-87db-b62d6d5d6dc9-scripts\") pod \"ovn-controller-ovs-2krf2\" (UID: \"1283a11a-7581-4146-87db-b62d6d5d6dc9\") " pod="openstack/ovn-controller-ovs-2krf2" Oct 05 21:09:48 crc kubenswrapper[4754]: I1005 21:09:48.699082 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/1283a11a-7581-4146-87db-b62d6d5d6dc9-etc-ovs\") pod \"ovn-controller-ovs-2krf2\" (UID: \"1283a11a-7581-4146-87db-b62d6d5d6dc9\") " pod="openstack/ovn-controller-ovs-2krf2" Oct 05 21:09:48 crc kubenswrapper[4754]: I1005 21:09:48.700212 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/1283a11a-7581-4146-87db-b62d6d5d6dc9-var-log\") pod \"ovn-controller-ovs-2krf2\" (UID: \"1283a11a-7581-4146-87db-b62d6d5d6dc9\") " pod="openstack/ovn-controller-ovs-2krf2" Oct 05 21:09:48 crc kubenswrapper[4754]: I1005 21:09:48.700442 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/1283a11a-7581-4146-87db-b62d6d5d6dc9-var-lib\") pod \"ovn-controller-ovs-2krf2\" (UID: \"1283a11a-7581-4146-87db-b62d6d5d6dc9\") " pod="openstack/ovn-controller-ovs-2krf2" Oct 05 21:09:48 crc kubenswrapper[4754]: I1005 21:09:48.700543 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/1283a11a-7581-4146-87db-b62d6d5d6dc9-var-run\") pod \"ovn-controller-ovs-2krf2\" (UID: \"1283a11a-7581-4146-87db-b62d6d5d6dc9\") " pod="openstack/ovn-controller-ovs-2krf2" Oct 05 21:09:48 crc kubenswrapper[4754]: I1005 21:09:48.709444 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1283a11a-7581-4146-87db-b62d6d5d6dc9-scripts\") pod \"ovn-controller-ovs-2krf2\" (UID: \"1283a11a-7581-4146-87db-b62d6d5d6dc9\") " pod="openstack/ovn-controller-ovs-2krf2" Oct 05 21:09:48 crc kubenswrapper[4754]: I1005 21:09:48.709857 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-pp728" Oct 05 21:09:48 crc kubenswrapper[4754]: I1005 21:09:48.740067 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5h75z\" (UniqueName: \"kubernetes.io/projected/1283a11a-7581-4146-87db-b62d6d5d6dc9-kube-api-access-5h75z\") pod \"ovn-controller-ovs-2krf2\" (UID: \"1283a11a-7581-4146-87db-b62d6d5d6dc9\") " pod="openstack/ovn-controller-ovs-2krf2" Oct 05 21:09:48 crc kubenswrapper[4754]: I1005 21:09:48.753223 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-2krf2" Oct 05 21:09:48 crc kubenswrapper[4754]: I1005 21:09:48.799625 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a5fdaf77-8a4c-4267-a32e-22a1c6be8f97-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"a5fdaf77-8a4c-4267-a32e-22a1c6be8f97\") " pod="openstack/ovsdbserver-nb-0" Oct 05 21:09:48 crc kubenswrapper[4754]: I1005 21:09:48.799673 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/a5fdaf77-8a4c-4267-a32e-22a1c6be8f97-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"a5fdaf77-8a4c-4267-a32e-22a1c6be8f97\") " pod="openstack/ovsdbserver-nb-0" Oct 05 21:09:48 crc kubenswrapper[4754]: I1005 21:09:48.799749 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5fdaf77-8a4c-4267-a32e-22a1c6be8f97-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"a5fdaf77-8a4c-4267-a32e-22a1c6be8f97\") " pod="openstack/ovsdbserver-nb-0" Oct 05 21:09:48 crc kubenswrapper[4754]: I1005 21:09:48.799770 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a5fdaf77-8a4c-4267-a32e-22a1c6be8f97-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"a5fdaf77-8a4c-4267-a32e-22a1c6be8f97\") " pod="openstack/ovsdbserver-nb-0" Oct 05 21:09:48 crc kubenswrapper[4754]: I1005 21:09:48.799945 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/a5fdaf77-8a4c-4267-a32e-22a1c6be8f97-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"a5fdaf77-8a4c-4267-a32e-22a1c6be8f97\") " pod="openstack/ovsdbserver-nb-0" Oct 05 21:09:48 crc kubenswrapper[4754]: I1005 21:09:48.800002 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mk2bb\" (UniqueName: \"kubernetes.io/projected/a5fdaf77-8a4c-4267-a32e-22a1c6be8f97-kube-api-access-mk2bb\") pod \"ovsdbserver-nb-0\" (UID: \"a5fdaf77-8a4c-4267-a32e-22a1c6be8f97\") " pod="openstack/ovsdbserver-nb-0" Oct 05 21:09:48 crc kubenswrapper[4754]: I1005 21:09:48.800024 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a5fdaf77-8a4c-4267-a32e-22a1c6be8f97-config\") pod \"ovsdbserver-nb-0\" (UID: \"a5fdaf77-8a4c-4267-a32e-22a1c6be8f97\") " pod="openstack/ovsdbserver-nb-0" Oct 05 21:09:48 crc kubenswrapper[4754]: I1005 21:09:48.800143 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"ovsdbserver-nb-0\" (UID: \"a5fdaf77-8a4c-4267-a32e-22a1c6be8f97\") " pod="openstack/ovsdbserver-nb-0" Oct 05 21:09:48 crc kubenswrapper[4754]: I1005 21:09:48.901108 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a5fdaf77-8a4c-4267-a32e-22a1c6be8f97-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"a5fdaf77-8a4c-4267-a32e-22a1c6be8f97\") " pod="openstack/ovsdbserver-nb-0" Oct 05 21:09:48 crc kubenswrapper[4754]: I1005 21:09:48.901160 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/a5fdaf77-8a4c-4267-a32e-22a1c6be8f97-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"a5fdaf77-8a4c-4267-a32e-22a1c6be8f97\") " pod="openstack/ovsdbserver-nb-0" Oct 05 21:09:48 crc kubenswrapper[4754]: I1005 21:09:48.901213 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5fdaf77-8a4c-4267-a32e-22a1c6be8f97-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"a5fdaf77-8a4c-4267-a32e-22a1c6be8f97\") " pod="openstack/ovsdbserver-nb-0" Oct 05 21:09:48 crc kubenswrapper[4754]: I1005 21:09:48.901231 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a5fdaf77-8a4c-4267-a32e-22a1c6be8f97-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"a5fdaf77-8a4c-4267-a32e-22a1c6be8f97\") " pod="openstack/ovsdbserver-nb-0" Oct 05 21:09:48 crc kubenswrapper[4754]: I1005 21:09:48.901280 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/a5fdaf77-8a4c-4267-a32e-22a1c6be8f97-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"a5fdaf77-8a4c-4267-a32e-22a1c6be8f97\") " pod="openstack/ovsdbserver-nb-0" Oct 05 21:09:48 crc kubenswrapper[4754]: I1005 21:09:48.901303 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mk2bb\" (UniqueName: \"kubernetes.io/projected/a5fdaf77-8a4c-4267-a32e-22a1c6be8f97-kube-api-access-mk2bb\") pod \"ovsdbserver-nb-0\" (UID: \"a5fdaf77-8a4c-4267-a32e-22a1c6be8f97\") " pod="openstack/ovsdbserver-nb-0" Oct 05 21:09:48 crc kubenswrapper[4754]: I1005 21:09:48.901318 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a5fdaf77-8a4c-4267-a32e-22a1c6be8f97-config\") pod \"ovsdbserver-nb-0\" (UID: \"a5fdaf77-8a4c-4267-a32e-22a1c6be8f97\") " pod="openstack/ovsdbserver-nb-0" Oct 05 21:09:48 crc kubenswrapper[4754]: I1005 21:09:48.901355 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"ovsdbserver-nb-0\" (UID: \"a5fdaf77-8a4c-4267-a32e-22a1c6be8f97\") " pod="openstack/ovsdbserver-nb-0" Oct 05 21:09:48 crc kubenswrapper[4754]: I1005 21:09:48.902203 4754 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"ovsdbserver-nb-0\" (UID: \"a5fdaf77-8a4c-4267-a32e-22a1c6be8f97\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/ovsdbserver-nb-0" Oct 05 21:09:48 crc kubenswrapper[4754]: I1005 21:09:48.902649 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a5fdaf77-8a4c-4267-a32e-22a1c6be8f97-config\") pod \"ovsdbserver-nb-0\" (UID: \"a5fdaf77-8a4c-4267-a32e-22a1c6be8f97\") " pod="openstack/ovsdbserver-nb-0" Oct 05 21:09:48 crc kubenswrapper[4754]: I1005 21:09:48.903274 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/a5fdaf77-8a4c-4267-a32e-22a1c6be8f97-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"a5fdaf77-8a4c-4267-a32e-22a1c6be8f97\") " pod="openstack/ovsdbserver-nb-0" Oct 05 21:09:48 crc kubenswrapper[4754]: I1005 21:09:48.905346 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a5fdaf77-8a4c-4267-a32e-22a1c6be8f97-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"a5fdaf77-8a4c-4267-a32e-22a1c6be8f97\") " pod="openstack/ovsdbserver-nb-0" Oct 05 21:09:48 crc kubenswrapper[4754]: I1005 21:09:48.907994 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5fdaf77-8a4c-4267-a32e-22a1c6be8f97-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"a5fdaf77-8a4c-4267-a32e-22a1c6be8f97\") " pod="openstack/ovsdbserver-nb-0" Oct 05 21:09:48 crc kubenswrapper[4754]: I1005 21:09:48.916268 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a5fdaf77-8a4c-4267-a32e-22a1c6be8f97-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"a5fdaf77-8a4c-4267-a32e-22a1c6be8f97\") " pod="openstack/ovsdbserver-nb-0" Oct 05 21:09:48 crc kubenswrapper[4754]: I1005 21:09:48.920859 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mk2bb\" (UniqueName: \"kubernetes.io/projected/a5fdaf77-8a4c-4267-a32e-22a1c6be8f97-kube-api-access-mk2bb\") pod \"ovsdbserver-nb-0\" (UID: \"a5fdaf77-8a4c-4267-a32e-22a1c6be8f97\") " pod="openstack/ovsdbserver-nb-0" Oct 05 21:09:48 crc kubenswrapper[4754]: I1005 21:09:48.932014 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/a5fdaf77-8a4c-4267-a32e-22a1c6be8f97-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"a5fdaf77-8a4c-4267-a32e-22a1c6be8f97\") " pod="openstack/ovsdbserver-nb-0" Oct 05 21:09:48 crc kubenswrapper[4754]: I1005 21:09:48.956600 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"ovsdbserver-nb-0\" (UID: \"a5fdaf77-8a4c-4267-a32e-22a1c6be8f97\") " pod="openstack/ovsdbserver-nb-0" Oct 05 21:09:49 crc kubenswrapper[4754]: I1005 21:09:49.023800 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 05 21:09:51 crc kubenswrapper[4754]: I1005 21:09:51.028415 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 05 21:09:51 crc kubenswrapper[4754]: I1005 21:09:51.031199 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 05 21:09:51 crc kubenswrapper[4754]: I1005 21:09:51.035942 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 05 21:09:51 crc kubenswrapper[4754]: I1005 21:09:51.038942 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Oct 05 21:09:51 crc kubenswrapper[4754]: I1005 21:09:51.039823 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Oct 05 21:09:51 crc kubenswrapper[4754]: I1005 21:09:51.039914 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-8hqcr" Oct 05 21:09:51 crc kubenswrapper[4754]: I1005 21:09:51.041827 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Oct 05 21:09:51 crc kubenswrapper[4754]: I1005 21:09:51.170595 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/3c38a205-1079-46c2-9db5-c380d19399e3-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"3c38a205-1079-46c2-9db5-c380d19399e3\") " pod="openstack/ovsdbserver-sb-0" Oct 05 21:09:51 crc kubenswrapper[4754]: I1005 21:09:51.170658 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8vz7t\" (UniqueName: \"kubernetes.io/projected/3c38a205-1079-46c2-9db5-c380d19399e3-kube-api-access-8vz7t\") pod \"ovsdbserver-sb-0\" (UID: \"3c38a205-1079-46c2-9db5-c380d19399e3\") " pod="openstack/ovsdbserver-sb-0" Oct 05 21:09:51 crc kubenswrapper[4754]: I1005 21:09:51.170688 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3c38a205-1079-46c2-9db5-c380d19399e3-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"3c38a205-1079-46c2-9db5-c380d19399e3\") " pod="openstack/ovsdbserver-sb-0" Oct 05 21:09:51 crc kubenswrapper[4754]: I1005 21:09:51.170739 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/3c38a205-1079-46c2-9db5-c380d19399e3-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"3c38a205-1079-46c2-9db5-c380d19399e3\") " pod="openstack/ovsdbserver-sb-0" Oct 05 21:09:51 crc kubenswrapper[4754]: I1005 21:09:51.170768 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/3c38a205-1079-46c2-9db5-c380d19399e3-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"3c38a205-1079-46c2-9db5-c380d19399e3\") " pod="openstack/ovsdbserver-sb-0" Oct 05 21:09:51 crc kubenswrapper[4754]: I1005 21:09:51.170812 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-sb-0\" (UID: \"3c38a205-1079-46c2-9db5-c380d19399e3\") " pod="openstack/ovsdbserver-sb-0" Oct 05 21:09:51 crc kubenswrapper[4754]: I1005 21:09:51.170838 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3c38a205-1079-46c2-9db5-c380d19399e3-config\") pod \"ovsdbserver-sb-0\" (UID: \"3c38a205-1079-46c2-9db5-c380d19399e3\") " pod="openstack/ovsdbserver-sb-0" Oct 05 21:09:51 crc kubenswrapper[4754]: I1005 21:09:51.170854 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c38a205-1079-46c2-9db5-c380d19399e3-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"3c38a205-1079-46c2-9db5-c380d19399e3\") " pod="openstack/ovsdbserver-sb-0" Oct 05 21:09:51 crc kubenswrapper[4754]: I1005 21:09:51.272107 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-sb-0\" (UID: \"3c38a205-1079-46c2-9db5-c380d19399e3\") " pod="openstack/ovsdbserver-sb-0" Oct 05 21:09:51 crc kubenswrapper[4754]: I1005 21:09:51.272176 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3c38a205-1079-46c2-9db5-c380d19399e3-config\") pod \"ovsdbserver-sb-0\" (UID: \"3c38a205-1079-46c2-9db5-c380d19399e3\") " pod="openstack/ovsdbserver-sb-0" Oct 05 21:09:51 crc kubenswrapper[4754]: I1005 21:09:51.272203 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c38a205-1079-46c2-9db5-c380d19399e3-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"3c38a205-1079-46c2-9db5-c380d19399e3\") " pod="openstack/ovsdbserver-sb-0" Oct 05 21:09:51 crc kubenswrapper[4754]: I1005 21:09:51.272248 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/3c38a205-1079-46c2-9db5-c380d19399e3-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"3c38a205-1079-46c2-9db5-c380d19399e3\") " pod="openstack/ovsdbserver-sb-0" Oct 05 21:09:51 crc kubenswrapper[4754]: I1005 21:09:51.272284 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8vz7t\" (UniqueName: \"kubernetes.io/projected/3c38a205-1079-46c2-9db5-c380d19399e3-kube-api-access-8vz7t\") pod \"ovsdbserver-sb-0\" (UID: \"3c38a205-1079-46c2-9db5-c380d19399e3\") " pod="openstack/ovsdbserver-sb-0" Oct 05 21:09:51 crc kubenswrapper[4754]: I1005 21:09:51.272316 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3c38a205-1079-46c2-9db5-c380d19399e3-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"3c38a205-1079-46c2-9db5-c380d19399e3\") " pod="openstack/ovsdbserver-sb-0" Oct 05 21:09:51 crc kubenswrapper[4754]: I1005 21:09:51.272375 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/3c38a205-1079-46c2-9db5-c380d19399e3-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"3c38a205-1079-46c2-9db5-c380d19399e3\") " pod="openstack/ovsdbserver-sb-0" Oct 05 21:09:51 crc kubenswrapper[4754]: I1005 21:09:51.272413 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/3c38a205-1079-46c2-9db5-c380d19399e3-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"3c38a205-1079-46c2-9db5-c380d19399e3\") " pod="openstack/ovsdbserver-sb-0" Oct 05 21:09:51 crc kubenswrapper[4754]: I1005 21:09:51.273090 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/3c38a205-1079-46c2-9db5-c380d19399e3-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"3c38a205-1079-46c2-9db5-c380d19399e3\") " pod="openstack/ovsdbserver-sb-0" Oct 05 21:09:51 crc kubenswrapper[4754]: I1005 21:09:51.273469 4754 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-sb-0\" (UID: \"3c38a205-1079-46c2-9db5-c380d19399e3\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/ovsdbserver-sb-0" Oct 05 21:09:51 crc kubenswrapper[4754]: I1005 21:09:51.274617 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3c38a205-1079-46c2-9db5-c380d19399e3-config\") pod \"ovsdbserver-sb-0\" (UID: \"3c38a205-1079-46c2-9db5-c380d19399e3\") " pod="openstack/ovsdbserver-sb-0" Oct 05 21:09:51 crc kubenswrapper[4754]: I1005 21:09:51.276149 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3c38a205-1079-46c2-9db5-c380d19399e3-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"3c38a205-1079-46c2-9db5-c380d19399e3\") " pod="openstack/ovsdbserver-sb-0" Oct 05 21:09:51 crc kubenswrapper[4754]: I1005 21:09:51.284352 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/3c38a205-1079-46c2-9db5-c380d19399e3-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"3c38a205-1079-46c2-9db5-c380d19399e3\") " pod="openstack/ovsdbserver-sb-0" Oct 05 21:09:51 crc kubenswrapper[4754]: I1005 21:09:51.284681 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/3c38a205-1079-46c2-9db5-c380d19399e3-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"3c38a205-1079-46c2-9db5-c380d19399e3\") " pod="openstack/ovsdbserver-sb-0" Oct 05 21:09:51 crc kubenswrapper[4754]: I1005 21:09:51.285345 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c38a205-1079-46c2-9db5-c380d19399e3-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"3c38a205-1079-46c2-9db5-c380d19399e3\") " pod="openstack/ovsdbserver-sb-0" Oct 05 21:09:51 crc kubenswrapper[4754]: I1005 21:09:51.311209 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-sb-0\" (UID: \"3c38a205-1079-46c2-9db5-c380d19399e3\") " pod="openstack/ovsdbserver-sb-0" Oct 05 21:09:51 crc kubenswrapper[4754]: I1005 21:09:51.314324 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8vz7t\" (UniqueName: \"kubernetes.io/projected/3c38a205-1079-46c2-9db5-c380d19399e3-kube-api-access-8vz7t\") pod \"ovsdbserver-sb-0\" (UID: \"3c38a205-1079-46c2-9db5-c380d19399e3\") " pod="openstack/ovsdbserver-sb-0" Oct 05 21:09:51 crc kubenswrapper[4754]: I1005 21:09:51.370810 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 05 21:10:05 crc kubenswrapper[4754]: I1005 21:10:05.244990 4754 patch_prober.go:28] interesting pod/machine-config-daemon-b2h9k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 21:10:05 crc kubenswrapper[4754]: I1005 21:10:05.245872 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 21:10:11 crc kubenswrapper[4754]: E1005 21:10:11.072138 4754 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified" Oct 05 21:10:11 crc kubenswrapper[4754]: E1005 21:10:11.072884 4754 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6ltd4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cell1-server-0_openstack(1d96e280-8c26-41ea-ab29-64ce5b9c6a2c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 05 21:10:11 crc kubenswrapper[4754]: E1005 21:10:11.074174 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/rabbitmq-cell1-server-0" podUID="1d96e280-8c26-41ea-ab29-64ce5b9c6a2c" Oct 05 21:10:11 crc kubenswrapper[4754]: E1005 21:10:11.120178 4754 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified" Oct 05 21:10:11 crc kubenswrapper[4754]: E1005 21:10:11.120382 4754 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-jkfvq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-server-0_openstack(39aff8a0-2905-4ad3-ad2d-5bf6b5de3858): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 05 21:10:11 crc kubenswrapper[4754]: E1005 21:10:11.122078 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/rabbitmq-server-0" podUID="39aff8a0-2905-4ad3-ad2d-5bf6b5de3858" Oct 05 21:10:11 crc kubenswrapper[4754]: E1005 21:10:11.707706 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified\\\"\"" pod="openstack/rabbitmq-cell1-server-0" podUID="1d96e280-8c26-41ea-ab29-64ce5b9c6a2c" Oct 05 21:10:11 crc kubenswrapper[4754]: E1005 21:10:11.709716 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified\\\"\"" pod="openstack/rabbitmq-server-0" podUID="39aff8a0-2905-4ad3-ad2d-5bf6b5de3858" Oct 05 21:10:18 crc kubenswrapper[4754]: E1005 21:10:18.331375 4754 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-mariadb:current-podified" Oct 05 21:10:18 crc kubenswrapper[4754]: E1005 21:10:18.332209 4754 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:mysql-bootstrap,Image:quay.io/podified-antelope-centos9/openstack-mariadb:current-podified,Command:[bash /var/lib/operator-scripts/mysql_bootstrap.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:True,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:DB_ROOT_PASSWORD,Value:,ValueFrom:&EnvVarSource{FieldRef:nil,ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:&SecretKeySelector{LocalObjectReference:LocalObjectReference{Name:osp-secret,},Key:DbRootPassword,Optional:nil,},},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:mysql-db,ReadOnly:false,MountPath:/var/lib/mysql,SubPath:mysql,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-default,ReadOnly:true,MountPath:/var/lib/config-data/default,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-generated,ReadOnly:false,MountPath:/var/lib/config-data/generated,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:secrets,ReadOnly:true,MountPath:/var/lib/secrets,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:operator-scripts,ReadOnly:true,MountPath:/var/lib/operator-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kolla-config,ReadOnly:true,MountPath:/var/lib/kolla/config_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-7jn87,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-cell1-galera-0_openstack(d97f6b34-ef35-4b21-81eb-9e7009945fa6): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 05 21:10:18 crc kubenswrapper[4754]: E1005 21:10:18.334481 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/openstack-cell1-galera-0" podUID="d97f6b34-ef35-4b21-81eb-9e7009945fa6" Oct 05 21:10:18 crc kubenswrapper[4754]: E1005 21:10:18.764985 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-mariadb:current-podified\\\"\"" pod="openstack/openstack-cell1-galera-0" podUID="d97f6b34-ef35-4b21-81eb-9e7009945fa6" Oct 05 21:10:19 crc kubenswrapper[4754]: E1005 21:10:19.092484 4754 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-memcached:current-podified" Oct 05 21:10:19 crc kubenswrapper[4754]: E1005 21:10:19.094736 4754 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:memcached,Image:quay.io/podified-antelope-centos9/openstack-memcached:current-podified,Command:[/usr/bin/dumb-init -- /usr/local/bin/kolla_start],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:memcached,HostPort:0,ContainerPort:11211,Protocol:TCP,HostIP:,},ContainerPort{Name:memcached-tls,HostPort:0,ContainerPort:11212,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:POD_IPS,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIPs,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:CONFIG_HASH,Value:n5dbh8dh56bh5f8h84h5c7h695h567h5fdh67ch568h59ch64h65h58dh5bfhdch5fh657h664h56hb5h58hdhd8hfch575h677h5ddh75hd4h6fq,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/src,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kolla-config,ReadOnly:true,MountPath:/var/lib/kolla/config_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:memcached-tls-certs,ReadOnly:true,MountPath:/var/lib/config-data/tls/certs/memcached.crt,SubPath:tls.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:memcached-tls-certs,ReadOnly:true,MountPath:/var/lib/config-data/tls/private/memcached.key,SubPath:tls.key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-jnlj8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 11211 },Host:,},GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 11211 },Host:,},GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42457,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42457,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod memcached-0_openstack(bdbca489-8f9b-420e-a306-65575175af99): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 05 21:10:19 crc kubenswrapper[4754]: E1005 21:10:19.095963 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"memcached\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/memcached-0" podUID="bdbca489-8f9b-420e-a306-65575175af99" Oct 05 21:10:19 crc kubenswrapper[4754]: E1005 21:10:19.776637 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"memcached\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-memcached:current-podified\\\"\"" pod="openstack/memcached-0" podUID="bdbca489-8f9b-420e-a306-65575175af99" Oct 05 21:10:20 crc kubenswrapper[4754]: E1005 21:10:20.163780 4754 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Oct 05 21:10:20 crc kubenswrapper[4754]: E1005 21:10:20.164336 4754 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-67tht,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78dd6ddcc-crmqb_openstack(bcac2b2f-d2f3-438a-8bbd-92586d284527): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 05 21:10:20 crc kubenswrapper[4754]: E1005 21:10:20.166771 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78dd6ddcc-crmqb" podUID="bcac2b2f-d2f3-438a-8bbd-92586d284527" Oct 05 21:10:20 crc kubenswrapper[4754]: E1005 21:10:20.173865 4754 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Oct 05 21:10:20 crc kubenswrapper[4754]: E1005 21:10:20.176687 4754 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n68chd6h679hbfh55fhc6h5ffh5d8h94h56ch589hb4hc5h57bh677hcdh655h8dh667h675h654h66ch567h8fh659h5b4h675h566h55bh54h67dh6dq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-tdg46,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-666b6646f7-wfrrs_openstack(c21fb2d2-38b1-41a0-8df8-d34aedfee879): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 05 21:10:20 crc kubenswrapper[4754]: E1005 21:10:20.177875 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-666b6646f7-wfrrs" podUID="c21fb2d2-38b1-41a0-8df8-d34aedfee879" Oct 05 21:10:20 crc kubenswrapper[4754]: E1005 21:10:20.256457 4754 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Oct 05 21:10:20 crc kubenswrapper[4754]: E1005 21:10:20.257073 4754 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-7lg65,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-675f4bcbfc-l4bph_openstack(de3fa71c-9583-40d3-a47d-48ce71a7793c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 05 21:10:20 crc kubenswrapper[4754]: E1005 21:10:20.258474 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-675f4bcbfc-l4bph" podUID="de3fa71c-9583-40d3-a47d-48ce71a7793c" Oct 05 21:10:20 crc kubenswrapper[4754]: E1005 21:10:20.339353 4754 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Oct 05 21:10:20 crc kubenswrapper[4754]: E1005 21:10:20.339587 4754 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n659h4h664hbh658h587h67ch89h587h8fh679hc6hf9h55fh644h5d5h698h68dh5cdh5ffh669h54ch9h689hb8hd4h5bfhd8h5d7h5fh665h574q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-dts6z,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-57d769cc4f-d5rsv_openstack(22f48eb7-09d6-44bc-bc47-ba7026cf4cc3): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 05 21:10:20 crc kubenswrapper[4754]: E1005 21:10:20.342977 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-57d769cc4f-d5rsv" podUID="22f48eb7-09d6-44bc-bc47-ba7026cf4cc3" Oct 05 21:10:20 crc kubenswrapper[4754]: E1005 21:10:20.793360 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-666b6646f7-wfrrs" podUID="c21fb2d2-38b1-41a0-8df8-d34aedfee879" Oct 05 21:10:20 crc kubenswrapper[4754]: E1005 21:10:20.796336 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-57d769cc4f-d5rsv" podUID="22f48eb7-09d6-44bc-bc47-ba7026cf4cc3" Oct 05 21:10:20 crc kubenswrapper[4754]: I1005 21:10:20.927754 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-pp728"] Oct 05 21:10:21 crc kubenswrapper[4754]: W1005 21:10:21.140380 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode3785031_57bd_44ab_b30e_b609cca6dcaf.slice/crio-fa67733b522ec248bfff558a5b2c377f403905066a331471e65f4c302dbd3859 WatchSource:0}: Error finding container fa67733b522ec248bfff558a5b2c377f403905066a331471e65f4c302dbd3859: Status 404 returned error can't find the container with id fa67733b522ec248bfff558a5b2c377f403905066a331471e65f4c302dbd3859 Oct 05 21:10:21 crc kubenswrapper[4754]: I1005 21:10:21.228699 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-2krf2"] Oct 05 21:10:21 crc kubenswrapper[4754]: I1005 21:10:21.330544 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-crmqb" Oct 05 21:10:21 crc kubenswrapper[4754]: W1005 21:10:21.360598 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1283a11a_7581_4146_87db_b62d6d5d6dc9.slice/crio-5ded116737630a1d61a74d4aad7fc458b55417f874fccb7da4bab02efcbaff90 WatchSource:0}: Error finding container 5ded116737630a1d61a74d4aad7fc458b55417f874fccb7da4bab02efcbaff90: Status 404 returned error can't find the container with id 5ded116737630a1d61a74d4aad7fc458b55417f874fccb7da4bab02efcbaff90 Oct 05 21:10:21 crc kubenswrapper[4754]: I1005 21:10:21.427352 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-l4bph" Oct 05 21:10:21 crc kubenswrapper[4754]: I1005 21:10:21.471817 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bcac2b2f-d2f3-438a-8bbd-92586d284527-config\") pod \"bcac2b2f-d2f3-438a-8bbd-92586d284527\" (UID: \"bcac2b2f-d2f3-438a-8bbd-92586d284527\") " Oct 05 21:10:21 crc kubenswrapper[4754]: I1005 21:10:21.471870 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bcac2b2f-d2f3-438a-8bbd-92586d284527-dns-svc\") pod \"bcac2b2f-d2f3-438a-8bbd-92586d284527\" (UID: \"bcac2b2f-d2f3-438a-8bbd-92586d284527\") " Oct 05 21:10:21 crc kubenswrapper[4754]: I1005 21:10:21.472060 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-67tht\" (UniqueName: \"kubernetes.io/projected/bcac2b2f-d2f3-438a-8bbd-92586d284527-kube-api-access-67tht\") pod \"bcac2b2f-d2f3-438a-8bbd-92586d284527\" (UID: \"bcac2b2f-d2f3-438a-8bbd-92586d284527\") " Oct 05 21:10:21 crc kubenswrapper[4754]: I1005 21:10:21.473234 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bcac2b2f-d2f3-438a-8bbd-92586d284527-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "bcac2b2f-d2f3-438a-8bbd-92586d284527" (UID: "bcac2b2f-d2f3-438a-8bbd-92586d284527"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 21:10:21 crc kubenswrapper[4754]: I1005 21:10:21.473269 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bcac2b2f-d2f3-438a-8bbd-92586d284527-config" (OuterVolumeSpecName: "config") pod "bcac2b2f-d2f3-438a-8bbd-92586d284527" (UID: "bcac2b2f-d2f3-438a-8bbd-92586d284527"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 21:10:21 crc kubenswrapper[4754]: I1005 21:10:21.485724 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bcac2b2f-d2f3-438a-8bbd-92586d284527-kube-api-access-67tht" (OuterVolumeSpecName: "kube-api-access-67tht") pod "bcac2b2f-d2f3-438a-8bbd-92586d284527" (UID: "bcac2b2f-d2f3-438a-8bbd-92586d284527"). InnerVolumeSpecName "kube-api-access-67tht". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:10:21 crc kubenswrapper[4754]: I1005 21:10:21.573642 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7lg65\" (UniqueName: \"kubernetes.io/projected/de3fa71c-9583-40d3-a47d-48ce71a7793c-kube-api-access-7lg65\") pod \"de3fa71c-9583-40d3-a47d-48ce71a7793c\" (UID: \"de3fa71c-9583-40d3-a47d-48ce71a7793c\") " Oct 05 21:10:21 crc kubenswrapper[4754]: I1005 21:10:21.573700 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/de3fa71c-9583-40d3-a47d-48ce71a7793c-config\") pod \"de3fa71c-9583-40d3-a47d-48ce71a7793c\" (UID: \"de3fa71c-9583-40d3-a47d-48ce71a7793c\") " Oct 05 21:10:21 crc kubenswrapper[4754]: I1005 21:10:21.574263 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-67tht\" (UniqueName: \"kubernetes.io/projected/bcac2b2f-d2f3-438a-8bbd-92586d284527-kube-api-access-67tht\") on node \"crc\" DevicePath \"\"" Oct 05 21:10:21 crc kubenswrapper[4754]: I1005 21:10:21.574277 4754 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bcac2b2f-d2f3-438a-8bbd-92586d284527-config\") on node \"crc\" DevicePath \"\"" Oct 05 21:10:21 crc kubenswrapper[4754]: I1005 21:10:21.574287 4754 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bcac2b2f-d2f3-438a-8bbd-92586d284527-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 05 21:10:21 crc kubenswrapper[4754]: I1005 21:10:21.575118 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/de3fa71c-9583-40d3-a47d-48ce71a7793c-config" (OuterVolumeSpecName: "config") pod "de3fa71c-9583-40d3-a47d-48ce71a7793c" (UID: "de3fa71c-9583-40d3-a47d-48ce71a7793c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 21:10:21 crc kubenswrapper[4754]: I1005 21:10:21.577140 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/de3fa71c-9583-40d3-a47d-48ce71a7793c-kube-api-access-7lg65" (OuterVolumeSpecName: "kube-api-access-7lg65") pod "de3fa71c-9583-40d3-a47d-48ce71a7793c" (UID: "de3fa71c-9583-40d3-a47d-48ce71a7793c"). InnerVolumeSpecName "kube-api-access-7lg65". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:10:21 crc kubenswrapper[4754]: I1005 21:10:21.675713 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7lg65\" (UniqueName: \"kubernetes.io/projected/de3fa71c-9583-40d3-a47d-48ce71a7793c-kube-api-access-7lg65\") on node \"crc\" DevicePath \"\"" Oct 05 21:10:21 crc kubenswrapper[4754]: I1005 21:10:21.675743 4754 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/de3fa71c-9583-40d3-a47d-48ce71a7793c-config\") on node \"crc\" DevicePath \"\"" Oct 05 21:10:21 crc kubenswrapper[4754]: I1005 21:10:21.781166 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 05 21:10:21 crc kubenswrapper[4754]: I1005 21:10:21.787849 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"d311946e-91c6-4be6-9d48-1b823ee77607","Type":"ContainerStarted","Data":"ec56f5e8176d9ae61641d3a197a6391d110b4913ff0b634a91334b4e35414421"} Oct 05 21:10:21 crc kubenswrapper[4754]: I1005 21:10:21.789291 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-crmqb" Oct 05 21:10:21 crc kubenswrapper[4754]: I1005 21:10:21.789322 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-crmqb" event={"ID":"bcac2b2f-d2f3-438a-8bbd-92586d284527","Type":"ContainerDied","Data":"d741907e2653480b7e663c8b051a35f64e26d72e09ce813aec3a33cecc7e8918"} Oct 05 21:10:21 crc kubenswrapper[4754]: I1005 21:10:21.790601 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-l4bph" Oct 05 21:10:21 crc kubenswrapper[4754]: I1005 21:10:21.800049 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-l4bph" event={"ID":"de3fa71c-9583-40d3-a47d-48ce71a7793c","Type":"ContainerDied","Data":"08fc02582a25331039f5d399075e8fb74a8e294bb09c5834fab9f7b6ed76551c"} Oct 05 21:10:21 crc kubenswrapper[4754]: I1005 21:10:21.800097 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-pp728" event={"ID":"e3785031-57bd-44ab-b30e-b609cca6dcaf","Type":"ContainerStarted","Data":"fa67733b522ec248bfff558a5b2c377f403905066a331471e65f4c302dbd3859"} Oct 05 21:10:21 crc kubenswrapper[4754]: I1005 21:10:21.800111 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-2krf2" event={"ID":"1283a11a-7581-4146-87db-b62d6d5d6dc9","Type":"ContainerStarted","Data":"5ded116737630a1d61a74d4aad7fc458b55417f874fccb7da4bab02efcbaff90"} Oct 05 21:10:21 crc kubenswrapper[4754]: I1005 21:10:21.915557 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-crmqb"] Oct 05 21:10:21 crc kubenswrapper[4754]: I1005 21:10:21.933544 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-crmqb"] Oct 05 21:10:21 crc kubenswrapper[4754]: I1005 21:10:21.974941 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-l4bph"] Oct 05 21:10:21 crc kubenswrapper[4754]: I1005 21:10:21.989148 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-l4bph"] Oct 05 21:10:22 crc kubenswrapper[4754]: I1005 21:10:22.133603 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-8xfw8"] Oct 05 21:10:22 crc kubenswrapper[4754]: I1005 21:10:22.134635 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-8xfw8" Oct 05 21:10:22 crc kubenswrapper[4754]: I1005 21:10:22.137158 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Oct 05 21:10:22 crc kubenswrapper[4754]: I1005 21:10:22.159645 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-8xfw8"] Oct 05 21:10:22 crc kubenswrapper[4754]: I1005 21:10:22.294246 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/06242eb6-ddde-49ea-b4a1-c61aad6f6402-ovs-rundir\") pod \"ovn-controller-metrics-8xfw8\" (UID: \"06242eb6-ddde-49ea-b4a1-c61aad6f6402\") " pod="openstack/ovn-controller-metrics-8xfw8" Oct 05 21:10:22 crc kubenswrapper[4754]: I1005 21:10:22.294302 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t8w6b\" (UniqueName: \"kubernetes.io/projected/06242eb6-ddde-49ea-b4a1-c61aad6f6402-kube-api-access-t8w6b\") pod \"ovn-controller-metrics-8xfw8\" (UID: \"06242eb6-ddde-49ea-b4a1-c61aad6f6402\") " pod="openstack/ovn-controller-metrics-8xfw8" Oct 05 21:10:22 crc kubenswrapper[4754]: I1005 21:10:22.294332 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06242eb6-ddde-49ea-b4a1-c61aad6f6402-combined-ca-bundle\") pod \"ovn-controller-metrics-8xfw8\" (UID: \"06242eb6-ddde-49ea-b4a1-c61aad6f6402\") " pod="openstack/ovn-controller-metrics-8xfw8" Oct 05 21:10:22 crc kubenswrapper[4754]: I1005 21:10:22.294367 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/06242eb6-ddde-49ea-b4a1-c61aad6f6402-config\") pod \"ovn-controller-metrics-8xfw8\" (UID: \"06242eb6-ddde-49ea-b4a1-c61aad6f6402\") " pod="openstack/ovn-controller-metrics-8xfw8" Oct 05 21:10:22 crc kubenswrapper[4754]: I1005 21:10:22.294419 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/06242eb6-ddde-49ea-b4a1-c61aad6f6402-ovn-rundir\") pod \"ovn-controller-metrics-8xfw8\" (UID: \"06242eb6-ddde-49ea-b4a1-c61aad6f6402\") " pod="openstack/ovn-controller-metrics-8xfw8" Oct 05 21:10:22 crc kubenswrapper[4754]: I1005 21:10:22.294434 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/06242eb6-ddde-49ea-b4a1-c61aad6f6402-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-8xfw8\" (UID: \"06242eb6-ddde-49ea-b4a1-c61aad6f6402\") " pod="openstack/ovn-controller-metrics-8xfw8" Oct 05 21:10:22 crc kubenswrapper[4754]: I1005 21:10:22.354372 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 05 21:10:22 crc kubenswrapper[4754]: I1005 21:10:22.389210 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-d5rsv"] Oct 05 21:10:22 crc kubenswrapper[4754]: I1005 21:10:22.402474 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/06242eb6-ddde-49ea-b4a1-c61aad6f6402-config\") pod \"ovn-controller-metrics-8xfw8\" (UID: \"06242eb6-ddde-49ea-b4a1-c61aad6f6402\") " pod="openstack/ovn-controller-metrics-8xfw8" Oct 05 21:10:22 crc kubenswrapper[4754]: I1005 21:10:22.402565 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/06242eb6-ddde-49ea-b4a1-c61aad6f6402-ovn-rundir\") pod \"ovn-controller-metrics-8xfw8\" (UID: \"06242eb6-ddde-49ea-b4a1-c61aad6f6402\") " pod="openstack/ovn-controller-metrics-8xfw8" Oct 05 21:10:22 crc kubenswrapper[4754]: I1005 21:10:22.402586 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/06242eb6-ddde-49ea-b4a1-c61aad6f6402-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-8xfw8\" (UID: \"06242eb6-ddde-49ea-b4a1-c61aad6f6402\") " pod="openstack/ovn-controller-metrics-8xfw8" Oct 05 21:10:22 crc kubenswrapper[4754]: I1005 21:10:22.402647 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/06242eb6-ddde-49ea-b4a1-c61aad6f6402-ovs-rundir\") pod \"ovn-controller-metrics-8xfw8\" (UID: \"06242eb6-ddde-49ea-b4a1-c61aad6f6402\") " pod="openstack/ovn-controller-metrics-8xfw8" Oct 05 21:10:22 crc kubenswrapper[4754]: I1005 21:10:22.402708 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t8w6b\" (UniqueName: \"kubernetes.io/projected/06242eb6-ddde-49ea-b4a1-c61aad6f6402-kube-api-access-t8w6b\") pod \"ovn-controller-metrics-8xfw8\" (UID: \"06242eb6-ddde-49ea-b4a1-c61aad6f6402\") " pod="openstack/ovn-controller-metrics-8xfw8" Oct 05 21:10:22 crc kubenswrapper[4754]: I1005 21:10:22.402734 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06242eb6-ddde-49ea-b4a1-c61aad6f6402-combined-ca-bundle\") pod \"ovn-controller-metrics-8xfw8\" (UID: \"06242eb6-ddde-49ea-b4a1-c61aad6f6402\") " pod="openstack/ovn-controller-metrics-8xfw8" Oct 05 21:10:22 crc kubenswrapper[4754]: I1005 21:10:22.403360 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/06242eb6-ddde-49ea-b4a1-c61aad6f6402-ovn-rundir\") pod \"ovn-controller-metrics-8xfw8\" (UID: \"06242eb6-ddde-49ea-b4a1-c61aad6f6402\") " pod="openstack/ovn-controller-metrics-8xfw8" Oct 05 21:10:22 crc kubenswrapper[4754]: I1005 21:10:22.403371 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/06242eb6-ddde-49ea-b4a1-c61aad6f6402-ovs-rundir\") pod \"ovn-controller-metrics-8xfw8\" (UID: \"06242eb6-ddde-49ea-b4a1-c61aad6f6402\") " pod="openstack/ovn-controller-metrics-8xfw8" Oct 05 21:10:22 crc kubenswrapper[4754]: I1005 21:10:22.404923 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/06242eb6-ddde-49ea-b4a1-c61aad6f6402-config\") pod \"ovn-controller-metrics-8xfw8\" (UID: \"06242eb6-ddde-49ea-b4a1-c61aad6f6402\") " pod="openstack/ovn-controller-metrics-8xfw8" Oct 05 21:10:22 crc kubenswrapper[4754]: I1005 21:10:22.407677 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06242eb6-ddde-49ea-b4a1-c61aad6f6402-combined-ca-bundle\") pod \"ovn-controller-metrics-8xfw8\" (UID: \"06242eb6-ddde-49ea-b4a1-c61aad6f6402\") " pod="openstack/ovn-controller-metrics-8xfw8" Oct 05 21:10:22 crc kubenswrapper[4754]: I1005 21:10:22.436953 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/06242eb6-ddde-49ea-b4a1-c61aad6f6402-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-8xfw8\" (UID: \"06242eb6-ddde-49ea-b4a1-c61aad6f6402\") " pod="openstack/ovn-controller-metrics-8xfw8" Oct 05 21:10:22 crc kubenswrapper[4754]: I1005 21:10:22.490293 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-hktv6"] Oct 05 21:10:22 crc kubenswrapper[4754]: I1005 21:10:22.497579 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-hktv6" Oct 05 21:10:22 crc kubenswrapper[4754]: I1005 21:10:22.501728 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Oct 05 21:10:22 crc kubenswrapper[4754]: I1005 21:10:22.506184 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t8w6b\" (UniqueName: \"kubernetes.io/projected/06242eb6-ddde-49ea-b4a1-c61aad6f6402-kube-api-access-t8w6b\") pod \"ovn-controller-metrics-8xfw8\" (UID: \"06242eb6-ddde-49ea-b4a1-c61aad6f6402\") " pod="openstack/ovn-controller-metrics-8xfw8" Oct 05 21:10:22 crc kubenswrapper[4754]: I1005 21:10:22.578555 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-hktv6"] Oct 05 21:10:22 crc kubenswrapper[4754]: I1005 21:10:22.608703 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b2dadb10-4ed0-4032-a049-a5b0289bd893-config\") pod \"dnsmasq-dns-7fd796d7df-hktv6\" (UID: \"b2dadb10-4ed0-4032-a049-a5b0289bd893\") " pod="openstack/dnsmasq-dns-7fd796d7df-hktv6" Oct 05 21:10:22 crc kubenswrapper[4754]: I1005 21:10:22.608793 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b2dadb10-4ed0-4032-a049-a5b0289bd893-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-hktv6\" (UID: \"b2dadb10-4ed0-4032-a049-a5b0289bd893\") " pod="openstack/dnsmasq-dns-7fd796d7df-hktv6" Oct 05 21:10:22 crc kubenswrapper[4754]: I1005 21:10:22.608823 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ch2t7\" (UniqueName: \"kubernetes.io/projected/b2dadb10-4ed0-4032-a049-a5b0289bd893-kube-api-access-ch2t7\") pod \"dnsmasq-dns-7fd796d7df-hktv6\" (UID: \"b2dadb10-4ed0-4032-a049-a5b0289bd893\") " pod="openstack/dnsmasq-dns-7fd796d7df-hktv6" Oct 05 21:10:22 crc kubenswrapper[4754]: I1005 21:10:22.608864 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b2dadb10-4ed0-4032-a049-a5b0289bd893-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-hktv6\" (UID: \"b2dadb10-4ed0-4032-a049-a5b0289bd893\") " pod="openstack/dnsmasq-dns-7fd796d7df-hktv6" Oct 05 21:10:22 crc kubenswrapper[4754]: I1005 21:10:22.629247 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-wfrrs"] Oct 05 21:10:22 crc kubenswrapper[4754]: I1005 21:10:22.680711 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-r8m5k"] Oct 05 21:10:22 crc kubenswrapper[4754]: I1005 21:10:22.682051 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-r8m5k" Oct 05 21:10:22 crc kubenswrapper[4754]: I1005 21:10:22.685336 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Oct 05 21:10:22 crc kubenswrapper[4754]: I1005 21:10:22.710172 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-r8m5k"] Oct 05 21:10:22 crc kubenswrapper[4754]: I1005 21:10:22.710370 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b2dadb10-4ed0-4032-a049-a5b0289bd893-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-hktv6\" (UID: \"b2dadb10-4ed0-4032-a049-a5b0289bd893\") " pod="openstack/dnsmasq-dns-7fd796d7df-hktv6" Oct 05 21:10:22 crc kubenswrapper[4754]: I1005 21:10:22.710460 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b2dadb10-4ed0-4032-a049-a5b0289bd893-config\") pod \"dnsmasq-dns-7fd796d7df-hktv6\" (UID: \"b2dadb10-4ed0-4032-a049-a5b0289bd893\") " pod="openstack/dnsmasq-dns-7fd796d7df-hktv6" Oct 05 21:10:22 crc kubenswrapper[4754]: I1005 21:10:22.710574 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b2dadb10-4ed0-4032-a049-a5b0289bd893-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-hktv6\" (UID: \"b2dadb10-4ed0-4032-a049-a5b0289bd893\") " pod="openstack/dnsmasq-dns-7fd796d7df-hktv6" Oct 05 21:10:22 crc kubenswrapper[4754]: I1005 21:10:22.710600 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ch2t7\" (UniqueName: \"kubernetes.io/projected/b2dadb10-4ed0-4032-a049-a5b0289bd893-kube-api-access-ch2t7\") pod \"dnsmasq-dns-7fd796d7df-hktv6\" (UID: \"b2dadb10-4ed0-4032-a049-a5b0289bd893\") " pod="openstack/dnsmasq-dns-7fd796d7df-hktv6" Oct 05 21:10:22 crc kubenswrapper[4754]: I1005 21:10:22.711646 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b2dadb10-4ed0-4032-a049-a5b0289bd893-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-hktv6\" (UID: \"b2dadb10-4ed0-4032-a049-a5b0289bd893\") " pod="openstack/dnsmasq-dns-7fd796d7df-hktv6" Oct 05 21:10:22 crc kubenswrapper[4754]: I1005 21:10:22.711803 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b2dadb10-4ed0-4032-a049-a5b0289bd893-config\") pod \"dnsmasq-dns-7fd796d7df-hktv6\" (UID: \"b2dadb10-4ed0-4032-a049-a5b0289bd893\") " pod="openstack/dnsmasq-dns-7fd796d7df-hktv6" Oct 05 21:10:22 crc kubenswrapper[4754]: I1005 21:10:22.712182 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b2dadb10-4ed0-4032-a049-a5b0289bd893-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-hktv6\" (UID: \"b2dadb10-4ed0-4032-a049-a5b0289bd893\") " pod="openstack/dnsmasq-dns-7fd796d7df-hktv6" Oct 05 21:10:22 crc kubenswrapper[4754]: I1005 21:10:22.742184 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ch2t7\" (UniqueName: \"kubernetes.io/projected/b2dadb10-4ed0-4032-a049-a5b0289bd893-kube-api-access-ch2t7\") pod \"dnsmasq-dns-7fd796d7df-hktv6\" (UID: \"b2dadb10-4ed0-4032-a049-a5b0289bd893\") " pod="openstack/dnsmasq-dns-7fd796d7df-hktv6" Oct 05 21:10:22 crc kubenswrapper[4754]: I1005 21:10:22.759857 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-8xfw8" Oct 05 21:10:22 crc kubenswrapper[4754]: I1005 21:10:22.812460 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a40e02a7-900e-4da2-8a4a-73773ca90876-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-r8m5k\" (UID: \"a40e02a7-900e-4da2-8a4a-73773ca90876\") " pod="openstack/dnsmasq-dns-86db49b7ff-r8m5k" Oct 05 21:10:22 crc kubenswrapper[4754]: I1005 21:10:22.812561 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a40e02a7-900e-4da2-8a4a-73773ca90876-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-r8m5k\" (UID: \"a40e02a7-900e-4da2-8a4a-73773ca90876\") " pod="openstack/dnsmasq-dns-86db49b7ff-r8m5k" Oct 05 21:10:22 crc kubenswrapper[4754]: I1005 21:10:22.812586 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a40e02a7-900e-4da2-8a4a-73773ca90876-config\") pod \"dnsmasq-dns-86db49b7ff-r8m5k\" (UID: \"a40e02a7-900e-4da2-8a4a-73773ca90876\") " pod="openstack/dnsmasq-dns-86db49b7ff-r8m5k" Oct 05 21:10:22 crc kubenswrapper[4754]: I1005 21:10:22.812628 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a40e02a7-900e-4da2-8a4a-73773ca90876-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-r8m5k\" (UID: \"a40e02a7-900e-4da2-8a4a-73773ca90876\") " pod="openstack/dnsmasq-dns-86db49b7ff-r8m5k" Oct 05 21:10:22 crc kubenswrapper[4754]: I1005 21:10:22.812671 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8ls6k\" (UniqueName: \"kubernetes.io/projected/a40e02a7-900e-4da2-8a4a-73773ca90876-kube-api-access-8ls6k\") pod \"dnsmasq-dns-86db49b7ff-r8m5k\" (UID: \"a40e02a7-900e-4da2-8a4a-73773ca90876\") " pod="openstack/dnsmasq-dns-86db49b7ff-r8m5k" Oct 05 21:10:22 crc kubenswrapper[4754]: I1005 21:10:22.868259 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bcac2b2f-d2f3-438a-8bbd-92586d284527" path="/var/lib/kubelet/pods/bcac2b2f-d2f3-438a-8bbd-92586d284527/volumes" Oct 05 21:10:22 crc kubenswrapper[4754]: I1005 21:10:22.868826 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="de3fa71c-9583-40d3-a47d-48ce71a7793c" path="/var/lib/kubelet/pods/de3fa71c-9583-40d3-a47d-48ce71a7793c/volumes" Oct 05 21:10:22 crc kubenswrapper[4754]: I1005 21:10:22.881549 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-hktv6" Oct 05 21:10:22 crc kubenswrapper[4754]: I1005 21:10:22.914759 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a40e02a7-900e-4da2-8a4a-73773ca90876-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-r8m5k\" (UID: \"a40e02a7-900e-4da2-8a4a-73773ca90876\") " pod="openstack/dnsmasq-dns-86db49b7ff-r8m5k" Oct 05 21:10:22 crc kubenswrapper[4754]: I1005 21:10:22.914905 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a40e02a7-900e-4da2-8a4a-73773ca90876-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-r8m5k\" (UID: \"a40e02a7-900e-4da2-8a4a-73773ca90876\") " pod="openstack/dnsmasq-dns-86db49b7ff-r8m5k" Oct 05 21:10:22 crc kubenswrapper[4754]: I1005 21:10:22.914927 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a40e02a7-900e-4da2-8a4a-73773ca90876-config\") pod \"dnsmasq-dns-86db49b7ff-r8m5k\" (UID: \"a40e02a7-900e-4da2-8a4a-73773ca90876\") " pod="openstack/dnsmasq-dns-86db49b7ff-r8m5k" Oct 05 21:10:22 crc kubenswrapper[4754]: I1005 21:10:22.915038 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a40e02a7-900e-4da2-8a4a-73773ca90876-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-r8m5k\" (UID: \"a40e02a7-900e-4da2-8a4a-73773ca90876\") " pod="openstack/dnsmasq-dns-86db49b7ff-r8m5k" Oct 05 21:10:22 crc kubenswrapper[4754]: I1005 21:10:22.915108 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8ls6k\" (UniqueName: \"kubernetes.io/projected/a40e02a7-900e-4da2-8a4a-73773ca90876-kube-api-access-8ls6k\") pod \"dnsmasq-dns-86db49b7ff-r8m5k\" (UID: \"a40e02a7-900e-4da2-8a4a-73773ca90876\") " pod="openstack/dnsmasq-dns-86db49b7ff-r8m5k" Oct 05 21:10:22 crc kubenswrapper[4754]: I1005 21:10:22.916199 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a40e02a7-900e-4da2-8a4a-73773ca90876-config\") pod \"dnsmasq-dns-86db49b7ff-r8m5k\" (UID: \"a40e02a7-900e-4da2-8a4a-73773ca90876\") " pod="openstack/dnsmasq-dns-86db49b7ff-r8m5k" Oct 05 21:10:22 crc kubenswrapper[4754]: I1005 21:10:22.917339 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a40e02a7-900e-4da2-8a4a-73773ca90876-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-r8m5k\" (UID: \"a40e02a7-900e-4da2-8a4a-73773ca90876\") " pod="openstack/dnsmasq-dns-86db49b7ff-r8m5k" Oct 05 21:10:22 crc kubenswrapper[4754]: I1005 21:10:22.918315 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a40e02a7-900e-4da2-8a4a-73773ca90876-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-r8m5k\" (UID: \"a40e02a7-900e-4da2-8a4a-73773ca90876\") " pod="openstack/dnsmasq-dns-86db49b7ff-r8m5k" Oct 05 21:10:22 crc kubenswrapper[4754]: I1005 21:10:22.918676 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a40e02a7-900e-4da2-8a4a-73773ca90876-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-r8m5k\" (UID: \"a40e02a7-900e-4da2-8a4a-73773ca90876\") " pod="openstack/dnsmasq-dns-86db49b7ff-r8m5k" Oct 05 21:10:22 crc kubenswrapper[4754]: I1005 21:10:22.942259 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8ls6k\" (UniqueName: \"kubernetes.io/projected/a40e02a7-900e-4da2-8a4a-73773ca90876-kube-api-access-8ls6k\") pod \"dnsmasq-dns-86db49b7ff-r8m5k\" (UID: \"a40e02a7-900e-4da2-8a4a-73773ca90876\") " pod="openstack/dnsmasq-dns-86db49b7ff-r8m5k" Oct 05 21:10:23 crc kubenswrapper[4754]: I1005 21:10:23.007510 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-r8m5k" Oct 05 21:10:23 crc kubenswrapper[4754]: W1005 21:10:23.300183 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda5fdaf77_8a4c_4267_a32e_22a1c6be8f97.slice/crio-889ddb0166b138db89f77d95c243b2d378127d42b27fc304131c4c010b708d99 WatchSource:0}: Error finding container 889ddb0166b138db89f77d95c243b2d378127d42b27fc304131c4c010b708d99: Status 404 returned error can't find the container with id 889ddb0166b138db89f77d95c243b2d378127d42b27fc304131c4c010b708d99 Oct 05 21:10:23 crc kubenswrapper[4754]: I1005 21:10:23.360694 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-wfrrs" Oct 05 21:10:23 crc kubenswrapper[4754]: I1005 21:10:23.363300 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-d5rsv" Oct 05 21:10:23 crc kubenswrapper[4754]: I1005 21:10:23.423161 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22f48eb7-09d6-44bc-bc47-ba7026cf4cc3-config\") pod \"22f48eb7-09d6-44bc-bc47-ba7026cf4cc3\" (UID: \"22f48eb7-09d6-44bc-bc47-ba7026cf4cc3\") " Oct 05 21:10:23 crc kubenswrapper[4754]: I1005 21:10:23.423308 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tdg46\" (UniqueName: \"kubernetes.io/projected/c21fb2d2-38b1-41a0-8df8-d34aedfee879-kube-api-access-tdg46\") pod \"c21fb2d2-38b1-41a0-8df8-d34aedfee879\" (UID: \"c21fb2d2-38b1-41a0-8df8-d34aedfee879\") " Oct 05 21:10:23 crc kubenswrapper[4754]: I1005 21:10:23.423380 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c21fb2d2-38b1-41a0-8df8-d34aedfee879-config\") pod \"c21fb2d2-38b1-41a0-8df8-d34aedfee879\" (UID: \"c21fb2d2-38b1-41a0-8df8-d34aedfee879\") " Oct 05 21:10:23 crc kubenswrapper[4754]: I1005 21:10:23.423406 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dts6z\" (UniqueName: \"kubernetes.io/projected/22f48eb7-09d6-44bc-bc47-ba7026cf4cc3-kube-api-access-dts6z\") pod \"22f48eb7-09d6-44bc-bc47-ba7026cf4cc3\" (UID: \"22f48eb7-09d6-44bc-bc47-ba7026cf4cc3\") " Oct 05 21:10:23 crc kubenswrapper[4754]: I1005 21:10:23.423443 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/22f48eb7-09d6-44bc-bc47-ba7026cf4cc3-dns-svc\") pod \"22f48eb7-09d6-44bc-bc47-ba7026cf4cc3\" (UID: \"22f48eb7-09d6-44bc-bc47-ba7026cf4cc3\") " Oct 05 21:10:23 crc kubenswrapper[4754]: I1005 21:10:23.423480 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c21fb2d2-38b1-41a0-8df8-d34aedfee879-dns-svc\") pod \"c21fb2d2-38b1-41a0-8df8-d34aedfee879\" (UID: \"c21fb2d2-38b1-41a0-8df8-d34aedfee879\") " Oct 05 21:10:23 crc kubenswrapper[4754]: I1005 21:10:23.423774 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22f48eb7-09d6-44bc-bc47-ba7026cf4cc3-config" (OuterVolumeSpecName: "config") pod "22f48eb7-09d6-44bc-bc47-ba7026cf4cc3" (UID: "22f48eb7-09d6-44bc-bc47-ba7026cf4cc3"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 21:10:23 crc kubenswrapper[4754]: I1005 21:10:23.424115 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c21fb2d2-38b1-41a0-8df8-d34aedfee879-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "c21fb2d2-38b1-41a0-8df8-d34aedfee879" (UID: "c21fb2d2-38b1-41a0-8df8-d34aedfee879"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 21:10:23 crc kubenswrapper[4754]: I1005 21:10:23.424159 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c21fb2d2-38b1-41a0-8df8-d34aedfee879-config" (OuterVolumeSpecName: "config") pod "c21fb2d2-38b1-41a0-8df8-d34aedfee879" (UID: "c21fb2d2-38b1-41a0-8df8-d34aedfee879"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 21:10:23 crc kubenswrapper[4754]: I1005 21:10:23.424232 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22f48eb7-09d6-44bc-bc47-ba7026cf4cc3-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "22f48eb7-09d6-44bc-bc47-ba7026cf4cc3" (UID: "22f48eb7-09d6-44bc-bc47-ba7026cf4cc3"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 21:10:23 crc kubenswrapper[4754]: I1005 21:10:23.433766 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22f48eb7-09d6-44bc-bc47-ba7026cf4cc3-kube-api-access-dts6z" (OuterVolumeSpecName: "kube-api-access-dts6z") pod "22f48eb7-09d6-44bc-bc47-ba7026cf4cc3" (UID: "22f48eb7-09d6-44bc-bc47-ba7026cf4cc3"). InnerVolumeSpecName "kube-api-access-dts6z". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:10:23 crc kubenswrapper[4754]: I1005 21:10:23.433814 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c21fb2d2-38b1-41a0-8df8-d34aedfee879-kube-api-access-tdg46" (OuterVolumeSpecName: "kube-api-access-tdg46") pod "c21fb2d2-38b1-41a0-8df8-d34aedfee879" (UID: "c21fb2d2-38b1-41a0-8df8-d34aedfee879"). InnerVolumeSpecName "kube-api-access-tdg46". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:10:23 crc kubenswrapper[4754]: I1005 21:10:23.530595 4754 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/22f48eb7-09d6-44bc-bc47-ba7026cf4cc3-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 05 21:10:23 crc kubenswrapper[4754]: I1005 21:10:23.530627 4754 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c21fb2d2-38b1-41a0-8df8-d34aedfee879-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 05 21:10:23 crc kubenswrapper[4754]: I1005 21:10:23.530637 4754 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22f48eb7-09d6-44bc-bc47-ba7026cf4cc3-config\") on node \"crc\" DevicePath \"\"" Oct 05 21:10:23 crc kubenswrapper[4754]: I1005 21:10:23.530645 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tdg46\" (UniqueName: \"kubernetes.io/projected/c21fb2d2-38b1-41a0-8df8-d34aedfee879-kube-api-access-tdg46\") on node \"crc\" DevicePath \"\"" Oct 05 21:10:23 crc kubenswrapper[4754]: I1005 21:10:23.530657 4754 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c21fb2d2-38b1-41a0-8df8-d34aedfee879-config\") on node \"crc\" DevicePath \"\"" Oct 05 21:10:23 crc kubenswrapper[4754]: I1005 21:10:23.530665 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dts6z\" (UniqueName: \"kubernetes.io/projected/22f48eb7-09d6-44bc-bc47-ba7026cf4cc3-kube-api-access-dts6z\") on node \"crc\" DevicePath \"\"" Oct 05 21:10:23 crc kubenswrapper[4754]: I1005 21:10:23.814554 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-wfrrs" event={"ID":"c21fb2d2-38b1-41a0-8df8-d34aedfee879","Type":"ContainerDied","Data":"4fa73152575258c410743c0061bcd9f156a2333f04bcf31056053bdc8b58743b"} Oct 05 21:10:23 crc kubenswrapper[4754]: I1005 21:10:23.814582 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-wfrrs" Oct 05 21:10:23 crc kubenswrapper[4754]: I1005 21:10:23.817240 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"a5fdaf77-8a4c-4267-a32e-22a1c6be8f97","Type":"ContainerStarted","Data":"889ddb0166b138db89f77d95c243b2d378127d42b27fc304131c4c010b708d99"} Oct 05 21:10:23 crc kubenswrapper[4754]: I1005 21:10:23.831196 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-d5rsv" event={"ID":"22f48eb7-09d6-44bc-bc47-ba7026cf4cc3","Type":"ContainerDied","Data":"d36c72ea046cc25ee3cebcab4fdf3d09ac2f457fe2bf71446d7f8e45626f3854"} Oct 05 21:10:23 crc kubenswrapper[4754]: I1005 21:10:23.831283 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-d5rsv" Oct 05 21:10:23 crc kubenswrapper[4754]: I1005 21:10:23.839521 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"3c38a205-1079-46c2-9db5-c380d19399e3","Type":"ContainerStarted","Data":"86615b1181aeb708568db5d6446c880bf2412f9ee7937289cd68df5422717b49"} Oct 05 21:10:23 crc kubenswrapper[4754]: I1005 21:10:23.899146 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-wfrrs"] Oct 05 21:10:23 crc kubenswrapper[4754]: I1005 21:10:23.908574 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-wfrrs"] Oct 05 21:10:23 crc kubenswrapper[4754]: I1005 21:10:23.924639 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-d5rsv"] Oct 05 21:10:23 crc kubenswrapper[4754]: I1005 21:10:23.929363 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-d5rsv"] Oct 05 21:10:24 crc kubenswrapper[4754]: I1005 21:10:24.008013 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-8xfw8"] Oct 05 21:10:24 crc kubenswrapper[4754]: I1005 21:10:24.236841 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-r8m5k"] Oct 05 21:10:24 crc kubenswrapper[4754]: I1005 21:10:24.249137 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-hktv6"] Oct 05 21:10:24 crc kubenswrapper[4754]: W1005 21:10:24.281947 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb2dadb10_4ed0_4032_a049_a5b0289bd893.slice/crio-04a829472f4a9d43f8489bfe20ce1bf2de8a34f1be7a476042e9a26af5dc62a4 WatchSource:0}: Error finding container 04a829472f4a9d43f8489bfe20ce1bf2de8a34f1be7a476042e9a26af5dc62a4: Status 404 returned error can't find the container with id 04a829472f4a9d43f8489bfe20ce1bf2de8a34f1be7a476042e9a26af5dc62a4 Oct 05 21:10:24 crc kubenswrapper[4754]: I1005 21:10:24.846315 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22f48eb7-09d6-44bc-bc47-ba7026cf4cc3" path="/var/lib/kubelet/pods/22f48eb7-09d6-44bc-bc47-ba7026cf4cc3/volumes" Oct 05 21:10:24 crc kubenswrapper[4754]: I1005 21:10:24.846884 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c21fb2d2-38b1-41a0-8df8-d34aedfee879" path="/var/lib/kubelet/pods/c21fb2d2-38b1-41a0-8df8-d34aedfee879/volumes" Oct 05 21:10:24 crc kubenswrapper[4754]: I1005 21:10:24.849223 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-8xfw8" event={"ID":"06242eb6-ddde-49ea-b4a1-c61aad6f6402","Type":"ContainerStarted","Data":"e6f8e15ae38f6f8bf44a20fc87839e7b6d1e5c14f3132ef1b8b8eff9f769784d"} Oct 05 21:10:24 crc kubenswrapper[4754]: I1005 21:10:24.851646 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-r8m5k" event={"ID":"a40e02a7-900e-4da2-8a4a-73773ca90876","Type":"ContainerStarted","Data":"4426e1a6008d7785f3df3b9ede9af06a0344676ecb05f3f03ffd1e742721e76a"} Oct 05 21:10:24 crc kubenswrapper[4754]: I1005 21:10:24.854932 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"280f4e00-063d-486e-bd58-f4b246f54814","Type":"ContainerStarted","Data":"a59a9eb09e684bcff101063ecb5b4c6e28d60f5edeefe13b2f5db1d3257258be"} Oct 05 21:10:24 crc kubenswrapper[4754]: I1005 21:10:24.855387 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Oct 05 21:10:24 crc kubenswrapper[4754]: I1005 21:10:24.856545 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-hktv6" event={"ID":"b2dadb10-4ed0-4032-a049-a5b0289bd893","Type":"ContainerStarted","Data":"04a829472f4a9d43f8489bfe20ce1bf2de8a34f1be7a476042e9a26af5dc62a4"} Oct 05 21:10:24 crc kubenswrapper[4754]: I1005 21:10:24.917101 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.939830217 podStartE2EDuration="40.91708173s" podCreationTimestamp="2025-10-05 21:09:44 +0000 UTC" firstStartedPulling="2025-10-05 21:09:45.675917315 +0000 UTC m=+909.580036025" lastFinishedPulling="2025-10-05 21:10:23.653168828 +0000 UTC m=+947.557287538" observedRunningTime="2025-10-05 21:10:24.892581548 +0000 UTC m=+948.796700258" watchObservedRunningTime="2025-10-05 21:10:24.91708173 +0000 UTC m=+948.821200440" Oct 05 21:10:25 crc kubenswrapper[4754]: I1005 21:10:25.867123 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"1d96e280-8c26-41ea-ab29-64ce5b9c6a2c","Type":"ContainerStarted","Data":"1a0e1b3cbb7432cf1c8d8d5b93db680fd2af14a635594bd4eb551779f7e4ba30"} Oct 05 21:10:25 crc kubenswrapper[4754]: I1005 21:10:25.874598 4754 generic.go:334] "Generic (PLEG): container finished" podID="a40e02a7-900e-4da2-8a4a-73773ca90876" containerID="d0afd180d66849b57e27b8e3892816da0c5c63ff7bfef21d39cd5bbb23b8b76a" exitCode=0 Oct 05 21:10:25 crc kubenswrapper[4754]: I1005 21:10:25.874677 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-r8m5k" event={"ID":"a40e02a7-900e-4da2-8a4a-73773ca90876","Type":"ContainerDied","Data":"d0afd180d66849b57e27b8e3892816da0c5c63ff7bfef21d39cd5bbb23b8b76a"} Oct 05 21:10:25 crc kubenswrapper[4754]: I1005 21:10:25.879968 4754 generic.go:334] "Generic (PLEG): container finished" podID="b2dadb10-4ed0-4032-a049-a5b0289bd893" containerID="78b96b0e102e5a8075322c252ee856bd3546b2bdb59a2cdc9731f19b2523ffcf" exitCode=0 Oct 05 21:10:25 crc kubenswrapper[4754]: I1005 21:10:25.881136 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-hktv6" event={"ID":"b2dadb10-4ed0-4032-a049-a5b0289bd893","Type":"ContainerDied","Data":"78b96b0e102e5a8075322c252ee856bd3546b2bdb59a2cdc9731f19b2523ffcf"} Oct 05 21:10:26 crc kubenswrapper[4754]: I1005 21:10:26.897667 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"39aff8a0-2905-4ad3-ad2d-5bf6b5de3858","Type":"ContainerStarted","Data":"493f2a671031a6ad481270c980a4f16b3d2567e2e2da7677960e8f9fba245eaf"} Oct 05 21:10:26 crc kubenswrapper[4754]: I1005 21:10:26.901461 4754 generic.go:334] "Generic (PLEG): container finished" podID="d311946e-91c6-4be6-9d48-1b823ee77607" containerID="ec56f5e8176d9ae61641d3a197a6391d110b4913ff0b634a91334b4e35414421" exitCode=0 Oct 05 21:10:26 crc kubenswrapper[4754]: I1005 21:10:26.901583 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"d311946e-91c6-4be6-9d48-1b823ee77607","Type":"ContainerDied","Data":"ec56f5e8176d9ae61641d3a197a6391d110b4913ff0b634a91334b4e35414421"} Oct 05 21:10:26 crc kubenswrapper[4754]: I1005 21:10:26.906673 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-hktv6" event={"ID":"b2dadb10-4ed0-4032-a049-a5b0289bd893","Type":"ContainerStarted","Data":"3d10e4b14d4ea32038244ea1beb387d171d0c8d9f7fb741b2931dc1753c7d7fa"} Oct 05 21:10:26 crc kubenswrapper[4754]: I1005 21:10:26.906767 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7fd796d7df-hktv6" Oct 05 21:10:26 crc kubenswrapper[4754]: I1005 21:10:26.988240 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7fd796d7df-hktv6" podStartSLOduration=4.362493004 podStartE2EDuration="4.988214215s" podCreationTimestamp="2025-10-05 21:10:22 +0000 UTC" firstStartedPulling="2025-10-05 21:10:24.286155561 +0000 UTC m=+948.190274271" lastFinishedPulling="2025-10-05 21:10:24.911876772 +0000 UTC m=+948.815995482" observedRunningTime="2025-10-05 21:10:26.983408517 +0000 UTC m=+950.887527227" watchObservedRunningTime="2025-10-05 21:10:26.988214215 +0000 UTC m=+950.892332925" Oct 05 21:10:29 crc kubenswrapper[4754]: I1005 21:10:29.939317 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-r8m5k" event={"ID":"a40e02a7-900e-4da2-8a4a-73773ca90876","Type":"ContainerStarted","Data":"20598aa4e239724712ce00389212a2b4a21ab17443b6890e775b769028651f20"} Oct 05 21:10:29 crc kubenswrapper[4754]: I1005 21:10:29.939830 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-86db49b7ff-r8m5k" Oct 05 21:10:29 crc kubenswrapper[4754]: I1005 21:10:29.962569 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-86db49b7ff-r8m5k" podStartSLOduration=7.322216308 podStartE2EDuration="7.962542607s" podCreationTimestamp="2025-10-05 21:10:22 +0000 UTC" firstStartedPulling="2025-10-05 21:10:24.30563496 +0000 UTC m=+948.209753670" lastFinishedPulling="2025-10-05 21:10:24.945961259 +0000 UTC m=+948.850079969" observedRunningTime="2025-10-05 21:10:29.959944918 +0000 UTC m=+953.864063638" watchObservedRunningTime="2025-10-05 21:10:29.962542607 +0000 UTC m=+953.866661357" Oct 05 21:10:32 crc kubenswrapper[4754]: I1005 21:10:32.883974 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7fd796d7df-hktv6" Oct 05 21:10:34 crc kubenswrapper[4754]: I1005 21:10:34.886295 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Oct 05 21:10:35 crc kubenswrapper[4754]: I1005 21:10:35.245075 4754 patch_prober.go:28] interesting pod/machine-config-daemon-b2h9k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 21:10:35 crc kubenswrapper[4754]: I1005 21:10:35.245705 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 21:10:35 crc kubenswrapper[4754]: I1005 21:10:35.245817 4754 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" Oct 05 21:10:35 crc kubenswrapper[4754]: I1005 21:10:35.246818 4754 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"204a9d4b26fceb896f0714f752c6bee1970d2349381ac50fae1a7322b50bafde"} pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 05 21:10:35 crc kubenswrapper[4754]: I1005 21:10:35.246956 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" containerName="machine-config-daemon" containerID="cri-o://204a9d4b26fceb896f0714f752c6bee1970d2349381ac50fae1a7322b50bafde" gracePeriod=600 Oct 05 21:10:36 crc kubenswrapper[4754]: I1005 21:10:36.001368 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"3c38a205-1079-46c2-9db5-c380d19399e3","Type":"ContainerStarted","Data":"60feb1bc566918e063c4e266fea59262b62b8f72e26458b2eb484c30cbfa7ff0"} Oct 05 21:10:36 crc kubenswrapper[4754]: I1005 21:10:36.009509 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-2krf2" event={"ID":"1283a11a-7581-4146-87db-b62d6d5d6dc9","Type":"ContainerStarted","Data":"94f1da06b814c982d33e6eac19a7121191720cac9c406b9ab10a6076346c2125"} Oct 05 21:10:36 crc kubenswrapper[4754]: I1005 21:10:36.034536 4754 generic.go:334] "Generic (PLEG): container finished" podID="be95a413-6a59-45b4-84b7-b43cae694a26" containerID="204a9d4b26fceb896f0714f752c6bee1970d2349381ac50fae1a7322b50bafde" exitCode=0 Oct 05 21:10:36 crc kubenswrapper[4754]: I1005 21:10:36.034679 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" event={"ID":"be95a413-6a59-45b4-84b7-b43cae694a26","Type":"ContainerDied","Data":"204a9d4b26fceb896f0714f752c6bee1970d2349381ac50fae1a7322b50bafde"} Oct 05 21:10:36 crc kubenswrapper[4754]: I1005 21:10:36.034726 4754 scope.go:117] "RemoveContainer" containerID="42efe442464282aff7d7a4c1a4006bea20a662c700019a0224e31e9c8b2dfd60" Oct 05 21:10:36 crc kubenswrapper[4754]: I1005 21:10:36.056837 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"a5fdaf77-8a4c-4267-a32e-22a1c6be8f97","Type":"ContainerStarted","Data":"2e2056f3922fcc6808f97d6159c9946b95e8d78566bcd385564a121cc66cca51"} Oct 05 21:10:36 crc kubenswrapper[4754]: I1005 21:10:36.070948 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"d311946e-91c6-4be6-9d48-1b823ee77607","Type":"ContainerStarted","Data":"225856772abd5ea580a1d0118fc14079c085d90fce5f12bcea9f581185156f2e"} Oct 05 21:10:36 crc kubenswrapper[4754]: I1005 21:10:36.074559 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"d97f6b34-ef35-4b21-81eb-9e7009945fa6","Type":"ContainerStarted","Data":"2d60b8bdb35899756b0e913b46e6dd512e340ee387b5ede9da9c574a7066012b"} Oct 05 21:10:36 crc kubenswrapper[4754]: I1005 21:10:36.103663 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=18.658615037 podStartE2EDuration="56.10363905s" podCreationTimestamp="2025-10-05 21:09:40 +0000 UTC" firstStartedPulling="2025-10-05 21:09:42.774914364 +0000 UTC m=+906.679033074" lastFinishedPulling="2025-10-05 21:10:20.219938377 +0000 UTC m=+944.124057087" observedRunningTime="2025-10-05 21:10:36.096775277 +0000 UTC m=+960.000893987" watchObservedRunningTime="2025-10-05 21:10:36.10363905 +0000 UTC m=+960.007757760" Oct 05 21:10:37 crc kubenswrapper[4754]: I1005 21:10:37.085331 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-8xfw8" event={"ID":"06242eb6-ddde-49ea-b4a1-c61aad6f6402","Type":"ContainerStarted","Data":"b0f9e5623f85fda6aca72690d34990121f461da121b69e663451fc7469044fb5"} Oct 05 21:10:37 crc kubenswrapper[4754]: I1005 21:10:37.088716 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"3c38a205-1079-46c2-9db5-c380d19399e3","Type":"ContainerStarted","Data":"ce67bd43b2a59e998e63bedda6c6a8dc28afdf39fc8a841cb1f6a1b4fe138629"} Oct 05 21:10:37 crc kubenswrapper[4754]: I1005 21:10:37.093840 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-pp728" event={"ID":"e3785031-57bd-44ab-b30e-b609cca6dcaf","Type":"ContainerStarted","Data":"441d07b31fb450cf4f4ce8c1295da6f9c2045d7b642699fbc2af5ad5ab56b70e"} Oct 05 21:10:37 crc kubenswrapper[4754]: I1005 21:10:37.094061 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-pp728" Oct 05 21:10:37 crc kubenswrapper[4754]: I1005 21:10:37.095630 4754 generic.go:334] "Generic (PLEG): container finished" podID="1283a11a-7581-4146-87db-b62d6d5d6dc9" containerID="94f1da06b814c982d33e6eac19a7121191720cac9c406b9ab10a6076346c2125" exitCode=0 Oct 05 21:10:37 crc kubenswrapper[4754]: I1005 21:10:37.095683 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-2krf2" event={"ID":"1283a11a-7581-4146-87db-b62d6d5d6dc9","Type":"ContainerDied","Data":"94f1da06b814c982d33e6eac19a7121191720cac9c406b9ab10a6076346c2125"} Oct 05 21:10:37 crc kubenswrapper[4754]: I1005 21:10:37.106468 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-8xfw8" podStartSLOduration=2.742056719 podStartE2EDuration="15.106442859s" podCreationTimestamp="2025-10-05 21:10:22 +0000 UTC" firstStartedPulling="2025-10-05 21:10:24.052384207 +0000 UTC m=+947.956502917" lastFinishedPulling="2025-10-05 21:10:36.416770357 +0000 UTC m=+960.320889057" observedRunningTime="2025-10-05 21:10:37.105175365 +0000 UTC m=+961.009294075" watchObservedRunningTime="2025-10-05 21:10:37.106442859 +0000 UTC m=+961.010561589" Oct 05 21:10:37 crc kubenswrapper[4754]: I1005 21:10:37.144363 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=35.067337024 podStartE2EDuration="48.144344728s" podCreationTimestamp="2025-10-05 21:09:49 +0000 UTC" firstStartedPulling="2025-10-05 21:10:23.495242023 +0000 UTC m=+947.399360733" lastFinishedPulling="2025-10-05 21:10:36.572249727 +0000 UTC m=+960.476368437" observedRunningTime="2025-10-05 21:10:37.141584265 +0000 UTC m=+961.045703005" watchObservedRunningTime="2025-10-05 21:10:37.144344728 +0000 UTC m=+961.048463438" Oct 05 21:10:37 crc kubenswrapper[4754]: I1005 21:10:37.145716 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" event={"ID":"be95a413-6a59-45b4-84b7-b43cae694a26","Type":"ContainerStarted","Data":"03710aff934f922f0a7167007c25b57fd90897b7f134dd149dd0d1c75cb91da2"} Oct 05 21:10:37 crc kubenswrapper[4754]: I1005 21:10:37.150575 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"a5fdaf77-8a4c-4267-a32e-22a1c6be8f97","Type":"ContainerStarted","Data":"058093902ae0c84b386ee0cc23773b3cbe9950962dbab3756d47b8b619d7fbc2"} Oct 05 21:10:37 crc kubenswrapper[4754]: I1005 21:10:37.164007 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"bdbca489-8f9b-420e-a306-65575175af99","Type":"ContainerStarted","Data":"4b22530e2d983ce5fa22b69ac5a5d8ba554298cd4c085abdac54b1795c01282a"} Oct 05 21:10:37 crc kubenswrapper[4754]: I1005 21:10:37.164915 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Oct 05 21:10:37 crc kubenswrapper[4754]: I1005 21:10:37.209869 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-pp728" podStartSLOduration=34.028740927 podStartE2EDuration="49.209850583s" podCreationTimestamp="2025-10-05 21:09:48 +0000 UTC" firstStartedPulling="2025-10-05 21:10:21.142980843 +0000 UTC m=+945.047099553" lastFinishedPulling="2025-10-05 21:10:36.324090499 +0000 UTC m=+960.228209209" observedRunningTime="2025-10-05 21:10:37.16470515 +0000 UTC m=+961.068823870" watchObservedRunningTime="2025-10-05 21:10:37.209850583 +0000 UTC m=+961.113969293" Oct 05 21:10:37 crc kubenswrapper[4754]: I1005 21:10:37.240532 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=3.617669523 podStartE2EDuration="55.240513509s" podCreationTimestamp="2025-10-05 21:09:42 +0000 UTC" firstStartedPulling="2025-10-05 21:09:44.798964985 +0000 UTC m=+908.703083685" lastFinishedPulling="2025-10-05 21:10:36.421808961 +0000 UTC m=+960.325927671" observedRunningTime="2025-10-05 21:10:37.234880759 +0000 UTC m=+961.138999479" watchObservedRunningTime="2025-10-05 21:10:37.240513509 +0000 UTC m=+961.144632209" Oct 05 21:10:37 crc kubenswrapper[4754]: I1005 21:10:37.287525 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=37.365299877 podStartE2EDuration="50.28750409s" podCreationTimestamp="2025-10-05 21:09:47 +0000 UTC" firstStartedPulling="2025-10-05 21:10:23.495639343 +0000 UTC m=+947.399758053" lastFinishedPulling="2025-10-05 21:10:36.417843556 +0000 UTC m=+960.321962266" observedRunningTime="2025-10-05 21:10:37.28713294 +0000 UTC m=+961.191251660" watchObservedRunningTime="2025-10-05 21:10:37.28750409 +0000 UTC m=+961.191622800" Oct 05 21:10:38 crc kubenswrapper[4754]: I1005 21:10:38.009698 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-86db49b7ff-r8m5k" Oct 05 21:10:38 crc kubenswrapper[4754]: I1005 21:10:38.083556 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-hktv6"] Oct 05 21:10:38 crc kubenswrapper[4754]: I1005 21:10:38.083827 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7fd796d7df-hktv6" podUID="b2dadb10-4ed0-4032-a049-a5b0289bd893" containerName="dnsmasq-dns" containerID="cri-o://3d10e4b14d4ea32038244ea1beb387d171d0c8d9f7fb741b2931dc1753c7d7fa" gracePeriod=10 Oct 05 21:10:38 crc kubenswrapper[4754]: I1005 21:10:38.177909 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-2krf2" event={"ID":"1283a11a-7581-4146-87db-b62d6d5d6dc9","Type":"ContainerStarted","Data":"880d9e6d10934617784bc717bcda0dcc252f30d0b2d6f9f56dd18e2e408fea11"} Oct 05 21:10:38 crc kubenswrapper[4754]: I1005 21:10:38.177950 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-2krf2" event={"ID":"1283a11a-7581-4146-87db-b62d6d5d6dc9","Type":"ContainerStarted","Data":"f9fc1277423db60142a734357f7dbc10326be5d0106abb9ecc72222c4470ba6a"} Oct 05 21:10:38 crc kubenswrapper[4754]: I1005 21:10:38.179751 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-2krf2" Oct 05 21:10:38 crc kubenswrapper[4754]: I1005 21:10:38.179776 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-2krf2" Oct 05 21:10:38 crc kubenswrapper[4754]: I1005 21:10:38.247253 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-2krf2" podStartSLOduration=36.617580346 podStartE2EDuration="50.247226213s" podCreationTimestamp="2025-10-05 21:09:48 +0000 UTC" firstStartedPulling="2025-10-05 21:10:21.364363328 +0000 UTC m=+945.268482038" lastFinishedPulling="2025-10-05 21:10:34.994009195 +0000 UTC m=+958.898127905" observedRunningTime="2025-10-05 21:10:38.21706921 +0000 UTC m=+962.121187920" watchObservedRunningTime="2025-10-05 21:10:38.247226213 +0000 UTC m=+962.151344923" Oct 05 21:10:38 crc kubenswrapper[4754]: I1005 21:10:38.759423 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-hktv6" Oct 05 21:10:38 crc kubenswrapper[4754]: I1005 21:10:38.800049 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ch2t7\" (UniqueName: \"kubernetes.io/projected/b2dadb10-4ed0-4032-a049-a5b0289bd893-kube-api-access-ch2t7\") pod \"b2dadb10-4ed0-4032-a049-a5b0289bd893\" (UID: \"b2dadb10-4ed0-4032-a049-a5b0289bd893\") " Oct 05 21:10:38 crc kubenswrapper[4754]: I1005 21:10:38.800107 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b2dadb10-4ed0-4032-a049-a5b0289bd893-config\") pod \"b2dadb10-4ed0-4032-a049-a5b0289bd893\" (UID: \"b2dadb10-4ed0-4032-a049-a5b0289bd893\") " Oct 05 21:10:38 crc kubenswrapper[4754]: I1005 21:10:38.800227 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b2dadb10-4ed0-4032-a049-a5b0289bd893-dns-svc\") pod \"b2dadb10-4ed0-4032-a049-a5b0289bd893\" (UID: \"b2dadb10-4ed0-4032-a049-a5b0289bd893\") " Oct 05 21:10:38 crc kubenswrapper[4754]: I1005 21:10:38.800276 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b2dadb10-4ed0-4032-a049-a5b0289bd893-ovsdbserver-nb\") pod \"b2dadb10-4ed0-4032-a049-a5b0289bd893\" (UID: \"b2dadb10-4ed0-4032-a049-a5b0289bd893\") " Oct 05 21:10:38 crc kubenswrapper[4754]: I1005 21:10:38.811035 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b2dadb10-4ed0-4032-a049-a5b0289bd893-kube-api-access-ch2t7" (OuterVolumeSpecName: "kube-api-access-ch2t7") pod "b2dadb10-4ed0-4032-a049-a5b0289bd893" (UID: "b2dadb10-4ed0-4032-a049-a5b0289bd893"). InnerVolumeSpecName "kube-api-access-ch2t7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:10:38 crc kubenswrapper[4754]: I1005 21:10:38.851815 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b2dadb10-4ed0-4032-a049-a5b0289bd893-config" (OuterVolumeSpecName: "config") pod "b2dadb10-4ed0-4032-a049-a5b0289bd893" (UID: "b2dadb10-4ed0-4032-a049-a5b0289bd893"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 21:10:38 crc kubenswrapper[4754]: I1005 21:10:38.851861 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b2dadb10-4ed0-4032-a049-a5b0289bd893-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b2dadb10-4ed0-4032-a049-a5b0289bd893" (UID: "b2dadb10-4ed0-4032-a049-a5b0289bd893"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 21:10:38 crc kubenswrapper[4754]: I1005 21:10:38.893436 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b2dadb10-4ed0-4032-a049-a5b0289bd893-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "b2dadb10-4ed0-4032-a049-a5b0289bd893" (UID: "b2dadb10-4ed0-4032-a049-a5b0289bd893"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 21:10:38 crc kubenswrapper[4754]: I1005 21:10:38.902682 4754 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b2dadb10-4ed0-4032-a049-a5b0289bd893-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 05 21:10:38 crc kubenswrapper[4754]: I1005 21:10:38.902709 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ch2t7\" (UniqueName: \"kubernetes.io/projected/b2dadb10-4ed0-4032-a049-a5b0289bd893-kube-api-access-ch2t7\") on node \"crc\" DevicePath \"\"" Oct 05 21:10:38 crc kubenswrapper[4754]: I1005 21:10:38.902724 4754 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b2dadb10-4ed0-4032-a049-a5b0289bd893-config\") on node \"crc\" DevicePath \"\"" Oct 05 21:10:38 crc kubenswrapper[4754]: I1005 21:10:38.902734 4754 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b2dadb10-4ed0-4032-a049-a5b0289bd893-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 05 21:10:39 crc kubenswrapper[4754]: I1005 21:10:39.025328 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Oct 05 21:10:39 crc kubenswrapper[4754]: I1005 21:10:39.185060 4754 generic.go:334] "Generic (PLEG): container finished" podID="b2dadb10-4ed0-4032-a049-a5b0289bd893" containerID="3d10e4b14d4ea32038244ea1beb387d171d0c8d9f7fb741b2931dc1753c7d7fa" exitCode=0 Oct 05 21:10:39 crc kubenswrapper[4754]: I1005 21:10:39.185883 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-hktv6" Oct 05 21:10:39 crc kubenswrapper[4754]: I1005 21:10:39.190544 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-hktv6" event={"ID":"b2dadb10-4ed0-4032-a049-a5b0289bd893","Type":"ContainerDied","Data":"3d10e4b14d4ea32038244ea1beb387d171d0c8d9f7fb741b2931dc1753c7d7fa"} Oct 05 21:10:39 crc kubenswrapper[4754]: I1005 21:10:39.190578 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-hktv6" event={"ID":"b2dadb10-4ed0-4032-a049-a5b0289bd893","Type":"ContainerDied","Data":"04a829472f4a9d43f8489bfe20ce1bf2de8a34f1be7a476042e9a26af5dc62a4"} Oct 05 21:10:39 crc kubenswrapper[4754]: I1005 21:10:39.190597 4754 scope.go:117] "RemoveContainer" containerID="3d10e4b14d4ea32038244ea1beb387d171d0c8d9f7fb741b2931dc1753c7d7fa" Oct 05 21:10:39 crc kubenswrapper[4754]: I1005 21:10:39.242969 4754 scope.go:117] "RemoveContainer" containerID="78b96b0e102e5a8075322c252ee856bd3546b2bdb59a2cdc9731f19b2523ffcf" Oct 05 21:10:39 crc kubenswrapper[4754]: I1005 21:10:39.254596 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-hktv6"] Oct 05 21:10:39 crc kubenswrapper[4754]: I1005 21:10:39.263747 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-hktv6"] Oct 05 21:10:39 crc kubenswrapper[4754]: I1005 21:10:39.282284 4754 scope.go:117] "RemoveContainer" containerID="3d10e4b14d4ea32038244ea1beb387d171d0c8d9f7fb741b2931dc1753c7d7fa" Oct 05 21:10:39 crc kubenswrapper[4754]: E1005 21:10:39.286763 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3d10e4b14d4ea32038244ea1beb387d171d0c8d9f7fb741b2931dc1753c7d7fa\": container with ID starting with 3d10e4b14d4ea32038244ea1beb387d171d0c8d9f7fb741b2931dc1753c7d7fa not found: ID does not exist" containerID="3d10e4b14d4ea32038244ea1beb387d171d0c8d9f7fb741b2931dc1753c7d7fa" Oct 05 21:10:39 crc kubenswrapper[4754]: I1005 21:10:39.286808 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3d10e4b14d4ea32038244ea1beb387d171d0c8d9f7fb741b2931dc1753c7d7fa"} err="failed to get container status \"3d10e4b14d4ea32038244ea1beb387d171d0c8d9f7fb741b2931dc1753c7d7fa\": rpc error: code = NotFound desc = could not find container \"3d10e4b14d4ea32038244ea1beb387d171d0c8d9f7fb741b2931dc1753c7d7fa\": container with ID starting with 3d10e4b14d4ea32038244ea1beb387d171d0c8d9f7fb741b2931dc1753c7d7fa not found: ID does not exist" Oct 05 21:10:39 crc kubenswrapper[4754]: I1005 21:10:39.286836 4754 scope.go:117] "RemoveContainer" containerID="78b96b0e102e5a8075322c252ee856bd3546b2bdb59a2cdc9731f19b2523ffcf" Oct 05 21:10:39 crc kubenswrapper[4754]: E1005 21:10:39.287106 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"78b96b0e102e5a8075322c252ee856bd3546b2bdb59a2cdc9731f19b2523ffcf\": container with ID starting with 78b96b0e102e5a8075322c252ee856bd3546b2bdb59a2cdc9731f19b2523ffcf not found: ID does not exist" containerID="78b96b0e102e5a8075322c252ee856bd3546b2bdb59a2cdc9731f19b2523ffcf" Oct 05 21:10:39 crc kubenswrapper[4754]: I1005 21:10:39.287127 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"78b96b0e102e5a8075322c252ee856bd3546b2bdb59a2cdc9731f19b2523ffcf"} err="failed to get container status \"78b96b0e102e5a8075322c252ee856bd3546b2bdb59a2cdc9731f19b2523ffcf\": rpc error: code = NotFound desc = could not find container \"78b96b0e102e5a8075322c252ee856bd3546b2bdb59a2cdc9731f19b2523ffcf\": container with ID starting with 78b96b0e102e5a8075322c252ee856bd3546b2bdb59a2cdc9731f19b2523ffcf not found: ID does not exist" Oct 05 21:10:39 crc kubenswrapper[4754]: I1005 21:10:39.371569 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Oct 05 21:10:39 crc kubenswrapper[4754]: I1005 21:10:39.434040 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Oct 05 21:10:40 crc kubenswrapper[4754]: I1005 21:10:40.024767 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Oct 05 21:10:40 crc kubenswrapper[4754]: I1005 21:10:40.065472 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Oct 05 21:10:40 crc kubenswrapper[4754]: I1005 21:10:40.196342 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Oct 05 21:10:40 crc kubenswrapper[4754]: I1005 21:10:40.240669 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Oct 05 21:10:40 crc kubenswrapper[4754]: I1005 21:10:40.244078 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Oct 05 21:10:40 crc kubenswrapper[4754]: I1005 21:10:40.561943 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Oct 05 21:10:40 crc kubenswrapper[4754]: E1005 21:10:40.562719 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2dadb10-4ed0-4032-a049-a5b0289bd893" containerName="dnsmasq-dns" Oct 05 21:10:40 crc kubenswrapper[4754]: I1005 21:10:40.562743 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2dadb10-4ed0-4032-a049-a5b0289bd893" containerName="dnsmasq-dns" Oct 05 21:10:40 crc kubenswrapper[4754]: E1005 21:10:40.562759 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2dadb10-4ed0-4032-a049-a5b0289bd893" containerName="init" Oct 05 21:10:40 crc kubenswrapper[4754]: I1005 21:10:40.562764 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2dadb10-4ed0-4032-a049-a5b0289bd893" containerName="init" Oct 05 21:10:40 crc kubenswrapper[4754]: I1005 21:10:40.562939 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="b2dadb10-4ed0-4032-a049-a5b0289bd893" containerName="dnsmasq-dns" Oct 05 21:10:40 crc kubenswrapper[4754]: I1005 21:10:40.565267 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 05 21:10:40 crc kubenswrapper[4754]: I1005 21:10:40.568196 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-8f9j9" Oct 05 21:10:40 crc kubenswrapper[4754]: I1005 21:10:40.568412 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Oct 05 21:10:40 crc kubenswrapper[4754]: I1005 21:10:40.570937 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Oct 05 21:10:40 crc kubenswrapper[4754]: I1005 21:10:40.574650 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Oct 05 21:10:40 crc kubenswrapper[4754]: I1005 21:10:40.585539 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Oct 05 21:10:40 crc kubenswrapper[4754]: I1005 21:10:40.731285 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/316121e1-b760-4eaf-9b8e-ec5bacc6b117-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"316121e1-b760-4eaf-9b8e-ec5bacc6b117\") " pod="openstack/ovn-northd-0" Oct 05 21:10:40 crc kubenswrapper[4754]: I1005 21:10:40.731345 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/316121e1-b760-4eaf-9b8e-ec5bacc6b117-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"316121e1-b760-4eaf-9b8e-ec5bacc6b117\") " pod="openstack/ovn-northd-0" Oct 05 21:10:40 crc kubenswrapper[4754]: I1005 21:10:40.731370 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-chqww\" (UniqueName: \"kubernetes.io/projected/316121e1-b760-4eaf-9b8e-ec5bacc6b117-kube-api-access-chqww\") pod \"ovn-northd-0\" (UID: \"316121e1-b760-4eaf-9b8e-ec5bacc6b117\") " pod="openstack/ovn-northd-0" Oct 05 21:10:40 crc kubenswrapper[4754]: I1005 21:10:40.731433 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/316121e1-b760-4eaf-9b8e-ec5bacc6b117-scripts\") pod \"ovn-northd-0\" (UID: \"316121e1-b760-4eaf-9b8e-ec5bacc6b117\") " pod="openstack/ovn-northd-0" Oct 05 21:10:40 crc kubenswrapper[4754]: I1005 21:10:40.731684 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/316121e1-b760-4eaf-9b8e-ec5bacc6b117-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"316121e1-b760-4eaf-9b8e-ec5bacc6b117\") " pod="openstack/ovn-northd-0" Oct 05 21:10:40 crc kubenswrapper[4754]: I1005 21:10:40.731844 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/316121e1-b760-4eaf-9b8e-ec5bacc6b117-config\") pod \"ovn-northd-0\" (UID: \"316121e1-b760-4eaf-9b8e-ec5bacc6b117\") " pod="openstack/ovn-northd-0" Oct 05 21:10:40 crc kubenswrapper[4754]: I1005 21:10:40.731908 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/316121e1-b760-4eaf-9b8e-ec5bacc6b117-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"316121e1-b760-4eaf-9b8e-ec5bacc6b117\") " pod="openstack/ovn-northd-0" Oct 05 21:10:40 crc kubenswrapper[4754]: I1005 21:10:40.833778 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/316121e1-b760-4eaf-9b8e-ec5bacc6b117-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"316121e1-b760-4eaf-9b8e-ec5bacc6b117\") " pod="openstack/ovn-northd-0" Oct 05 21:10:40 crc kubenswrapper[4754]: I1005 21:10:40.833863 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/316121e1-b760-4eaf-9b8e-ec5bacc6b117-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"316121e1-b760-4eaf-9b8e-ec5bacc6b117\") " pod="openstack/ovn-northd-0" Oct 05 21:10:40 crc kubenswrapper[4754]: I1005 21:10:40.833894 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-chqww\" (UniqueName: \"kubernetes.io/projected/316121e1-b760-4eaf-9b8e-ec5bacc6b117-kube-api-access-chqww\") pod \"ovn-northd-0\" (UID: \"316121e1-b760-4eaf-9b8e-ec5bacc6b117\") " pod="openstack/ovn-northd-0" Oct 05 21:10:40 crc kubenswrapper[4754]: I1005 21:10:40.833942 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/316121e1-b760-4eaf-9b8e-ec5bacc6b117-scripts\") pod \"ovn-northd-0\" (UID: \"316121e1-b760-4eaf-9b8e-ec5bacc6b117\") " pod="openstack/ovn-northd-0" Oct 05 21:10:40 crc kubenswrapper[4754]: I1005 21:10:40.834010 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/316121e1-b760-4eaf-9b8e-ec5bacc6b117-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"316121e1-b760-4eaf-9b8e-ec5bacc6b117\") " pod="openstack/ovn-northd-0" Oct 05 21:10:40 crc kubenswrapper[4754]: I1005 21:10:40.835079 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/316121e1-b760-4eaf-9b8e-ec5bacc6b117-scripts\") pod \"ovn-northd-0\" (UID: \"316121e1-b760-4eaf-9b8e-ec5bacc6b117\") " pod="openstack/ovn-northd-0" Oct 05 21:10:40 crc kubenswrapper[4754]: I1005 21:10:40.835223 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/316121e1-b760-4eaf-9b8e-ec5bacc6b117-config\") pod \"ovn-northd-0\" (UID: \"316121e1-b760-4eaf-9b8e-ec5bacc6b117\") " pod="openstack/ovn-northd-0" Oct 05 21:10:40 crc kubenswrapper[4754]: I1005 21:10:40.835952 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/316121e1-b760-4eaf-9b8e-ec5bacc6b117-config\") pod \"ovn-northd-0\" (UID: \"316121e1-b760-4eaf-9b8e-ec5bacc6b117\") " pod="openstack/ovn-northd-0" Oct 05 21:10:40 crc kubenswrapper[4754]: I1005 21:10:40.836000 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/316121e1-b760-4eaf-9b8e-ec5bacc6b117-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"316121e1-b760-4eaf-9b8e-ec5bacc6b117\") " pod="openstack/ovn-northd-0" Oct 05 21:10:40 crc kubenswrapper[4754]: I1005 21:10:40.836360 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/316121e1-b760-4eaf-9b8e-ec5bacc6b117-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"316121e1-b760-4eaf-9b8e-ec5bacc6b117\") " pod="openstack/ovn-northd-0" Oct 05 21:10:40 crc kubenswrapper[4754]: I1005 21:10:40.842194 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/316121e1-b760-4eaf-9b8e-ec5bacc6b117-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"316121e1-b760-4eaf-9b8e-ec5bacc6b117\") " pod="openstack/ovn-northd-0" Oct 05 21:10:40 crc kubenswrapper[4754]: I1005 21:10:40.843409 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/316121e1-b760-4eaf-9b8e-ec5bacc6b117-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"316121e1-b760-4eaf-9b8e-ec5bacc6b117\") " pod="openstack/ovn-northd-0" Oct 05 21:10:40 crc kubenswrapper[4754]: I1005 21:10:40.850173 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/316121e1-b760-4eaf-9b8e-ec5bacc6b117-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"316121e1-b760-4eaf-9b8e-ec5bacc6b117\") " pod="openstack/ovn-northd-0" Oct 05 21:10:40 crc kubenswrapper[4754]: I1005 21:10:40.856072 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b2dadb10-4ed0-4032-a049-a5b0289bd893" path="/var/lib/kubelet/pods/b2dadb10-4ed0-4032-a049-a5b0289bd893/volumes" Oct 05 21:10:40 crc kubenswrapper[4754]: I1005 21:10:40.866172 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-chqww\" (UniqueName: \"kubernetes.io/projected/316121e1-b760-4eaf-9b8e-ec5bacc6b117-kube-api-access-chqww\") pod \"ovn-northd-0\" (UID: \"316121e1-b760-4eaf-9b8e-ec5bacc6b117\") " pod="openstack/ovn-northd-0" Oct 05 21:10:40 crc kubenswrapper[4754]: I1005 21:10:40.891253 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 05 21:10:41 crc kubenswrapper[4754]: I1005 21:10:41.177567 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Oct 05 21:10:41 crc kubenswrapper[4754]: I1005 21:10:41.207123 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"316121e1-b760-4eaf-9b8e-ec5bacc6b117","Type":"ContainerStarted","Data":"d5605ab4874971b52e1310bab78c9053ca50eb9c0da48a1a87855fa5767c5057"} Oct 05 21:10:41 crc kubenswrapper[4754]: I1005 21:10:41.213129 4754 generic.go:334] "Generic (PLEG): container finished" podID="d97f6b34-ef35-4b21-81eb-9e7009945fa6" containerID="2d60b8bdb35899756b0e913b46e6dd512e340ee387b5ede9da9c574a7066012b" exitCode=0 Oct 05 21:10:41 crc kubenswrapper[4754]: I1005 21:10:41.213383 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"d97f6b34-ef35-4b21-81eb-9e7009945fa6","Type":"ContainerDied","Data":"2d60b8bdb35899756b0e913b46e6dd512e340ee387b5ede9da9c574a7066012b"} Oct 05 21:10:41 crc kubenswrapper[4754]: I1005 21:10:41.711943 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Oct 05 21:10:41 crc kubenswrapper[4754]: I1005 21:10:41.712437 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Oct 05 21:10:42 crc kubenswrapper[4754]: I1005 21:10:42.226707 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"d97f6b34-ef35-4b21-81eb-9e7009945fa6","Type":"ContainerStarted","Data":"1a481c0e5ae2991293f88da31aa1211e4555ada5abf84f68e8f709455a4bcc47"} Oct 05 21:10:42 crc kubenswrapper[4754]: I1005 21:10:42.251466 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=-9223371975.603363 podStartE2EDuration="1m1.251413646s" podCreationTimestamp="2025-10-05 21:09:41 +0000 UTC" firstStartedPulling="2025-10-05 21:09:44.030829293 +0000 UTC m=+907.934948003" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 21:10:42.249102325 +0000 UTC m=+966.153221035" watchObservedRunningTime="2025-10-05 21:10:42.251413646 +0000 UTC m=+966.155532356" Oct 05 21:10:42 crc kubenswrapper[4754]: I1005 21:10:42.880878 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Oct 05 21:10:42 crc kubenswrapper[4754]: I1005 21:10:42.880930 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Oct 05 21:10:43 crc kubenswrapper[4754]: I1005 21:10:43.419926 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Oct 05 21:10:45 crc kubenswrapper[4754]: I1005 21:10:45.173417 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-698758b865-hh4gc"] Oct 05 21:10:45 crc kubenswrapper[4754]: I1005 21:10:45.175026 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-hh4gc" Oct 05 21:10:45 crc kubenswrapper[4754]: I1005 21:10:45.207699 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-hh4gc"] Oct 05 21:10:45 crc kubenswrapper[4754]: I1005 21:10:45.274792 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c636e2d4-bd61-47ff-ba53-e7da442521ce-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-hh4gc\" (UID: \"c636e2d4-bd61-47ff-ba53-e7da442521ce\") " pod="openstack/dnsmasq-dns-698758b865-hh4gc" Oct 05 21:10:45 crc kubenswrapper[4754]: I1005 21:10:45.274925 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c636e2d4-bd61-47ff-ba53-e7da442521ce-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-hh4gc\" (UID: \"c636e2d4-bd61-47ff-ba53-e7da442521ce\") " pod="openstack/dnsmasq-dns-698758b865-hh4gc" Oct 05 21:10:45 crc kubenswrapper[4754]: I1005 21:10:45.275101 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c636e2d4-bd61-47ff-ba53-e7da442521ce-dns-svc\") pod \"dnsmasq-dns-698758b865-hh4gc\" (UID: \"c636e2d4-bd61-47ff-ba53-e7da442521ce\") " pod="openstack/dnsmasq-dns-698758b865-hh4gc" Oct 05 21:10:45 crc kubenswrapper[4754]: I1005 21:10:45.275174 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d6hvh\" (UniqueName: \"kubernetes.io/projected/c636e2d4-bd61-47ff-ba53-e7da442521ce-kube-api-access-d6hvh\") pod \"dnsmasq-dns-698758b865-hh4gc\" (UID: \"c636e2d4-bd61-47ff-ba53-e7da442521ce\") " pod="openstack/dnsmasq-dns-698758b865-hh4gc" Oct 05 21:10:45 crc kubenswrapper[4754]: I1005 21:10:45.275367 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c636e2d4-bd61-47ff-ba53-e7da442521ce-config\") pod \"dnsmasq-dns-698758b865-hh4gc\" (UID: \"c636e2d4-bd61-47ff-ba53-e7da442521ce\") " pod="openstack/dnsmasq-dns-698758b865-hh4gc" Oct 05 21:10:45 crc kubenswrapper[4754]: I1005 21:10:45.380100 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c636e2d4-bd61-47ff-ba53-e7da442521ce-dns-svc\") pod \"dnsmasq-dns-698758b865-hh4gc\" (UID: \"c636e2d4-bd61-47ff-ba53-e7da442521ce\") " pod="openstack/dnsmasq-dns-698758b865-hh4gc" Oct 05 21:10:45 crc kubenswrapper[4754]: I1005 21:10:45.380445 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d6hvh\" (UniqueName: \"kubernetes.io/projected/c636e2d4-bd61-47ff-ba53-e7da442521ce-kube-api-access-d6hvh\") pod \"dnsmasq-dns-698758b865-hh4gc\" (UID: \"c636e2d4-bd61-47ff-ba53-e7da442521ce\") " pod="openstack/dnsmasq-dns-698758b865-hh4gc" Oct 05 21:10:45 crc kubenswrapper[4754]: I1005 21:10:45.380538 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c636e2d4-bd61-47ff-ba53-e7da442521ce-config\") pod \"dnsmasq-dns-698758b865-hh4gc\" (UID: \"c636e2d4-bd61-47ff-ba53-e7da442521ce\") " pod="openstack/dnsmasq-dns-698758b865-hh4gc" Oct 05 21:10:45 crc kubenswrapper[4754]: I1005 21:10:45.380599 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c636e2d4-bd61-47ff-ba53-e7da442521ce-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-hh4gc\" (UID: \"c636e2d4-bd61-47ff-ba53-e7da442521ce\") " pod="openstack/dnsmasq-dns-698758b865-hh4gc" Oct 05 21:10:45 crc kubenswrapper[4754]: I1005 21:10:45.380622 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c636e2d4-bd61-47ff-ba53-e7da442521ce-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-hh4gc\" (UID: \"c636e2d4-bd61-47ff-ba53-e7da442521ce\") " pod="openstack/dnsmasq-dns-698758b865-hh4gc" Oct 05 21:10:45 crc kubenswrapper[4754]: I1005 21:10:45.381637 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c636e2d4-bd61-47ff-ba53-e7da442521ce-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-hh4gc\" (UID: \"c636e2d4-bd61-47ff-ba53-e7da442521ce\") " pod="openstack/dnsmasq-dns-698758b865-hh4gc" Oct 05 21:10:45 crc kubenswrapper[4754]: I1005 21:10:45.382272 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c636e2d4-bd61-47ff-ba53-e7da442521ce-dns-svc\") pod \"dnsmasq-dns-698758b865-hh4gc\" (UID: \"c636e2d4-bd61-47ff-ba53-e7da442521ce\") " pod="openstack/dnsmasq-dns-698758b865-hh4gc" Oct 05 21:10:45 crc kubenswrapper[4754]: I1005 21:10:45.383637 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c636e2d4-bd61-47ff-ba53-e7da442521ce-config\") pod \"dnsmasq-dns-698758b865-hh4gc\" (UID: \"c636e2d4-bd61-47ff-ba53-e7da442521ce\") " pod="openstack/dnsmasq-dns-698758b865-hh4gc" Oct 05 21:10:45 crc kubenswrapper[4754]: I1005 21:10:45.383633 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c636e2d4-bd61-47ff-ba53-e7da442521ce-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-hh4gc\" (UID: \"c636e2d4-bd61-47ff-ba53-e7da442521ce\") " pod="openstack/dnsmasq-dns-698758b865-hh4gc" Oct 05 21:10:45 crc kubenswrapper[4754]: I1005 21:10:45.415952 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d6hvh\" (UniqueName: \"kubernetes.io/projected/c636e2d4-bd61-47ff-ba53-e7da442521ce-kube-api-access-d6hvh\") pod \"dnsmasq-dns-698758b865-hh4gc\" (UID: \"c636e2d4-bd61-47ff-ba53-e7da442521ce\") " pod="openstack/dnsmasq-dns-698758b865-hh4gc" Oct 05 21:10:45 crc kubenswrapper[4754]: I1005 21:10:45.502293 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-hh4gc" Oct 05 21:10:46 crc kubenswrapper[4754]: I1005 21:10:46.121989 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-hh4gc"] Oct 05 21:10:46 crc kubenswrapper[4754]: I1005 21:10:46.470711 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Oct 05 21:10:46 crc kubenswrapper[4754]: W1005 21:10:46.474746 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc636e2d4_bd61_47ff_ba53_e7da442521ce.slice/crio-e92f12d5a4e18bd997af43c419ac2d9e9ae5bd5ce362214f7e6bce07f42cda4b WatchSource:0}: Error finding container e92f12d5a4e18bd997af43c419ac2d9e9ae5bd5ce362214f7e6bce07f42cda4b: Status 404 returned error can't find the container with id e92f12d5a4e18bd997af43c419ac2d9e9ae5bd5ce362214f7e6bce07f42cda4b Oct 05 21:10:46 crc kubenswrapper[4754]: I1005 21:10:46.552790 4754 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-galera-0" podUID="d311946e-91c6-4be6-9d48-1b823ee77607" containerName="galera" probeResult="failure" output=< Oct 05 21:10:46 crc kubenswrapper[4754]: wsrep_local_state_comment (Joined) differs from Synced Oct 05 21:10:46 crc kubenswrapper[4754]: > Oct 05 21:10:46 crc kubenswrapper[4754]: I1005 21:10:46.613026 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Oct 05 21:10:46 crc kubenswrapper[4754]: I1005 21:10:46.621270 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Oct 05 21:10:46 crc kubenswrapper[4754]: I1005 21:10:46.625533 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Oct 05 21:10:46 crc kubenswrapper[4754]: I1005 21:10:46.625787 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Oct 05 21:10:46 crc kubenswrapper[4754]: I1005 21:10:46.625989 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-s9nm5" Oct 05 21:10:46 crc kubenswrapper[4754]: I1005 21:10:46.626199 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Oct 05 21:10:46 crc kubenswrapper[4754]: I1005 21:10:46.634273 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Oct 05 21:10:46 crc kubenswrapper[4754]: I1005 21:10:46.700287 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"swift-storage-0\" (UID: \"9af24378-0013-4510-aee6-d2fe2172935c\") " pod="openstack/swift-storage-0" Oct 05 21:10:46 crc kubenswrapper[4754]: I1005 21:10:46.700363 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/9af24378-0013-4510-aee6-d2fe2172935c-etc-swift\") pod \"swift-storage-0\" (UID: \"9af24378-0013-4510-aee6-d2fe2172935c\") " pod="openstack/swift-storage-0" Oct 05 21:10:46 crc kubenswrapper[4754]: I1005 21:10:46.700436 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/9af24378-0013-4510-aee6-d2fe2172935c-lock\") pod \"swift-storage-0\" (UID: \"9af24378-0013-4510-aee6-d2fe2172935c\") " pod="openstack/swift-storage-0" Oct 05 21:10:46 crc kubenswrapper[4754]: I1005 21:10:46.700579 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/9af24378-0013-4510-aee6-d2fe2172935c-cache\") pod \"swift-storage-0\" (UID: \"9af24378-0013-4510-aee6-d2fe2172935c\") " pod="openstack/swift-storage-0" Oct 05 21:10:46 crc kubenswrapper[4754]: I1005 21:10:46.700602 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xvdqj\" (UniqueName: \"kubernetes.io/projected/9af24378-0013-4510-aee6-d2fe2172935c-kube-api-access-xvdqj\") pod \"swift-storage-0\" (UID: \"9af24378-0013-4510-aee6-d2fe2172935c\") " pod="openstack/swift-storage-0" Oct 05 21:10:46 crc kubenswrapper[4754]: I1005 21:10:46.802644 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/9af24378-0013-4510-aee6-d2fe2172935c-lock\") pod \"swift-storage-0\" (UID: \"9af24378-0013-4510-aee6-d2fe2172935c\") " pod="openstack/swift-storage-0" Oct 05 21:10:46 crc kubenswrapper[4754]: I1005 21:10:46.802758 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/9af24378-0013-4510-aee6-d2fe2172935c-cache\") pod \"swift-storage-0\" (UID: \"9af24378-0013-4510-aee6-d2fe2172935c\") " pod="openstack/swift-storage-0" Oct 05 21:10:46 crc kubenswrapper[4754]: I1005 21:10:46.802792 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xvdqj\" (UniqueName: \"kubernetes.io/projected/9af24378-0013-4510-aee6-d2fe2172935c-kube-api-access-xvdqj\") pod \"swift-storage-0\" (UID: \"9af24378-0013-4510-aee6-d2fe2172935c\") " pod="openstack/swift-storage-0" Oct 05 21:10:46 crc kubenswrapper[4754]: I1005 21:10:46.802850 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"swift-storage-0\" (UID: \"9af24378-0013-4510-aee6-d2fe2172935c\") " pod="openstack/swift-storage-0" Oct 05 21:10:46 crc kubenswrapper[4754]: I1005 21:10:46.802892 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/9af24378-0013-4510-aee6-d2fe2172935c-etc-swift\") pod \"swift-storage-0\" (UID: \"9af24378-0013-4510-aee6-d2fe2172935c\") " pod="openstack/swift-storage-0" Oct 05 21:10:46 crc kubenswrapper[4754]: E1005 21:10:46.803043 4754 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 05 21:10:46 crc kubenswrapper[4754]: E1005 21:10:46.803065 4754 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 05 21:10:46 crc kubenswrapper[4754]: I1005 21:10:46.803072 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/9af24378-0013-4510-aee6-d2fe2172935c-lock\") pod \"swift-storage-0\" (UID: \"9af24378-0013-4510-aee6-d2fe2172935c\") " pod="openstack/swift-storage-0" Oct 05 21:10:46 crc kubenswrapper[4754]: E1005 21:10:46.803115 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9af24378-0013-4510-aee6-d2fe2172935c-etc-swift podName:9af24378-0013-4510-aee6-d2fe2172935c nodeName:}" failed. No retries permitted until 2025-10-05 21:10:47.303097487 +0000 UTC m=+971.207216197 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/9af24378-0013-4510-aee6-d2fe2172935c-etc-swift") pod "swift-storage-0" (UID: "9af24378-0013-4510-aee6-d2fe2172935c") : configmap "swift-ring-files" not found Oct 05 21:10:46 crc kubenswrapper[4754]: I1005 21:10:46.803290 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/9af24378-0013-4510-aee6-d2fe2172935c-cache\") pod \"swift-storage-0\" (UID: \"9af24378-0013-4510-aee6-d2fe2172935c\") " pod="openstack/swift-storage-0" Oct 05 21:10:46 crc kubenswrapper[4754]: I1005 21:10:46.803399 4754 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"swift-storage-0\" (UID: \"9af24378-0013-4510-aee6-d2fe2172935c\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/swift-storage-0" Oct 05 21:10:46 crc kubenswrapper[4754]: I1005 21:10:46.823293 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xvdqj\" (UniqueName: \"kubernetes.io/projected/9af24378-0013-4510-aee6-d2fe2172935c-kube-api-access-xvdqj\") pod \"swift-storage-0\" (UID: \"9af24378-0013-4510-aee6-d2fe2172935c\") " pod="openstack/swift-storage-0" Oct 05 21:10:46 crc kubenswrapper[4754]: I1005 21:10:46.827208 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"swift-storage-0\" (UID: \"9af24378-0013-4510-aee6-d2fe2172935c\") " pod="openstack/swift-storage-0" Oct 05 21:10:47 crc kubenswrapper[4754]: E1005 21:10:47.054975 4754 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.20:57564->38.102.83.20:44369: write tcp 38.102.83.20:57564->38.102.83.20:44369: write: broken pipe Oct 05 21:10:47 crc kubenswrapper[4754]: I1005 21:10:47.132120 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-ljsff"] Oct 05 21:10:47 crc kubenswrapper[4754]: I1005 21:10:47.133066 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-ljsff" Oct 05 21:10:47 crc kubenswrapper[4754]: I1005 21:10:47.134806 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Oct 05 21:10:47 crc kubenswrapper[4754]: I1005 21:10:47.138654 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Oct 05 21:10:47 crc kubenswrapper[4754]: I1005 21:10:47.142139 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Oct 05 21:10:47 crc kubenswrapper[4754]: I1005 21:10:47.161898 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-ljsff"] Oct 05 21:10:47 crc kubenswrapper[4754]: I1005 21:10:47.210195 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/89a55f59-6693-4e0d-b8f8-dfb258d7a078-combined-ca-bundle\") pod \"swift-ring-rebalance-ljsff\" (UID: \"89a55f59-6693-4e0d-b8f8-dfb258d7a078\") " pod="openstack/swift-ring-rebalance-ljsff" Oct 05 21:10:47 crc kubenswrapper[4754]: I1005 21:10:47.210241 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wmphp\" (UniqueName: \"kubernetes.io/projected/89a55f59-6693-4e0d-b8f8-dfb258d7a078-kube-api-access-wmphp\") pod \"swift-ring-rebalance-ljsff\" (UID: \"89a55f59-6693-4e0d-b8f8-dfb258d7a078\") " pod="openstack/swift-ring-rebalance-ljsff" Oct 05 21:10:47 crc kubenswrapper[4754]: I1005 21:10:47.210273 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/89a55f59-6693-4e0d-b8f8-dfb258d7a078-dispersionconf\") pod \"swift-ring-rebalance-ljsff\" (UID: \"89a55f59-6693-4e0d-b8f8-dfb258d7a078\") " pod="openstack/swift-ring-rebalance-ljsff" Oct 05 21:10:47 crc kubenswrapper[4754]: I1005 21:10:47.210302 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/89a55f59-6693-4e0d-b8f8-dfb258d7a078-scripts\") pod \"swift-ring-rebalance-ljsff\" (UID: \"89a55f59-6693-4e0d-b8f8-dfb258d7a078\") " pod="openstack/swift-ring-rebalance-ljsff" Oct 05 21:10:47 crc kubenswrapper[4754]: I1005 21:10:47.210335 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/89a55f59-6693-4e0d-b8f8-dfb258d7a078-swiftconf\") pod \"swift-ring-rebalance-ljsff\" (UID: \"89a55f59-6693-4e0d-b8f8-dfb258d7a078\") " pod="openstack/swift-ring-rebalance-ljsff" Oct 05 21:10:47 crc kubenswrapper[4754]: I1005 21:10:47.210384 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/89a55f59-6693-4e0d-b8f8-dfb258d7a078-ring-data-devices\") pod \"swift-ring-rebalance-ljsff\" (UID: \"89a55f59-6693-4e0d-b8f8-dfb258d7a078\") " pod="openstack/swift-ring-rebalance-ljsff" Oct 05 21:10:47 crc kubenswrapper[4754]: I1005 21:10:47.210419 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/89a55f59-6693-4e0d-b8f8-dfb258d7a078-etc-swift\") pod \"swift-ring-rebalance-ljsff\" (UID: \"89a55f59-6693-4e0d-b8f8-dfb258d7a078\") " pod="openstack/swift-ring-rebalance-ljsff" Oct 05 21:10:47 crc kubenswrapper[4754]: I1005 21:10:47.274029 4754 generic.go:334] "Generic (PLEG): container finished" podID="c636e2d4-bd61-47ff-ba53-e7da442521ce" containerID="378fc5007a8d78723eabecf80c7b1dda63d132593907e383e3dae39d94e14ea2" exitCode=0 Oct 05 21:10:47 crc kubenswrapper[4754]: I1005 21:10:47.274089 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-hh4gc" event={"ID":"c636e2d4-bd61-47ff-ba53-e7da442521ce","Type":"ContainerDied","Data":"378fc5007a8d78723eabecf80c7b1dda63d132593907e383e3dae39d94e14ea2"} Oct 05 21:10:47 crc kubenswrapper[4754]: I1005 21:10:47.274114 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-hh4gc" event={"ID":"c636e2d4-bd61-47ff-ba53-e7da442521ce","Type":"ContainerStarted","Data":"e92f12d5a4e18bd997af43c419ac2d9e9ae5bd5ce362214f7e6bce07f42cda4b"} Oct 05 21:10:47 crc kubenswrapper[4754]: I1005 21:10:47.276272 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"316121e1-b760-4eaf-9b8e-ec5bacc6b117","Type":"ContainerStarted","Data":"a7a9f4fe4f1a847a422214e203e9c8f1a0b286ee29e41fd9608a83f27260b311"} Oct 05 21:10:47 crc kubenswrapper[4754]: I1005 21:10:47.276294 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"316121e1-b760-4eaf-9b8e-ec5bacc6b117","Type":"ContainerStarted","Data":"27903c688276561dd0822b261e950223f1075a2d803319f66467317cdd15eb07"} Oct 05 21:10:47 crc kubenswrapper[4754]: I1005 21:10:47.276457 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Oct 05 21:10:47 crc kubenswrapper[4754]: I1005 21:10:47.311737 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/9af24378-0013-4510-aee6-d2fe2172935c-etc-swift\") pod \"swift-storage-0\" (UID: \"9af24378-0013-4510-aee6-d2fe2172935c\") " pod="openstack/swift-storage-0" Oct 05 21:10:47 crc kubenswrapper[4754]: I1005 21:10:47.311773 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/89a55f59-6693-4e0d-b8f8-dfb258d7a078-ring-data-devices\") pod \"swift-ring-rebalance-ljsff\" (UID: \"89a55f59-6693-4e0d-b8f8-dfb258d7a078\") " pod="openstack/swift-ring-rebalance-ljsff" Oct 05 21:10:47 crc kubenswrapper[4754]: I1005 21:10:47.311853 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/89a55f59-6693-4e0d-b8f8-dfb258d7a078-etc-swift\") pod \"swift-ring-rebalance-ljsff\" (UID: \"89a55f59-6693-4e0d-b8f8-dfb258d7a078\") " pod="openstack/swift-ring-rebalance-ljsff" Oct 05 21:10:47 crc kubenswrapper[4754]: I1005 21:10:47.311925 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/89a55f59-6693-4e0d-b8f8-dfb258d7a078-combined-ca-bundle\") pod \"swift-ring-rebalance-ljsff\" (UID: \"89a55f59-6693-4e0d-b8f8-dfb258d7a078\") " pod="openstack/swift-ring-rebalance-ljsff" Oct 05 21:10:47 crc kubenswrapper[4754]: E1005 21:10:47.311929 4754 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 05 21:10:47 crc kubenswrapper[4754]: I1005 21:10:47.311946 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wmphp\" (UniqueName: \"kubernetes.io/projected/89a55f59-6693-4e0d-b8f8-dfb258d7a078-kube-api-access-wmphp\") pod \"swift-ring-rebalance-ljsff\" (UID: \"89a55f59-6693-4e0d-b8f8-dfb258d7a078\") " pod="openstack/swift-ring-rebalance-ljsff" Oct 05 21:10:47 crc kubenswrapper[4754]: I1005 21:10:47.311986 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/89a55f59-6693-4e0d-b8f8-dfb258d7a078-dispersionconf\") pod \"swift-ring-rebalance-ljsff\" (UID: \"89a55f59-6693-4e0d-b8f8-dfb258d7a078\") " pod="openstack/swift-ring-rebalance-ljsff" Oct 05 21:10:47 crc kubenswrapper[4754]: I1005 21:10:47.312027 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/89a55f59-6693-4e0d-b8f8-dfb258d7a078-scripts\") pod \"swift-ring-rebalance-ljsff\" (UID: \"89a55f59-6693-4e0d-b8f8-dfb258d7a078\") " pod="openstack/swift-ring-rebalance-ljsff" Oct 05 21:10:47 crc kubenswrapper[4754]: I1005 21:10:47.312072 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/89a55f59-6693-4e0d-b8f8-dfb258d7a078-swiftconf\") pod \"swift-ring-rebalance-ljsff\" (UID: \"89a55f59-6693-4e0d-b8f8-dfb258d7a078\") " pod="openstack/swift-ring-rebalance-ljsff" Oct 05 21:10:47 crc kubenswrapper[4754]: E1005 21:10:47.311950 4754 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 05 21:10:47 crc kubenswrapper[4754]: E1005 21:10:47.312248 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9af24378-0013-4510-aee6-d2fe2172935c-etc-swift podName:9af24378-0013-4510-aee6-d2fe2172935c nodeName:}" failed. No retries permitted until 2025-10-05 21:10:48.312221213 +0000 UTC m=+972.216339923 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/9af24378-0013-4510-aee6-d2fe2172935c-etc-swift") pod "swift-storage-0" (UID: "9af24378-0013-4510-aee6-d2fe2172935c") : configmap "swift-ring-files" not found Oct 05 21:10:47 crc kubenswrapper[4754]: I1005 21:10:47.312712 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/89a55f59-6693-4e0d-b8f8-dfb258d7a078-etc-swift\") pod \"swift-ring-rebalance-ljsff\" (UID: \"89a55f59-6693-4e0d-b8f8-dfb258d7a078\") " pod="openstack/swift-ring-rebalance-ljsff" Oct 05 21:10:47 crc kubenswrapper[4754]: I1005 21:10:47.312874 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/89a55f59-6693-4e0d-b8f8-dfb258d7a078-ring-data-devices\") pod \"swift-ring-rebalance-ljsff\" (UID: \"89a55f59-6693-4e0d-b8f8-dfb258d7a078\") " pod="openstack/swift-ring-rebalance-ljsff" Oct 05 21:10:47 crc kubenswrapper[4754]: I1005 21:10:47.313332 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/89a55f59-6693-4e0d-b8f8-dfb258d7a078-scripts\") pod \"swift-ring-rebalance-ljsff\" (UID: \"89a55f59-6693-4e0d-b8f8-dfb258d7a078\") " pod="openstack/swift-ring-rebalance-ljsff" Oct 05 21:10:47 crc kubenswrapper[4754]: I1005 21:10:47.316677 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/89a55f59-6693-4e0d-b8f8-dfb258d7a078-dispersionconf\") pod \"swift-ring-rebalance-ljsff\" (UID: \"89a55f59-6693-4e0d-b8f8-dfb258d7a078\") " pod="openstack/swift-ring-rebalance-ljsff" Oct 05 21:10:47 crc kubenswrapper[4754]: I1005 21:10:47.319196 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/89a55f59-6693-4e0d-b8f8-dfb258d7a078-swiftconf\") pod \"swift-ring-rebalance-ljsff\" (UID: \"89a55f59-6693-4e0d-b8f8-dfb258d7a078\") " pod="openstack/swift-ring-rebalance-ljsff" Oct 05 21:10:47 crc kubenswrapper[4754]: I1005 21:10:47.319898 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=1.912733549 podStartE2EDuration="7.319881377s" podCreationTimestamp="2025-10-05 21:10:40 +0000 UTC" firstStartedPulling="2025-10-05 21:10:41.196879969 +0000 UTC m=+965.100998679" lastFinishedPulling="2025-10-05 21:10:46.604027797 +0000 UTC m=+970.508146507" observedRunningTime="2025-10-05 21:10:47.313370034 +0000 UTC m=+971.217488784" watchObservedRunningTime="2025-10-05 21:10:47.319881377 +0000 UTC m=+971.224000087" Oct 05 21:10:47 crc kubenswrapper[4754]: I1005 21:10:47.325437 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/89a55f59-6693-4e0d-b8f8-dfb258d7a078-combined-ca-bundle\") pod \"swift-ring-rebalance-ljsff\" (UID: \"89a55f59-6693-4e0d-b8f8-dfb258d7a078\") " pod="openstack/swift-ring-rebalance-ljsff" Oct 05 21:10:47 crc kubenswrapper[4754]: I1005 21:10:47.340340 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wmphp\" (UniqueName: \"kubernetes.io/projected/89a55f59-6693-4e0d-b8f8-dfb258d7a078-kube-api-access-wmphp\") pod \"swift-ring-rebalance-ljsff\" (UID: \"89a55f59-6693-4e0d-b8f8-dfb258d7a078\") " pod="openstack/swift-ring-rebalance-ljsff" Oct 05 21:10:47 crc kubenswrapper[4754]: I1005 21:10:47.458194 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-ljsff" Oct 05 21:10:47 crc kubenswrapper[4754]: I1005 21:10:47.921194 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-ljsff"] Oct 05 21:10:48 crc kubenswrapper[4754]: I1005 21:10:48.285008 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-ljsff" event={"ID":"89a55f59-6693-4e0d-b8f8-dfb258d7a078","Type":"ContainerStarted","Data":"a0c4a11b0dcc0302ed3649ca159b0d1debc77c5994d3896ecd01551568da0fc7"} Oct 05 21:10:48 crc kubenswrapper[4754]: I1005 21:10:48.287397 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-hh4gc" event={"ID":"c636e2d4-bd61-47ff-ba53-e7da442521ce","Type":"ContainerStarted","Data":"d487267b338cdf082ec536549e664a2ceda532b59c086ea25c0f0500412b65b9"} Oct 05 21:10:48 crc kubenswrapper[4754]: I1005 21:10:48.287799 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-698758b865-hh4gc" Oct 05 21:10:48 crc kubenswrapper[4754]: I1005 21:10:48.306069 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-698758b865-hh4gc" podStartSLOduration=3.306049903 podStartE2EDuration="3.306049903s" podCreationTimestamp="2025-10-05 21:10:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 21:10:48.301165343 +0000 UTC m=+972.205284053" watchObservedRunningTime="2025-10-05 21:10:48.306049903 +0000 UTC m=+972.210168613" Oct 05 21:10:48 crc kubenswrapper[4754]: I1005 21:10:48.330462 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/9af24378-0013-4510-aee6-d2fe2172935c-etc-swift\") pod \"swift-storage-0\" (UID: \"9af24378-0013-4510-aee6-d2fe2172935c\") " pod="openstack/swift-storage-0" Oct 05 21:10:48 crc kubenswrapper[4754]: E1005 21:10:48.330701 4754 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 05 21:10:48 crc kubenswrapper[4754]: E1005 21:10:48.330740 4754 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 05 21:10:48 crc kubenswrapper[4754]: E1005 21:10:48.330826 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9af24378-0013-4510-aee6-d2fe2172935c-etc-swift podName:9af24378-0013-4510-aee6-d2fe2172935c nodeName:}" failed. No retries permitted until 2025-10-05 21:10:50.330798102 +0000 UTC m=+974.234916802 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/9af24378-0013-4510-aee6-d2fe2172935c-etc-swift") pod "swift-storage-0" (UID: "9af24378-0013-4510-aee6-d2fe2172935c") : configmap "swift-ring-files" not found Oct 05 21:10:48 crc kubenswrapper[4754]: I1005 21:10:48.980387 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Oct 05 21:10:49 crc kubenswrapper[4754]: I1005 21:10:49.041276 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Oct 05 21:10:50 crc kubenswrapper[4754]: I1005 21:10:50.371081 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/9af24378-0013-4510-aee6-d2fe2172935c-etc-swift\") pod \"swift-storage-0\" (UID: \"9af24378-0013-4510-aee6-d2fe2172935c\") " pod="openstack/swift-storage-0" Oct 05 21:10:50 crc kubenswrapper[4754]: E1005 21:10:50.371296 4754 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 05 21:10:50 crc kubenswrapper[4754]: E1005 21:10:50.371629 4754 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 05 21:10:50 crc kubenswrapper[4754]: E1005 21:10:50.371684 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9af24378-0013-4510-aee6-d2fe2172935c-etc-swift podName:9af24378-0013-4510-aee6-d2fe2172935c nodeName:}" failed. No retries permitted until 2025-10-05 21:10:54.371665742 +0000 UTC m=+978.275784452 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/9af24378-0013-4510-aee6-d2fe2172935c-etc-swift") pod "swift-storage-0" (UID: "9af24378-0013-4510-aee6-d2fe2172935c") : configmap "swift-ring-files" not found Oct 05 21:10:51 crc kubenswrapper[4754]: I1005 21:10:51.848952 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Oct 05 21:10:52 crc kubenswrapper[4754]: I1005 21:10:52.323122 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-ljsff" event={"ID":"89a55f59-6693-4e0d-b8f8-dfb258d7a078","Type":"ContainerStarted","Data":"d2805c7e07be7017d0492b595038648706c23628e8be64e28d3484d0af8becc1"} Oct 05 21:10:52 crc kubenswrapper[4754]: I1005 21:10:52.345899 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-ljsff" podStartSLOduration=1.335023385 podStartE2EDuration="5.345875456s" podCreationTimestamp="2025-10-05 21:10:47 +0000 UTC" firstStartedPulling="2025-10-05 21:10:47.931198153 +0000 UTC m=+971.835316903" lastFinishedPulling="2025-10-05 21:10:51.942050264 +0000 UTC m=+975.846168974" observedRunningTime="2025-10-05 21:10:52.342031683 +0000 UTC m=+976.246150413" watchObservedRunningTime="2025-10-05 21:10:52.345875456 +0000 UTC m=+976.249994176" Oct 05 21:10:52 crc kubenswrapper[4754]: I1005 21:10:52.977347 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-g8hwz"] Oct 05 21:10:52 crc kubenswrapper[4754]: I1005 21:10:52.978643 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-g8hwz" Oct 05 21:10:52 crc kubenswrapper[4754]: I1005 21:10:52.994707 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-g8hwz"] Oct 05 21:10:53 crc kubenswrapper[4754]: I1005 21:10:53.021663 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5nrjg\" (UniqueName: \"kubernetes.io/projected/b8d124eb-a543-4988-9523-d0e37ca113b8-kube-api-access-5nrjg\") pod \"keystone-db-create-g8hwz\" (UID: \"b8d124eb-a543-4988-9523-d0e37ca113b8\") " pod="openstack/keystone-db-create-g8hwz" Oct 05 21:10:53 crc kubenswrapper[4754]: I1005 21:10:53.122913 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5nrjg\" (UniqueName: \"kubernetes.io/projected/b8d124eb-a543-4988-9523-d0e37ca113b8-kube-api-access-5nrjg\") pod \"keystone-db-create-g8hwz\" (UID: \"b8d124eb-a543-4988-9523-d0e37ca113b8\") " pod="openstack/keystone-db-create-g8hwz" Oct 05 21:10:53 crc kubenswrapper[4754]: I1005 21:10:53.150699 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5nrjg\" (UniqueName: \"kubernetes.io/projected/b8d124eb-a543-4988-9523-d0e37ca113b8-kube-api-access-5nrjg\") pod \"keystone-db-create-g8hwz\" (UID: \"b8d124eb-a543-4988-9523-d0e37ca113b8\") " pod="openstack/keystone-db-create-g8hwz" Oct 05 21:10:53 crc kubenswrapper[4754]: I1005 21:10:53.210904 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-c7gj8"] Oct 05 21:10:53 crc kubenswrapper[4754]: I1005 21:10:53.211970 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-c7gj8" Oct 05 21:10:53 crc kubenswrapper[4754]: I1005 21:10:53.227107 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fvjck\" (UniqueName: \"kubernetes.io/projected/28dab52f-2bfa-409f-8ec3-b47ae0641ce2-kube-api-access-fvjck\") pod \"placement-db-create-c7gj8\" (UID: \"28dab52f-2bfa-409f-8ec3-b47ae0641ce2\") " pod="openstack/placement-db-create-c7gj8" Oct 05 21:10:53 crc kubenswrapper[4754]: I1005 21:10:53.228530 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-c7gj8"] Oct 05 21:10:53 crc kubenswrapper[4754]: I1005 21:10:53.295967 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-g8hwz" Oct 05 21:10:53 crc kubenswrapper[4754]: I1005 21:10:53.329078 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fvjck\" (UniqueName: \"kubernetes.io/projected/28dab52f-2bfa-409f-8ec3-b47ae0641ce2-kube-api-access-fvjck\") pod \"placement-db-create-c7gj8\" (UID: \"28dab52f-2bfa-409f-8ec3-b47ae0641ce2\") " pod="openstack/placement-db-create-c7gj8" Oct 05 21:10:53 crc kubenswrapper[4754]: I1005 21:10:53.354251 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fvjck\" (UniqueName: \"kubernetes.io/projected/28dab52f-2bfa-409f-8ec3-b47ae0641ce2-kube-api-access-fvjck\") pod \"placement-db-create-c7gj8\" (UID: \"28dab52f-2bfa-409f-8ec3-b47ae0641ce2\") " pod="openstack/placement-db-create-c7gj8" Oct 05 21:10:53 crc kubenswrapper[4754]: I1005 21:10:53.546857 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-c7gj8" Oct 05 21:10:53 crc kubenswrapper[4754]: I1005 21:10:53.612838 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-5h8xj"] Oct 05 21:10:53 crc kubenswrapper[4754]: I1005 21:10:53.634516 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-5h8xj" Oct 05 21:10:53 crc kubenswrapper[4754]: I1005 21:10:53.656057 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-5h8xj"] Oct 05 21:10:53 crc kubenswrapper[4754]: I1005 21:10:53.754353 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bwmm5\" (UniqueName: \"kubernetes.io/projected/cee454de-74cf-4913-8792-a231ddd4c691-kube-api-access-bwmm5\") pod \"glance-db-create-5h8xj\" (UID: \"cee454de-74cf-4913-8792-a231ddd4c691\") " pod="openstack/glance-db-create-5h8xj" Oct 05 21:10:53 crc kubenswrapper[4754]: I1005 21:10:53.783125 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-g8hwz"] Oct 05 21:10:53 crc kubenswrapper[4754]: I1005 21:10:53.857372 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bwmm5\" (UniqueName: \"kubernetes.io/projected/cee454de-74cf-4913-8792-a231ddd4c691-kube-api-access-bwmm5\") pod \"glance-db-create-5h8xj\" (UID: \"cee454de-74cf-4913-8792-a231ddd4c691\") " pod="openstack/glance-db-create-5h8xj" Oct 05 21:10:53 crc kubenswrapper[4754]: I1005 21:10:53.876354 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bwmm5\" (UniqueName: \"kubernetes.io/projected/cee454de-74cf-4913-8792-a231ddd4c691-kube-api-access-bwmm5\") pod \"glance-db-create-5h8xj\" (UID: \"cee454de-74cf-4913-8792-a231ddd4c691\") " pod="openstack/glance-db-create-5h8xj" Oct 05 21:10:53 crc kubenswrapper[4754]: I1005 21:10:53.979953 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-5h8xj" Oct 05 21:10:54 crc kubenswrapper[4754]: I1005 21:10:54.164947 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-c7gj8"] Oct 05 21:10:54 crc kubenswrapper[4754]: I1005 21:10:54.343746 4754 generic.go:334] "Generic (PLEG): container finished" podID="b8d124eb-a543-4988-9523-d0e37ca113b8" containerID="e5cc49e57eb855cc1e05ded8fe904f5da764a35f1685c58e8914f9a18b0f54c0" exitCode=0 Oct 05 21:10:54 crc kubenswrapper[4754]: I1005 21:10:54.343834 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-g8hwz" event={"ID":"b8d124eb-a543-4988-9523-d0e37ca113b8","Type":"ContainerDied","Data":"e5cc49e57eb855cc1e05ded8fe904f5da764a35f1685c58e8914f9a18b0f54c0"} Oct 05 21:10:54 crc kubenswrapper[4754]: I1005 21:10:54.343874 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-g8hwz" event={"ID":"b8d124eb-a543-4988-9523-d0e37ca113b8","Type":"ContainerStarted","Data":"dd7f567876f9909834dceb2405002aa7f28dfccf964a367843cf3ec670a1dca3"} Oct 05 21:10:54 crc kubenswrapper[4754]: I1005 21:10:54.345889 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-c7gj8" event={"ID":"28dab52f-2bfa-409f-8ec3-b47ae0641ce2","Type":"ContainerStarted","Data":"1036385fff7b217e6fde18404b163e65732a0261a276a04863fe63810076a4e1"} Oct 05 21:10:54 crc kubenswrapper[4754]: I1005 21:10:54.431105 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-5h8xj"] Oct 05 21:10:54 crc kubenswrapper[4754]: I1005 21:10:54.471147 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/9af24378-0013-4510-aee6-d2fe2172935c-etc-swift\") pod \"swift-storage-0\" (UID: \"9af24378-0013-4510-aee6-d2fe2172935c\") " pod="openstack/swift-storage-0" Oct 05 21:10:54 crc kubenswrapper[4754]: E1005 21:10:54.471222 4754 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 05 21:10:54 crc kubenswrapper[4754]: E1005 21:10:54.471240 4754 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 05 21:10:54 crc kubenswrapper[4754]: E1005 21:10:54.471294 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9af24378-0013-4510-aee6-d2fe2172935c-etc-swift podName:9af24378-0013-4510-aee6-d2fe2172935c nodeName:}" failed. No retries permitted until 2025-10-05 21:11:02.471275726 +0000 UTC m=+986.375394436 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/9af24378-0013-4510-aee6-d2fe2172935c-etc-swift") pod "swift-storage-0" (UID: "9af24378-0013-4510-aee6-d2fe2172935c") : configmap "swift-ring-files" not found Oct 05 21:10:55 crc kubenswrapper[4754]: I1005 21:10:55.356906 4754 generic.go:334] "Generic (PLEG): container finished" podID="28dab52f-2bfa-409f-8ec3-b47ae0641ce2" containerID="0e0159d153289e7a8a9958ccd54336126d3fa18d6bb00f21b2ef5061991767b6" exitCode=0 Oct 05 21:10:55 crc kubenswrapper[4754]: I1005 21:10:55.356949 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-c7gj8" event={"ID":"28dab52f-2bfa-409f-8ec3-b47ae0641ce2","Type":"ContainerDied","Data":"0e0159d153289e7a8a9958ccd54336126d3fa18d6bb00f21b2ef5061991767b6"} Oct 05 21:10:55 crc kubenswrapper[4754]: I1005 21:10:55.362391 4754 generic.go:334] "Generic (PLEG): container finished" podID="cee454de-74cf-4913-8792-a231ddd4c691" containerID="62d7716283de84282a34e0fa4992d2b086d38ba6c83499f982fd51187d96878a" exitCode=0 Oct 05 21:10:55 crc kubenswrapper[4754]: I1005 21:10:55.362519 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-5h8xj" event={"ID":"cee454de-74cf-4913-8792-a231ddd4c691","Type":"ContainerDied","Data":"62d7716283de84282a34e0fa4992d2b086d38ba6c83499f982fd51187d96878a"} Oct 05 21:10:55 crc kubenswrapper[4754]: I1005 21:10:55.362593 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-5h8xj" event={"ID":"cee454de-74cf-4913-8792-a231ddd4c691","Type":"ContainerStarted","Data":"4a98b3ea4e794078a58aecea9e0cf8486b7c1cc19be1c68a6534e50000454f5b"} Oct 05 21:10:55 crc kubenswrapper[4754]: I1005 21:10:55.503750 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-698758b865-hh4gc" Oct 05 21:10:55 crc kubenswrapper[4754]: I1005 21:10:55.583899 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-r8m5k"] Oct 05 21:10:55 crc kubenswrapper[4754]: I1005 21:10:55.584162 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-86db49b7ff-r8m5k" podUID="a40e02a7-900e-4da2-8a4a-73773ca90876" containerName="dnsmasq-dns" containerID="cri-o://20598aa4e239724712ce00389212a2b4a21ab17443b6890e775b769028651f20" gracePeriod=10 Oct 05 21:10:55 crc kubenswrapper[4754]: I1005 21:10:55.842106 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-g8hwz" Oct 05 21:10:55 crc kubenswrapper[4754]: I1005 21:10:55.911528 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5nrjg\" (UniqueName: \"kubernetes.io/projected/b8d124eb-a543-4988-9523-d0e37ca113b8-kube-api-access-5nrjg\") pod \"b8d124eb-a543-4988-9523-d0e37ca113b8\" (UID: \"b8d124eb-a543-4988-9523-d0e37ca113b8\") " Oct 05 21:10:55 crc kubenswrapper[4754]: I1005 21:10:55.918809 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b8d124eb-a543-4988-9523-d0e37ca113b8-kube-api-access-5nrjg" (OuterVolumeSpecName: "kube-api-access-5nrjg") pod "b8d124eb-a543-4988-9523-d0e37ca113b8" (UID: "b8d124eb-a543-4988-9523-d0e37ca113b8"). InnerVolumeSpecName "kube-api-access-5nrjg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:10:56 crc kubenswrapper[4754]: I1005 21:10:56.013790 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5nrjg\" (UniqueName: \"kubernetes.io/projected/b8d124eb-a543-4988-9523-d0e37ca113b8-kube-api-access-5nrjg\") on node \"crc\" DevicePath \"\"" Oct 05 21:10:56 crc kubenswrapper[4754]: I1005 21:10:56.156414 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-r8m5k" Oct 05 21:10:56 crc kubenswrapper[4754]: I1005 21:10:56.217539 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a40e02a7-900e-4da2-8a4a-73773ca90876-dns-svc\") pod \"a40e02a7-900e-4da2-8a4a-73773ca90876\" (UID: \"a40e02a7-900e-4da2-8a4a-73773ca90876\") " Oct 05 21:10:56 crc kubenswrapper[4754]: I1005 21:10:56.217608 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8ls6k\" (UniqueName: \"kubernetes.io/projected/a40e02a7-900e-4da2-8a4a-73773ca90876-kube-api-access-8ls6k\") pod \"a40e02a7-900e-4da2-8a4a-73773ca90876\" (UID: \"a40e02a7-900e-4da2-8a4a-73773ca90876\") " Oct 05 21:10:56 crc kubenswrapper[4754]: I1005 21:10:56.217630 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a40e02a7-900e-4da2-8a4a-73773ca90876-ovsdbserver-sb\") pod \"a40e02a7-900e-4da2-8a4a-73773ca90876\" (UID: \"a40e02a7-900e-4da2-8a4a-73773ca90876\") " Oct 05 21:10:56 crc kubenswrapper[4754]: I1005 21:10:56.217722 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a40e02a7-900e-4da2-8a4a-73773ca90876-config\") pod \"a40e02a7-900e-4da2-8a4a-73773ca90876\" (UID: \"a40e02a7-900e-4da2-8a4a-73773ca90876\") " Oct 05 21:10:56 crc kubenswrapper[4754]: I1005 21:10:56.217804 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a40e02a7-900e-4da2-8a4a-73773ca90876-ovsdbserver-nb\") pod \"a40e02a7-900e-4da2-8a4a-73773ca90876\" (UID: \"a40e02a7-900e-4da2-8a4a-73773ca90876\") " Oct 05 21:10:56 crc kubenswrapper[4754]: I1005 21:10:56.230022 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a40e02a7-900e-4da2-8a4a-73773ca90876-kube-api-access-8ls6k" (OuterVolumeSpecName: "kube-api-access-8ls6k") pod "a40e02a7-900e-4da2-8a4a-73773ca90876" (UID: "a40e02a7-900e-4da2-8a4a-73773ca90876"). InnerVolumeSpecName "kube-api-access-8ls6k". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:10:56 crc kubenswrapper[4754]: I1005 21:10:56.265198 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a40e02a7-900e-4da2-8a4a-73773ca90876-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "a40e02a7-900e-4da2-8a4a-73773ca90876" (UID: "a40e02a7-900e-4da2-8a4a-73773ca90876"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 21:10:56 crc kubenswrapper[4754]: I1005 21:10:56.269325 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a40e02a7-900e-4da2-8a4a-73773ca90876-config" (OuterVolumeSpecName: "config") pod "a40e02a7-900e-4da2-8a4a-73773ca90876" (UID: "a40e02a7-900e-4da2-8a4a-73773ca90876"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 21:10:56 crc kubenswrapper[4754]: I1005 21:10:56.274514 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a40e02a7-900e-4da2-8a4a-73773ca90876-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "a40e02a7-900e-4da2-8a4a-73773ca90876" (UID: "a40e02a7-900e-4da2-8a4a-73773ca90876"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 21:10:56 crc kubenswrapper[4754]: I1005 21:10:56.293893 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a40e02a7-900e-4da2-8a4a-73773ca90876-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a40e02a7-900e-4da2-8a4a-73773ca90876" (UID: "a40e02a7-900e-4da2-8a4a-73773ca90876"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 21:10:56 crc kubenswrapper[4754]: I1005 21:10:56.319271 4754 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a40e02a7-900e-4da2-8a4a-73773ca90876-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 05 21:10:56 crc kubenswrapper[4754]: I1005 21:10:56.319291 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8ls6k\" (UniqueName: \"kubernetes.io/projected/a40e02a7-900e-4da2-8a4a-73773ca90876-kube-api-access-8ls6k\") on node \"crc\" DevicePath \"\"" Oct 05 21:10:56 crc kubenswrapper[4754]: I1005 21:10:56.319304 4754 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a40e02a7-900e-4da2-8a4a-73773ca90876-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 05 21:10:56 crc kubenswrapper[4754]: I1005 21:10:56.319312 4754 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a40e02a7-900e-4da2-8a4a-73773ca90876-config\") on node \"crc\" DevicePath \"\"" Oct 05 21:10:56 crc kubenswrapper[4754]: I1005 21:10:56.319320 4754 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a40e02a7-900e-4da2-8a4a-73773ca90876-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 05 21:10:56 crc kubenswrapper[4754]: I1005 21:10:56.373327 4754 generic.go:334] "Generic (PLEG): container finished" podID="a40e02a7-900e-4da2-8a4a-73773ca90876" containerID="20598aa4e239724712ce00389212a2b4a21ab17443b6890e775b769028651f20" exitCode=0 Oct 05 21:10:56 crc kubenswrapper[4754]: I1005 21:10:56.373400 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-r8m5k" event={"ID":"a40e02a7-900e-4da2-8a4a-73773ca90876","Type":"ContainerDied","Data":"20598aa4e239724712ce00389212a2b4a21ab17443b6890e775b769028651f20"} Oct 05 21:10:56 crc kubenswrapper[4754]: I1005 21:10:56.373460 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-r8m5k" Oct 05 21:10:56 crc kubenswrapper[4754]: I1005 21:10:56.373510 4754 scope.go:117] "RemoveContainer" containerID="20598aa4e239724712ce00389212a2b4a21ab17443b6890e775b769028651f20" Oct 05 21:10:56 crc kubenswrapper[4754]: I1005 21:10:56.373469 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-r8m5k" event={"ID":"a40e02a7-900e-4da2-8a4a-73773ca90876","Type":"ContainerDied","Data":"4426e1a6008d7785f3df3b9ede9af06a0344676ecb05f3f03ffd1e742721e76a"} Oct 05 21:10:56 crc kubenswrapper[4754]: I1005 21:10:56.377976 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-g8hwz" Oct 05 21:10:56 crc kubenswrapper[4754]: I1005 21:10:56.377976 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-g8hwz" event={"ID":"b8d124eb-a543-4988-9523-d0e37ca113b8","Type":"ContainerDied","Data":"dd7f567876f9909834dceb2405002aa7f28dfccf964a367843cf3ec670a1dca3"} Oct 05 21:10:56 crc kubenswrapper[4754]: I1005 21:10:56.378116 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dd7f567876f9909834dceb2405002aa7f28dfccf964a367843cf3ec670a1dca3" Oct 05 21:10:56 crc kubenswrapper[4754]: I1005 21:10:56.417756 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-r8m5k"] Oct 05 21:10:56 crc kubenswrapper[4754]: I1005 21:10:56.429292 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-r8m5k"] Oct 05 21:10:56 crc kubenswrapper[4754]: I1005 21:10:56.432089 4754 scope.go:117] "RemoveContainer" containerID="d0afd180d66849b57e27b8e3892816da0c5c63ff7bfef21d39cd5bbb23b8b76a" Oct 05 21:10:56 crc kubenswrapper[4754]: I1005 21:10:56.470948 4754 scope.go:117] "RemoveContainer" containerID="20598aa4e239724712ce00389212a2b4a21ab17443b6890e775b769028651f20" Oct 05 21:10:56 crc kubenswrapper[4754]: E1005 21:10:56.471668 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"20598aa4e239724712ce00389212a2b4a21ab17443b6890e775b769028651f20\": container with ID starting with 20598aa4e239724712ce00389212a2b4a21ab17443b6890e775b769028651f20 not found: ID does not exist" containerID="20598aa4e239724712ce00389212a2b4a21ab17443b6890e775b769028651f20" Oct 05 21:10:56 crc kubenswrapper[4754]: I1005 21:10:56.471741 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"20598aa4e239724712ce00389212a2b4a21ab17443b6890e775b769028651f20"} err="failed to get container status \"20598aa4e239724712ce00389212a2b4a21ab17443b6890e775b769028651f20\": rpc error: code = NotFound desc = could not find container \"20598aa4e239724712ce00389212a2b4a21ab17443b6890e775b769028651f20\": container with ID starting with 20598aa4e239724712ce00389212a2b4a21ab17443b6890e775b769028651f20 not found: ID does not exist" Oct 05 21:10:56 crc kubenswrapper[4754]: I1005 21:10:56.471773 4754 scope.go:117] "RemoveContainer" containerID="d0afd180d66849b57e27b8e3892816da0c5c63ff7bfef21d39cd5bbb23b8b76a" Oct 05 21:10:56 crc kubenswrapper[4754]: E1005 21:10:56.472183 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d0afd180d66849b57e27b8e3892816da0c5c63ff7bfef21d39cd5bbb23b8b76a\": container with ID starting with d0afd180d66849b57e27b8e3892816da0c5c63ff7bfef21d39cd5bbb23b8b76a not found: ID does not exist" containerID="d0afd180d66849b57e27b8e3892816da0c5c63ff7bfef21d39cd5bbb23b8b76a" Oct 05 21:10:56 crc kubenswrapper[4754]: I1005 21:10:56.472223 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d0afd180d66849b57e27b8e3892816da0c5c63ff7bfef21d39cd5bbb23b8b76a"} err="failed to get container status \"d0afd180d66849b57e27b8e3892816da0c5c63ff7bfef21d39cd5bbb23b8b76a\": rpc error: code = NotFound desc = could not find container \"d0afd180d66849b57e27b8e3892816da0c5c63ff7bfef21d39cd5bbb23b8b76a\": container with ID starting with d0afd180d66849b57e27b8e3892816da0c5c63ff7bfef21d39cd5bbb23b8b76a not found: ID does not exist" Oct 05 21:10:56 crc kubenswrapper[4754]: I1005 21:10:56.626162 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-5h8xj" Oct 05 21:10:56 crc kubenswrapper[4754]: I1005 21:10:56.727642 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bwmm5\" (UniqueName: \"kubernetes.io/projected/cee454de-74cf-4913-8792-a231ddd4c691-kube-api-access-bwmm5\") pod \"cee454de-74cf-4913-8792-a231ddd4c691\" (UID: \"cee454de-74cf-4913-8792-a231ddd4c691\") " Oct 05 21:10:56 crc kubenswrapper[4754]: I1005 21:10:56.733342 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cee454de-74cf-4913-8792-a231ddd4c691-kube-api-access-bwmm5" (OuterVolumeSpecName: "kube-api-access-bwmm5") pod "cee454de-74cf-4913-8792-a231ddd4c691" (UID: "cee454de-74cf-4913-8792-a231ddd4c691"). InnerVolumeSpecName "kube-api-access-bwmm5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:10:56 crc kubenswrapper[4754]: I1005 21:10:56.798087 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-c7gj8" Oct 05 21:10:56 crc kubenswrapper[4754]: I1005 21:10:56.832800 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bwmm5\" (UniqueName: \"kubernetes.io/projected/cee454de-74cf-4913-8792-a231ddd4c691-kube-api-access-bwmm5\") on node \"crc\" DevicePath \"\"" Oct 05 21:10:56 crc kubenswrapper[4754]: I1005 21:10:56.853314 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a40e02a7-900e-4da2-8a4a-73773ca90876" path="/var/lib/kubelet/pods/a40e02a7-900e-4da2-8a4a-73773ca90876/volumes" Oct 05 21:10:56 crc kubenswrapper[4754]: I1005 21:10:56.934288 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fvjck\" (UniqueName: \"kubernetes.io/projected/28dab52f-2bfa-409f-8ec3-b47ae0641ce2-kube-api-access-fvjck\") pod \"28dab52f-2bfa-409f-8ec3-b47ae0641ce2\" (UID: \"28dab52f-2bfa-409f-8ec3-b47ae0641ce2\") " Oct 05 21:10:56 crc kubenswrapper[4754]: I1005 21:10:56.948701 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/28dab52f-2bfa-409f-8ec3-b47ae0641ce2-kube-api-access-fvjck" (OuterVolumeSpecName: "kube-api-access-fvjck") pod "28dab52f-2bfa-409f-8ec3-b47ae0641ce2" (UID: "28dab52f-2bfa-409f-8ec3-b47ae0641ce2"). InnerVolumeSpecName "kube-api-access-fvjck". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:10:57 crc kubenswrapper[4754]: I1005 21:10:57.036869 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fvjck\" (UniqueName: \"kubernetes.io/projected/28dab52f-2bfa-409f-8ec3-b47ae0641ce2-kube-api-access-fvjck\") on node \"crc\" DevicePath \"\"" Oct 05 21:10:57 crc kubenswrapper[4754]: I1005 21:10:57.397362 4754 generic.go:334] "Generic (PLEG): container finished" podID="1d96e280-8c26-41ea-ab29-64ce5b9c6a2c" containerID="1a0e1b3cbb7432cf1c8d8d5b93db680fd2af14a635594bd4eb551779f7e4ba30" exitCode=0 Oct 05 21:10:57 crc kubenswrapper[4754]: I1005 21:10:57.397562 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"1d96e280-8c26-41ea-ab29-64ce5b9c6a2c","Type":"ContainerDied","Data":"1a0e1b3cbb7432cf1c8d8d5b93db680fd2af14a635594bd4eb551779f7e4ba30"} Oct 05 21:10:57 crc kubenswrapper[4754]: I1005 21:10:57.400695 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-5h8xj" event={"ID":"cee454de-74cf-4913-8792-a231ddd4c691","Type":"ContainerDied","Data":"4a98b3ea4e794078a58aecea9e0cf8486b7c1cc19be1c68a6534e50000454f5b"} Oct 05 21:10:57 crc kubenswrapper[4754]: I1005 21:10:57.400788 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4a98b3ea4e794078a58aecea9e0cf8486b7c1cc19be1c68a6534e50000454f5b" Oct 05 21:10:57 crc kubenswrapper[4754]: I1005 21:10:57.400882 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-5h8xj" Oct 05 21:10:57 crc kubenswrapper[4754]: I1005 21:10:57.412940 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-c7gj8" event={"ID":"28dab52f-2bfa-409f-8ec3-b47ae0641ce2","Type":"ContainerDied","Data":"1036385fff7b217e6fde18404b163e65732a0261a276a04863fe63810076a4e1"} Oct 05 21:10:57 crc kubenswrapper[4754]: I1005 21:10:57.412975 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1036385fff7b217e6fde18404b163e65732a0261a276a04863fe63810076a4e1" Oct 05 21:10:57 crc kubenswrapper[4754]: I1005 21:10:57.413028 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-c7gj8" Oct 05 21:10:58 crc kubenswrapper[4754]: I1005 21:10:58.426537 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"1d96e280-8c26-41ea-ab29-64ce5b9c6a2c","Type":"ContainerStarted","Data":"ea5291016f4053d149506d9f4e1a17c9d6281d48b900045f7a5bc743cc386a39"} Oct 05 21:10:58 crc kubenswrapper[4754]: I1005 21:10:58.427075 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Oct 05 21:10:58 crc kubenswrapper[4754]: I1005 21:10:58.431743 4754 generic.go:334] "Generic (PLEG): container finished" podID="39aff8a0-2905-4ad3-ad2d-5bf6b5de3858" containerID="493f2a671031a6ad481270c980a4f16b3d2567e2e2da7677960e8f9fba245eaf" exitCode=0 Oct 05 21:10:58 crc kubenswrapper[4754]: I1005 21:10:58.431797 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"39aff8a0-2905-4ad3-ad2d-5bf6b5de3858","Type":"ContainerDied","Data":"493f2a671031a6ad481270c980a4f16b3d2567e2e2da7677960e8f9fba245eaf"} Oct 05 21:10:58 crc kubenswrapper[4754]: I1005 21:10:58.454288 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=37.691747018 podStartE2EDuration="1m20.454266284s" podCreationTimestamp="2025-10-05 21:09:38 +0000 UTC" firstStartedPulling="2025-10-05 21:09:41.089761993 +0000 UTC m=+904.993880703" lastFinishedPulling="2025-10-05 21:10:23.852281259 +0000 UTC m=+947.756399969" observedRunningTime="2025-10-05 21:10:58.452317002 +0000 UTC m=+982.356435732" watchObservedRunningTime="2025-10-05 21:10:58.454266284 +0000 UTC m=+982.358384994" Oct 05 21:10:59 crc kubenswrapper[4754]: I1005 21:10:59.444612 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"39aff8a0-2905-4ad3-ad2d-5bf6b5de3858","Type":"ContainerStarted","Data":"e4012712472c94cd11327163abb6c6f5c55cff5c6625d3242d418e93290fa5bf"} Oct 05 21:10:59 crc kubenswrapper[4754]: I1005 21:10:59.446291 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Oct 05 21:10:59 crc kubenswrapper[4754]: I1005 21:10:59.474612 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=-9223371955.380188 podStartE2EDuration="1m21.474588641s" podCreationTimestamp="2025-10-05 21:09:38 +0000 UTC" firstStartedPulling="2025-10-05 21:09:40.502003965 +0000 UTC m=+904.406122675" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 21:10:59.47043881 +0000 UTC m=+983.374557520" watchObservedRunningTime="2025-10-05 21:10:59.474588641 +0000 UTC m=+983.378707341" Oct 05 21:11:00 crc kubenswrapper[4754]: I1005 21:11:00.454570 4754 generic.go:334] "Generic (PLEG): container finished" podID="89a55f59-6693-4e0d-b8f8-dfb258d7a078" containerID="d2805c7e07be7017d0492b595038648706c23628e8be64e28d3484d0af8becc1" exitCode=0 Oct 05 21:11:00 crc kubenswrapper[4754]: I1005 21:11:00.454623 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-ljsff" event={"ID":"89a55f59-6693-4e0d-b8f8-dfb258d7a078","Type":"ContainerDied","Data":"d2805c7e07be7017d0492b595038648706c23628e8be64e28d3484d0af8becc1"} Oct 05 21:11:00 crc kubenswrapper[4754]: I1005 21:11:00.962762 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Oct 05 21:11:01 crc kubenswrapper[4754]: I1005 21:11:01.826046 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-ljsff" Oct 05 21:11:01 crc kubenswrapper[4754]: I1005 21:11:01.915101 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/89a55f59-6693-4e0d-b8f8-dfb258d7a078-dispersionconf\") pod \"89a55f59-6693-4e0d-b8f8-dfb258d7a078\" (UID: \"89a55f59-6693-4e0d-b8f8-dfb258d7a078\") " Oct 05 21:11:01 crc kubenswrapper[4754]: I1005 21:11:01.915240 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/89a55f59-6693-4e0d-b8f8-dfb258d7a078-ring-data-devices\") pod \"89a55f59-6693-4e0d-b8f8-dfb258d7a078\" (UID: \"89a55f59-6693-4e0d-b8f8-dfb258d7a078\") " Oct 05 21:11:01 crc kubenswrapper[4754]: I1005 21:11:01.915268 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/89a55f59-6693-4e0d-b8f8-dfb258d7a078-etc-swift\") pod \"89a55f59-6693-4e0d-b8f8-dfb258d7a078\" (UID: \"89a55f59-6693-4e0d-b8f8-dfb258d7a078\") " Oct 05 21:11:01 crc kubenswrapper[4754]: I1005 21:11:01.915296 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/89a55f59-6693-4e0d-b8f8-dfb258d7a078-scripts\") pod \"89a55f59-6693-4e0d-b8f8-dfb258d7a078\" (UID: \"89a55f59-6693-4e0d-b8f8-dfb258d7a078\") " Oct 05 21:11:01 crc kubenswrapper[4754]: I1005 21:11:01.915331 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/89a55f59-6693-4e0d-b8f8-dfb258d7a078-combined-ca-bundle\") pod \"89a55f59-6693-4e0d-b8f8-dfb258d7a078\" (UID: \"89a55f59-6693-4e0d-b8f8-dfb258d7a078\") " Oct 05 21:11:01 crc kubenswrapper[4754]: I1005 21:11:01.915360 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wmphp\" (UniqueName: \"kubernetes.io/projected/89a55f59-6693-4e0d-b8f8-dfb258d7a078-kube-api-access-wmphp\") pod \"89a55f59-6693-4e0d-b8f8-dfb258d7a078\" (UID: \"89a55f59-6693-4e0d-b8f8-dfb258d7a078\") " Oct 05 21:11:01 crc kubenswrapper[4754]: I1005 21:11:01.916158 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/89a55f59-6693-4e0d-b8f8-dfb258d7a078-swiftconf\") pod \"89a55f59-6693-4e0d-b8f8-dfb258d7a078\" (UID: \"89a55f59-6693-4e0d-b8f8-dfb258d7a078\") " Oct 05 21:11:01 crc kubenswrapper[4754]: I1005 21:11:01.916002 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/89a55f59-6693-4e0d-b8f8-dfb258d7a078-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "89a55f59-6693-4e0d-b8f8-dfb258d7a078" (UID: "89a55f59-6693-4e0d-b8f8-dfb258d7a078"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 21:11:01 crc kubenswrapper[4754]: I1005 21:11:01.916665 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/89a55f59-6693-4e0d-b8f8-dfb258d7a078-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "89a55f59-6693-4e0d-b8f8-dfb258d7a078" (UID: "89a55f59-6693-4e0d-b8f8-dfb258d7a078"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 21:11:01 crc kubenswrapper[4754]: I1005 21:11:01.916778 4754 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/89a55f59-6693-4e0d-b8f8-dfb258d7a078-ring-data-devices\") on node \"crc\" DevicePath \"\"" Oct 05 21:11:01 crc kubenswrapper[4754]: I1005 21:11:01.916793 4754 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/89a55f59-6693-4e0d-b8f8-dfb258d7a078-etc-swift\") on node \"crc\" DevicePath \"\"" Oct 05 21:11:01 crc kubenswrapper[4754]: I1005 21:11:01.941765 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/89a55f59-6693-4e0d-b8f8-dfb258d7a078-kube-api-access-wmphp" (OuterVolumeSpecName: "kube-api-access-wmphp") pod "89a55f59-6693-4e0d-b8f8-dfb258d7a078" (UID: "89a55f59-6693-4e0d-b8f8-dfb258d7a078"). InnerVolumeSpecName "kube-api-access-wmphp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:11:01 crc kubenswrapper[4754]: I1005 21:11:01.956992 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/89a55f59-6693-4e0d-b8f8-dfb258d7a078-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "89a55f59-6693-4e0d-b8f8-dfb258d7a078" (UID: "89a55f59-6693-4e0d-b8f8-dfb258d7a078"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:11:01 crc kubenswrapper[4754]: I1005 21:11:01.961955 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/89a55f59-6693-4e0d-b8f8-dfb258d7a078-scripts" (OuterVolumeSpecName: "scripts") pod "89a55f59-6693-4e0d-b8f8-dfb258d7a078" (UID: "89a55f59-6693-4e0d-b8f8-dfb258d7a078"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 21:11:01 crc kubenswrapper[4754]: I1005 21:11:01.965754 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/89a55f59-6693-4e0d-b8f8-dfb258d7a078-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "89a55f59-6693-4e0d-b8f8-dfb258d7a078" (UID: "89a55f59-6693-4e0d-b8f8-dfb258d7a078"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:11:01 crc kubenswrapper[4754]: I1005 21:11:01.977304 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/89a55f59-6693-4e0d-b8f8-dfb258d7a078-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "89a55f59-6693-4e0d-b8f8-dfb258d7a078" (UID: "89a55f59-6693-4e0d-b8f8-dfb258d7a078"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:11:02 crc kubenswrapper[4754]: I1005 21:11:02.020022 4754 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/89a55f59-6693-4e0d-b8f8-dfb258d7a078-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 21:11:02 crc kubenswrapper[4754]: I1005 21:11:02.020287 4754 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/89a55f59-6693-4e0d-b8f8-dfb258d7a078-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 21:11:02 crc kubenswrapper[4754]: I1005 21:11:02.020351 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wmphp\" (UniqueName: \"kubernetes.io/projected/89a55f59-6693-4e0d-b8f8-dfb258d7a078-kube-api-access-wmphp\") on node \"crc\" DevicePath \"\"" Oct 05 21:11:02 crc kubenswrapper[4754]: I1005 21:11:02.020440 4754 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/89a55f59-6693-4e0d-b8f8-dfb258d7a078-swiftconf\") on node \"crc\" DevicePath \"\"" Oct 05 21:11:02 crc kubenswrapper[4754]: I1005 21:11:02.020518 4754 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/89a55f59-6693-4e0d-b8f8-dfb258d7a078-dispersionconf\") on node \"crc\" DevicePath \"\"" Oct 05 21:11:02 crc kubenswrapper[4754]: I1005 21:11:02.472248 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-ljsff" event={"ID":"89a55f59-6693-4e0d-b8f8-dfb258d7a078","Type":"ContainerDied","Data":"a0c4a11b0dcc0302ed3649ca159b0d1debc77c5994d3896ecd01551568da0fc7"} Oct 05 21:11:02 crc kubenswrapper[4754]: I1005 21:11:02.472511 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a0c4a11b0dcc0302ed3649ca159b0d1debc77c5994d3896ecd01551568da0fc7" Oct 05 21:11:02 crc kubenswrapper[4754]: I1005 21:11:02.472300 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-ljsff" Oct 05 21:11:02 crc kubenswrapper[4754]: I1005 21:11:02.527842 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/9af24378-0013-4510-aee6-d2fe2172935c-etc-swift\") pod \"swift-storage-0\" (UID: \"9af24378-0013-4510-aee6-d2fe2172935c\") " pod="openstack/swift-storage-0" Oct 05 21:11:02 crc kubenswrapper[4754]: I1005 21:11:02.533130 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/9af24378-0013-4510-aee6-d2fe2172935c-etc-swift\") pod \"swift-storage-0\" (UID: \"9af24378-0013-4510-aee6-d2fe2172935c\") " pod="openstack/swift-storage-0" Oct 05 21:11:02 crc kubenswrapper[4754]: I1005 21:11:02.589776 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Oct 05 21:11:03 crc kubenswrapper[4754]: I1005 21:11:03.175921 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-e07b-account-create-wkgh8"] Oct 05 21:11:03 crc kubenswrapper[4754]: E1005 21:11:03.177693 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a40e02a7-900e-4da2-8a4a-73773ca90876" containerName="init" Oct 05 21:11:03 crc kubenswrapper[4754]: I1005 21:11:03.177782 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="a40e02a7-900e-4da2-8a4a-73773ca90876" containerName="init" Oct 05 21:11:03 crc kubenswrapper[4754]: E1005 21:11:03.177847 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cee454de-74cf-4913-8792-a231ddd4c691" containerName="mariadb-database-create" Oct 05 21:11:03 crc kubenswrapper[4754]: I1005 21:11:03.177900 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="cee454de-74cf-4913-8792-a231ddd4c691" containerName="mariadb-database-create" Oct 05 21:11:03 crc kubenswrapper[4754]: E1005 21:11:03.177990 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a40e02a7-900e-4da2-8a4a-73773ca90876" containerName="dnsmasq-dns" Oct 05 21:11:03 crc kubenswrapper[4754]: I1005 21:11:03.178075 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="a40e02a7-900e-4da2-8a4a-73773ca90876" containerName="dnsmasq-dns" Oct 05 21:11:03 crc kubenswrapper[4754]: E1005 21:11:03.178150 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8d124eb-a543-4988-9523-d0e37ca113b8" containerName="mariadb-database-create" Oct 05 21:11:03 crc kubenswrapper[4754]: I1005 21:11:03.178202 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8d124eb-a543-4988-9523-d0e37ca113b8" containerName="mariadb-database-create" Oct 05 21:11:03 crc kubenswrapper[4754]: E1005 21:11:03.178258 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89a55f59-6693-4e0d-b8f8-dfb258d7a078" containerName="swift-ring-rebalance" Oct 05 21:11:03 crc kubenswrapper[4754]: I1005 21:11:03.178318 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="89a55f59-6693-4e0d-b8f8-dfb258d7a078" containerName="swift-ring-rebalance" Oct 05 21:11:03 crc kubenswrapper[4754]: E1005 21:11:03.178385 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28dab52f-2bfa-409f-8ec3-b47ae0641ce2" containerName="mariadb-database-create" Oct 05 21:11:03 crc kubenswrapper[4754]: I1005 21:11:03.178442 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="28dab52f-2bfa-409f-8ec3-b47ae0641ce2" containerName="mariadb-database-create" Oct 05 21:11:03 crc kubenswrapper[4754]: I1005 21:11:03.178731 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="89a55f59-6693-4e0d-b8f8-dfb258d7a078" containerName="swift-ring-rebalance" Oct 05 21:11:03 crc kubenswrapper[4754]: I1005 21:11:03.178804 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="b8d124eb-a543-4988-9523-d0e37ca113b8" containerName="mariadb-database-create" Oct 05 21:11:03 crc kubenswrapper[4754]: I1005 21:11:03.178860 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="cee454de-74cf-4913-8792-a231ddd4c691" containerName="mariadb-database-create" Oct 05 21:11:03 crc kubenswrapper[4754]: I1005 21:11:03.178912 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="a40e02a7-900e-4da2-8a4a-73773ca90876" containerName="dnsmasq-dns" Oct 05 21:11:03 crc kubenswrapper[4754]: I1005 21:11:03.178988 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="28dab52f-2bfa-409f-8ec3-b47ae0641ce2" containerName="mariadb-database-create" Oct 05 21:11:03 crc kubenswrapper[4754]: I1005 21:11:03.179616 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-e07b-account-create-wkgh8" Oct 05 21:11:03 crc kubenswrapper[4754]: I1005 21:11:03.186358 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Oct 05 21:11:03 crc kubenswrapper[4754]: I1005 21:11:03.190211 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Oct 05 21:11:03 crc kubenswrapper[4754]: W1005 21:11:03.196223 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9af24378_0013_4510_aee6_d2fe2172935c.slice/crio-96e362847bae5e8c3490eeed20226051d03fbd3b888c0eb0b789a9e519d3d9e6 WatchSource:0}: Error finding container 96e362847bae5e8c3490eeed20226051d03fbd3b888c0eb0b789a9e519d3d9e6: Status 404 returned error can't find the container with id 96e362847bae5e8c3490eeed20226051d03fbd3b888c0eb0b789a9e519d3d9e6 Oct 05 21:11:03 crc kubenswrapper[4754]: I1005 21:11:03.215626 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-e07b-account-create-wkgh8"] Oct 05 21:11:03 crc kubenswrapper[4754]: I1005 21:11:03.241410 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-52dgc\" (UniqueName: \"kubernetes.io/projected/e3ddc43f-cb95-4106-a5ff-09fc6320aad8-kube-api-access-52dgc\") pod \"keystone-e07b-account-create-wkgh8\" (UID: \"e3ddc43f-cb95-4106-a5ff-09fc6320aad8\") " pod="openstack/keystone-e07b-account-create-wkgh8" Oct 05 21:11:03 crc kubenswrapper[4754]: I1005 21:11:03.332456 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-5428-account-create-gbvmw"] Oct 05 21:11:03 crc kubenswrapper[4754]: I1005 21:11:03.334045 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-5428-account-create-gbvmw" Oct 05 21:11:03 crc kubenswrapper[4754]: I1005 21:11:03.336000 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Oct 05 21:11:03 crc kubenswrapper[4754]: I1005 21:11:03.342543 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-52dgc\" (UniqueName: \"kubernetes.io/projected/e3ddc43f-cb95-4106-a5ff-09fc6320aad8-kube-api-access-52dgc\") pod \"keystone-e07b-account-create-wkgh8\" (UID: \"e3ddc43f-cb95-4106-a5ff-09fc6320aad8\") " pod="openstack/keystone-e07b-account-create-wkgh8" Oct 05 21:11:03 crc kubenswrapper[4754]: I1005 21:11:03.356010 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-5428-account-create-gbvmw"] Oct 05 21:11:03 crc kubenswrapper[4754]: I1005 21:11:03.398748 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-52dgc\" (UniqueName: \"kubernetes.io/projected/e3ddc43f-cb95-4106-a5ff-09fc6320aad8-kube-api-access-52dgc\") pod \"keystone-e07b-account-create-wkgh8\" (UID: \"e3ddc43f-cb95-4106-a5ff-09fc6320aad8\") " pod="openstack/keystone-e07b-account-create-wkgh8" Oct 05 21:11:03 crc kubenswrapper[4754]: I1005 21:11:03.443955 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-85gcf\" (UniqueName: \"kubernetes.io/projected/21391bdb-ce3c-45ed-bc20-1bccdbcc3c72-kube-api-access-85gcf\") pod \"placement-5428-account-create-gbvmw\" (UID: \"21391bdb-ce3c-45ed-bc20-1bccdbcc3c72\") " pod="openstack/placement-5428-account-create-gbvmw" Oct 05 21:11:03 crc kubenswrapper[4754]: I1005 21:11:03.479433 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"9af24378-0013-4510-aee6-d2fe2172935c","Type":"ContainerStarted","Data":"96e362847bae5e8c3490eeed20226051d03fbd3b888c0eb0b789a9e519d3d9e6"} Oct 05 21:11:03 crc kubenswrapper[4754]: I1005 21:11:03.545808 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-85gcf\" (UniqueName: \"kubernetes.io/projected/21391bdb-ce3c-45ed-bc20-1bccdbcc3c72-kube-api-access-85gcf\") pod \"placement-5428-account-create-gbvmw\" (UID: \"21391bdb-ce3c-45ed-bc20-1bccdbcc3c72\") " pod="openstack/placement-5428-account-create-gbvmw" Oct 05 21:11:03 crc kubenswrapper[4754]: I1005 21:11:03.551428 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-e07b-account-create-wkgh8" Oct 05 21:11:03 crc kubenswrapper[4754]: I1005 21:11:03.562018 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-85gcf\" (UniqueName: \"kubernetes.io/projected/21391bdb-ce3c-45ed-bc20-1bccdbcc3c72-kube-api-access-85gcf\") pod \"placement-5428-account-create-gbvmw\" (UID: \"21391bdb-ce3c-45ed-bc20-1bccdbcc3c72\") " pod="openstack/placement-5428-account-create-gbvmw" Oct 05 21:11:03 crc kubenswrapper[4754]: I1005 21:11:03.652822 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-5428-account-create-gbvmw" Oct 05 21:11:03 crc kubenswrapper[4754]: I1005 21:11:03.792472 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-8228-account-create-pf59x"] Oct 05 21:11:03 crc kubenswrapper[4754]: I1005 21:11:03.793464 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-8228-account-create-pf59x" Oct 05 21:11:03 crc kubenswrapper[4754]: I1005 21:11:03.799252 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Oct 05 21:11:03 crc kubenswrapper[4754]: I1005 21:11:03.813981 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-8228-account-create-pf59x"] Oct 05 21:11:03 crc kubenswrapper[4754]: I1005 21:11:03.857061 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x97ps\" (UniqueName: \"kubernetes.io/projected/5b380a91-da43-41a4-afcf-3d41e99b094f-kube-api-access-x97ps\") pod \"glance-8228-account-create-pf59x\" (UID: \"5b380a91-da43-41a4-afcf-3d41e99b094f\") " pod="openstack/glance-8228-account-create-pf59x" Oct 05 21:11:03 crc kubenswrapper[4754]: I1005 21:11:03.958235 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x97ps\" (UniqueName: \"kubernetes.io/projected/5b380a91-da43-41a4-afcf-3d41e99b094f-kube-api-access-x97ps\") pod \"glance-8228-account-create-pf59x\" (UID: \"5b380a91-da43-41a4-afcf-3d41e99b094f\") " pod="openstack/glance-8228-account-create-pf59x" Oct 05 21:11:03 crc kubenswrapper[4754]: I1005 21:11:03.981847 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x97ps\" (UniqueName: \"kubernetes.io/projected/5b380a91-da43-41a4-afcf-3d41e99b094f-kube-api-access-x97ps\") pod \"glance-8228-account-create-pf59x\" (UID: \"5b380a91-da43-41a4-afcf-3d41e99b094f\") " pod="openstack/glance-8228-account-create-pf59x" Oct 05 21:11:04 crc kubenswrapper[4754]: I1005 21:11:04.036049 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-e07b-account-create-wkgh8"] Oct 05 21:11:04 crc kubenswrapper[4754]: W1005 21:11:04.046847 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode3ddc43f_cb95_4106_a5ff_09fc6320aad8.slice/crio-70141fe5c945e196a717c359dfdbd2155a74e63b9367d1964d1c5df1c9bb52d2 WatchSource:0}: Error finding container 70141fe5c945e196a717c359dfdbd2155a74e63b9367d1964d1c5df1c9bb52d2: Status 404 returned error can't find the container with id 70141fe5c945e196a717c359dfdbd2155a74e63b9367d1964d1c5df1c9bb52d2 Oct 05 21:11:04 crc kubenswrapper[4754]: I1005 21:11:04.122608 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-8228-account-create-pf59x" Oct 05 21:11:04 crc kubenswrapper[4754]: I1005 21:11:04.215900 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-5428-account-create-gbvmw"] Oct 05 21:11:04 crc kubenswrapper[4754]: I1005 21:11:04.498609 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-e07b-account-create-wkgh8" event={"ID":"e3ddc43f-cb95-4106-a5ff-09fc6320aad8","Type":"ContainerStarted","Data":"70141fe5c945e196a717c359dfdbd2155a74e63b9367d1964d1c5df1c9bb52d2"} Oct 05 21:11:04 crc kubenswrapper[4754]: I1005 21:11:04.500631 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-5428-account-create-gbvmw" event={"ID":"21391bdb-ce3c-45ed-bc20-1bccdbcc3c72","Type":"ContainerStarted","Data":"684677c5bedf56c3583beb7f226abfbf4d2708174fc5174c20b2441d2a43d98f"} Oct 05 21:11:04 crc kubenswrapper[4754]: I1005 21:11:04.500674 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-5428-account-create-gbvmw" event={"ID":"21391bdb-ce3c-45ed-bc20-1bccdbcc3c72","Type":"ContainerStarted","Data":"0eed24dcfa8f4a1a39b051186024fd85039e7a4e3678068943c7b4d31956acda"} Oct 05 21:11:04 crc kubenswrapper[4754]: I1005 21:11:04.518679 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-5428-account-create-gbvmw" podStartSLOduration=1.518659582 podStartE2EDuration="1.518659582s" podCreationTimestamp="2025-10-05 21:11:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 21:11:04.513957036 +0000 UTC m=+988.418075746" watchObservedRunningTime="2025-10-05 21:11:04.518659582 +0000 UTC m=+988.422778292" Oct 05 21:11:04 crc kubenswrapper[4754]: I1005 21:11:04.710432 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-8228-account-create-pf59x"] Oct 05 21:11:05 crc kubenswrapper[4754]: I1005 21:11:05.511664 4754 generic.go:334] "Generic (PLEG): container finished" podID="21391bdb-ce3c-45ed-bc20-1bccdbcc3c72" containerID="684677c5bedf56c3583beb7f226abfbf4d2708174fc5174c20b2441d2a43d98f" exitCode=0 Oct 05 21:11:05 crc kubenswrapper[4754]: I1005 21:11:05.511946 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-5428-account-create-gbvmw" event={"ID":"21391bdb-ce3c-45ed-bc20-1bccdbcc3c72","Type":"ContainerDied","Data":"684677c5bedf56c3583beb7f226abfbf4d2708174fc5174c20b2441d2a43d98f"} Oct 05 21:11:05 crc kubenswrapper[4754]: I1005 21:11:05.518540 4754 generic.go:334] "Generic (PLEG): container finished" podID="5b380a91-da43-41a4-afcf-3d41e99b094f" containerID="41b465148330f3be15a27262db77f02a260ba3a73ea227b2e2caa0a0795e677d" exitCode=0 Oct 05 21:11:05 crc kubenswrapper[4754]: I1005 21:11:05.518635 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-8228-account-create-pf59x" event={"ID":"5b380a91-da43-41a4-afcf-3d41e99b094f","Type":"ContainerDied","Data":"41b465148330f3be15a27262db77f02a260ba3a73ea227b2e2caa0a0795e677d"} Oct 05 21:11:05 crc kubenswrapper[4754]: I1005 21:11:05.518678 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-8228-account-create-pf59x" event={"ID":"5b380a91-da43-41a4-afcf-3d41e99b094f","Type":"ContainerStarted","Data":"5ae9f820b82947f48081ec1b3376dfdcd1bf9d116e0aa59cba696ff76e20affa"} Oct 05 21:11:05 crc kubenswrapper[4754]: I1005 21:11:05.523820 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"9af24378-0013-4510-aee6-d2fe2172935c","Type":"ContainerStarted","Data":"670f83f0348c4bfa56056297c14b14a3460b9fde772652a813e9c78251203370"} Oct 05 21:11:05 crc kubenswrapper[4754]: I1005 21:11:05.523874 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"9af24378-0013-4510-aee6-d2fe2172935c","Type":"ContainerStarted","Data":"b9c1d49c59d6df6f324dcb5c064ebe5b5b519bfa328bbe16c93a1fe2619f81c6"} Oct 05 21:11:05 crc kubenswrapper[4754]: I1005 21:11:05.523888 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"9af24378-0013-4510-aee6-d2fe2172935c","Type":"ContainerStarted","Data":"82803d530ca833822eb6f10f879176c964b5cee19918b3fdca56c5f9609a7fe9"} Oct 05 21:11:05 crc kubenswrapper[4754]: I1005 21:11:05.525533 4754 generic.go:334] "Generic (PLEG): container finished" podID="e3ddc43f-cb95-4106-a5ff-09fc6320aad8" containerID="89f838a779a6be03f73263b78506025ee248a5cc9827aa7981ac52d5d68aab78" exitCode=0 Oct 05 21:11:05 crc kubenswrapper[4754]: I1005 21:11:05.525581 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-e07b-account-create-wkgh8" event={"ID":"e3ddc43f-cb95-4106-a5ff-09fc6320aad8","Type":"ContainerDied","Data":"89f838a779a6be03f73263b78506025ee248a5cc9827aa7981ac52d5d68aab78"} Oct 05 21:11:06 crc kubenswrapper[4754]: I1005 21:11:06.552898 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"9af24378-0013-4510-aee6-d2fe2172935c","Type":"ContainerStarted","Data":"26dbae24144bf1af19359923f0e3ddd8c5074232f5b647e6886d8cba7145180f"} Oct 05 21:11:07 crc kubenswrapper[4754]: I1005 21:11:07.243916 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-e07b-account-create-wkgh8" Oct 05 21:11:07 crc kubenswrapper[4754]: I1005 21:11:07.251140 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-5428-account-create-gbvmw" Oct 05 21:11:07 crc kubenswrapper[4754]: I1005 21:11:07.257435 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-8228-account-create-pf59x" Oct 05 21:11:07 crc kubenswrapper[4754]: I1005 21:11:07.408411 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-52dgc\" (UniqueName: \"kubernetes.io/projected/e3ddc43f-cb95-4106-a5ff-09fc6320aad8-kube-api-access-52dgc\") pod \"e3ddc43f-cb95-4106-a5ff-09fc6320aad8\" (UID: \"e3ddc43f-cb95-4106-a5ff-09fc6320aad8\") " Oct 05 21:11:07 crc kubenswrapper[4754]: I1005 21:11:07.408504 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-85gcf\" (UniqueName: \"kubernetes.io/projected/21391bdb-ce3c-45ed-bc20-1bccdbcc3c72-kube-api-access-85gcf\") pod \"21391bdb-ce3c-45ed-bc20-1bccdbcc3c72\" (UID: \"21391bdb-ce3c-45ed-bc20-1bccdbcc3c72\") " Oct 05 21:11:07 crc kubenswrapper[4754]: I1005 21:11:07.408524 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x97ps\" (UniqueName: \"kubernetes.io/projected/5b380a91-da43-41a4-afcf-3d41e99b094f-kube-api-access-x97ps\") pod \"5b380a91-da43-41a4-afcf-3d41e99b094f\" (UID: \"5b380a91-da43-41a4-afcf-3d41e99b094f\") " Oct 05 21:11:07 crc kubenswrapper[4754]: I1005 21:11:07.425348 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b380a91-da43-41a4-afcf-3d41e99b094f-kube-api-access-x97ps" (OuterVolumeSpecName: "kube-api-access-x97ps") pod "5b380a91-da43-41a4-afcf-3d41e99b094f" (UID: "5b380a91-da43-41a4-afcf-3d41e99b094f"). InnerVolumeSpecName "kube-api-access-x97ps". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:11:07 crc kubenswrapper[4754]: I1005 21:11:07.425716 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e3ddc43f-cb95-4106-a5ff-09fc6320aad8-kube-api-access-52dgc" (OuterVolumeSpecName: "kube-api-access-52dgc") pod "e3ddc43f-cb95-4106-a5ff-09fc6320aad8" (UID: "e3ddc43f-cb95-4106-a5ff-09fc6320aad8"). InnerVolumeSpecName "kube-api-access-52dgc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:11:07 crc kubenswrapper[4754]: I1005 21:11:07.425798 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/21391bdb-ce3c-45ed-bc20-1bccdbcc3c72-kube-api-access-85gcf" (OuterVolumeSpecName: "kube-api-access-85gcf") pod "21391bdb-ce3c-45ed-bc20-1bccdbcc3c72" (UID: "21391bdb-ce3c-45ed-bc20-1bccdbcc3c72"). InnerVolumeSpecName "kube-api-access-85gcf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:11:07 crc kubenswrapper[4754]: I1005 21:11:07.510394 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-52dgc\" (UniqueName: \"kubernetes.io/projected/e3ddc43f-cb95-4106-a5ff-09fc6320aad8-kube-api-access-52dgc\") on node \"crc\" DevicePath \"\"" Oct 05 21:11:07 crc kubenswrapper[4754]: I1005 21:11:07.510444 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-85gcf\" (UniqueName: \"kubernetes.io/projected/21391bdb-ce3c-45ed-bc20-1bccdbcc3c72-kube-api-access-85gcf\") on node \"crc\" DevicePath \"\"" Oct 05 21:11:07 crc kubenswrapper[4754]: I1005 21:11:07.510454 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x97ps\" (UniqueName: \"kubernetes.io/projected/5b380a91-da43-41a4-afcf-3d41e99b094f-kube-api-access-x97ps\") on node \"crc\" DevicePath \"\"" Oct 05 21:11:07 crc kubenswrapper[4754]: I1005 21:11:07.572842 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-8228-account-create-pf59x" Oct 05 21:11:07 crc kubenswrapper[4754]: I1005 21:11:07.574576 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-8228-account-create-pf59x" event={"ID":"5b380a91-da43-41a4-afcf-3d41e99b094f","Type":"ContainerDied","Data":"5ae9f820b82947f48081ec1b3376dfdcd1bf9d116e0aa59cba696ff76e20affa"} Oct 05 21:11:07 crc kubenswrapper[4754]: I1005 21:11:07.574665 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5ae9f820b82947f48081ec1b3376dfdcd1bf9d116e0aa59cba696ff76e20affa" Oct 05 21:11:07 crc kubenswrapper[4754]: I1005 21:11:07.587117 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"9af24378-0013-4510-aee6-d2fe2172935c","Type":"ContainerStarted","Data":"b9f1dd166b7369ce360c5033b4a540af37ef861110be6c962a2d453e62fbdfe0"} Oct 05 21:11:07 crc kubenswrapper[4754]: I1005 21:11:07.590582 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-e07b-account-create-wkgh8" event={"ID":"e3ddc43f-cb95-4106-a5ff-09fc6320aad8","Type":"ContainerDied","Data":"70141fe5c945e196a717c359dfdbd2155a74e63b9367d1964d1c5df1c9bb52d2"} Oct 05 21:11:07 crc kubenswrapper[4754]: I1005 21:11:07.590609 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="70141fe5c945e196a717c359dfdbd2155a74e63b9367d1964d1c5df1c9bb52d2" Oct 05 21:11:07 crc kubenswrapper[4754]: I1005 21:11:07.590637 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-e07b-account-create-wkgh8" Oct 05 21:11:07 crc kubenswrapper[4754]: I1005 21:11:07.599811 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-5428-account-create-gbvmw" event={"ID":"21391bdb-ce3c-45ed-bc20-1bccdbcc3c72","Type":"ContainerDied","Data":"0eed24dcfa8f4a1a39b051186024fd85039e7a4e3678068943c7b4d31956acda"} Oct 05 21:11:07 crc kubenswrapper[4754]: I1005 21:11:07.599865 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0eed24dcfa8f4a1a39b051186024fd85039e7a4e3678068943c7b4d31956acda" Oct 05 21:11:07 crc kubenswrapper[4754]: I1005 21:11:07.599940 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-5428-account-create-gbvmw" Oct 05 21:11:08 crc kubenswrapper[4754]: I1005 21:11:08.611376 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"9af24378-0013-4510-aee6-d2fe2172935c","Type":"ContainerStarted","Data":"630dab50b23df3692d8c61e9f1231cd20bc5f93df43a1d12e74353e28ea77d23"} Oct 05 21:11:08 crc kubenswrapper[4754]: I1005 21:11:08.611737 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"9af24378-0013-4510-aee6-d2fe2172935c","Type":"ContainerStarted","Data":"2f477fde6c7a8b068291b22b39a094f01b421729d29c122a5b167281724a9c3f"} Oct 05 21:11:08 crc kubenswrapper[4754]: I1005 21:11:08.611750 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"9af24378-0013-4510-aee6-d2fe2172935c","Type":"ContainerStarted","Data":"b7c8c62bac60c9abf234742d26081abce97106d98138263335872cf0e50d1304"} Oct 05 21:11:08 crc kubenswrapper[4754]: I1005 21:11:08.787848 4754 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-pp728" podUID="e3785031-57bd-44ab-b30e-b609cca6dcaf" containerName="ovn-controller" probeResult="failure" output=< Oct 05 21:11:08 crc kubenswrapper[4754]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Oct 05 21:11:08 crc kubenswrapper[4754]: > Oct 05 21:11:08 crc kubenswrapper[4754]: I1005 21:11:08.817259 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-2krf2" Oct 05 21:11:08 crc kubenswrapper[4754]: I1005 21:11:08.853383 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-2krf2" Oct 05 21:11:08 crc kubenswrapper[4754]: I1005 21:11:08.969737 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-dvzwp"] Oct 05 21:11:08 crc kubenswrapper[4754]: E1005 21:11:08.970176 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b380a91-da43-41a4-afcf-3d41e99b094f" containerName="mariadb-account-create" Oct 05 21:11:08 crc kubenswrapper[4754]: I1005 21:11:08.970199 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b380a91-da43-41a4-afcf-3d41e99b094f" containerName="mariadb-account-create" Oct 05 21:11:08 crc kubenswrapper[4754]: E1005 21:11:08.970215 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3ddc43f-cb95-4106-a5ff-09fc6320aad8" containerName="mariadb-account-create" Oct 05 21:11:08 crc kubenswrapper[4754]: I1005 21:11:08.970223 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3ddc43f-cb95-4106-a5ff-09fc6320aad8" containerName="mariadb-account-create" Oct 05 21:11:08 crc kubenswrapper[4754]: E1005 21:11:08.970236 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21391bdb-ce3c-45ed-bc20-1bccdbcc3c72" containerName="mariadb-account-create" Oct 05 21:11:08 crc kubenswrapper[4754]: I1005 21:11:08.970244 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="21391bdb-ce3c-45ed-bc20-1bccdbcc3c72" containerName="mariadb-account-create" Oct 05 21:11:08 crc kubenswrapper[4754]: I1005 21:11:08.970432 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b380a91-da43-41a4-afcf-3d41e99b094f" containerName="mariadb-account-create" Oct 05 21:11:08 crc kubenswrapper[4754]: I1005 21:11:08.970486 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="21391bdb-ce3c-45ed-bc20-1bccdbcc3c72" containerName="mariadb-account-create" Oct 05 21:11:08 crc kubenswrapper[4754]: I1005 21:11:08.970513 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="e3ddc43f-cb95-4106-a5ff-09fc6320aad8" containerName="mariadb-account-create" Oct 05 21:11:08 crc kubenswrapper[4754]: I1005 21:11:08.971214 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-dvzwp" Oct 05 21:11:08 crc kubenswrapper[4754]: I1005 21:11:08.974658 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Oct 05 21:11:08 crc kubenswrapper[4754]: I1005 21:11:08.980230 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-gfwbt" Oct 05 21:11:09 crc kubenswrapper[4754]: I1005 21:11:09.003678 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-dvzwp"] Oct 05 21:11:09 crc kubenswrapper[4754]: I1005 21:11:09.101460 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-pp728-config-xd2gh"] Oct 05 21:11:09 crc kubenswrapper[4754]: I1005 21:11:09.102470 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-pp728-config-xd2gh" Oct 05 21:11:09 crc kubenswrapper[4754]: I1005 21:11:09.107073 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Oct 05 21:11:09 crc kubenswrapper[4754]: I1005 21:11:09.126937 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-pp728-config-xd2gh"] Oct 05 21:11:09 crc kubenswrapper[4754]: I1005 21:11:09.135572 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d20989a7-98b9-4bf2-9f85-d72890ef7e1a-combined-ca-bundle\") pod \"glance-db-sync-dvzwp\" (UID: \"d20989a7-98b9-4bf2-9f85-d72890ef7e1a\") " pod="openstack/glance-db-sync-dvzwp" Oct 05 21:11:09 crc kubenswrapper[4754]: I1005 21:11:09.135646 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d20989a7-98b9-4bf2-9f85-d72890ef7e1a-config-data\") pod \"glance-db-sync-dvzwp\" (UID: \"d20989a7-98b9-4bf2-9f85-d72890ef7e1a\") " pod="openstack/glance-db-sync-dvzwp" Oct 05 21:11:09 crc kubenswrapper[4754]: I1005 21:11:09.135725 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d20989a7-98b9-4bf2-9f85-d72890ef7e1a-db-sync-config-data\") pod \"glance-db-sync-dvzwp\" (UID: \"d20989a7-98b9-4bf2-9f85-d72890ef7e1a\") " pod="openstack/glance-db-sync-dvzwp" Oct 05 21:11:09 crc kubenswrapper[4754]: I1005 21:11:09.135987 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5rfdn\" (UniqueName: \"kubernetes.io/projected/d20989a7-98b9-4bf2-9f85-d72890ef7e1a-kube-api-access-5rfdn\") pod \"glance-db-sync-dvzwp\" (UID: \"d20989a7-98b9-4bf2-9f85-d72890ef7e1a\") " pod="openstack/glance-db-sync-dvzwp" Oct 05 21:11:09 crc kubenswrapper[4754]: I1005 21:11:09.237837 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d20989a7-98b9-4bf2-9f85-d72890ef7e1a-combined-ca-bundle\") pod \"glance-db-sync-dvzwp\" (UID: \"d20989a7-98b9-4bf2-9f85-d72890ef7e1a\") " pod="openstack/glance-db-sync-dvzwp" Oct 05 21:11:09 crc kubenswrapper[4754]: I1005 21:11:09.238129 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/c0a1bb02-14d5-4596-938d-5db342200716-var-log-ovn\") pod \"ovn-controller-pp728-config-xd2gh\" (UID: \"c0a1bb02-14d5-4596-938d-5db342200716\") " pod="openstack/ovn-controller-pp728-config-xd2gh" Oct 05 21:11:09 crc kubenswrapper[4754]: I1005 21:11:09.238156 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d20989a7-98b9-4bf2-9f85-d72890ef7e1a-config-data\") pod \"glance-db-sync-dvzwp\" (UID: \"d20989a7-98b9-4bf2-9f85-d72890ef7e1a\") " pod="openstack/glance-db-sync-dvzwp" Oct 05 21:11:09 crc kubenswrapper[4754]: I1005 21:11:09.238311 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c0a1bb02-14d5-4596-938d-5db342200716-scripts\") pod \"ovn-controller-pp728-config-xd2gh\" (UID: \"c0a1bb02-14d5-4596-938d-5db342200716\") " pod="openstack/ovn-controller-pp728-config-xd2gh" Oct 05 21:11:09 crc kubenswrapper[4754]: I1005 21:11:09.238527 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d20989a7-98b9-4bf2-9f85-d72890ef7e1a-db-sync-config-data\") pod \"glance-db-sync-dvzwp\" (UID: \"d20989a7-98b9-4bf2-9f85-d72890ef7e1a\") " pod="openstack/glance-db-sync-dvzwp" Oct 05 21:11:09 crc kubenswrapper[4754]: I1005 21:11:09.238606 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/c0a1bb02-14d5-4596-938d-5db342200716-additional-scripts\") pod \"ovn-controller-pp728-config-xd2gh\" (UID: \"c0a1bb02-14d5-4596-938d-5db342200716\") " pod="openstack/ovn-controller-pp728-config-xd2gh" Oct 05 21:11:09 crc kubenswrapper[4754]: I1005 21:11:09.238640 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5rfdn\" (UniqueName: \"kubernetes.io/projected/d20989a7-98b9-4bf2-9f85-d72890ef7e1a-kube-api-access-5rfdn\") pod \"glance-db-sync-dvzwp\" (UID: \"d20989a7-98b9-4bf2-9f85-d72890ef7e1a\") " pod="openstack/glance-db-sync-dvzwp" Oct 05 21:11:09 crc kubenswrapper[4754]: I1005 21:11:09.238666 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/c0a1bb02-14d5-4596-938d-5db342200716-var-run\") pod \"ovn-controller-pp728-config-xd2gh\" (UID: \"c0a1bb02-14d5-4596-938d-5db342200716\") " pod="openstack/ovn-controller-pp728-config-xd2gh" Oct 05 21:11:09 crc kubenswrapper[4754]: I1005 21:11:09.238739 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/c0a1bb02-14d5-4596-938d-5db342200716-var-run-ovn\") pod \"ovn-controller-pp728-config-xd2gh\" (UID: \"c0a1bb02-14d5-4596-938d-5db342200716\") " pod="openstack/ovn-controller-pp728-config-xd2gh" Oct 05 21:11:09 crc kubenswrapper[4754]: I1005 21:11:09.238762 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mqk4j\" (UniqueName: \"kubernetes.io/projected/c0a1bb02-14d5-4596-938d-5db342200716-kube-api-access-mqk4j\") pod \"ovn-controller-pp728-config-xd2gh\" (UID: \"c0a1bb02-14d5-4596-938d-5db342200716\") " pod="openstack/ovn-controller-pp728-config-xd2gh" Oct 05 21:11:09 crc kubenswrapper[4754]: I1005 21:11:09.245117 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d20989a7-98b9-4bf2-9f85-d72890ef7e1a-db-sync-config-data\") pod \"glance-db-sync-dvzwp\" (UID: \"d20989a7-98b9-4bf2-9f85-d72890ef7e1a\") " pod="openstack/glance-db-sync-dvzwp" Oct 05 21:11:09 crc kubenswrapper[4754]: I1005 21:11:09.255137 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d20989a7-98b9-4bf2-9f85-d72890ef7e1a-config-data\") pod \"glance-db-sync-dvzwp\" (UID: \"d20989a7-98b9-4bf2-9f85-d72890ef7e1a\") " pod="openstack/glance-db-sync-dvzwp" Oct 05 21:11:09 crc kubenswrapper[4754]: I1005 21:11:09.263270 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d20989a7-98b9-4bf2-9f85-d72890ef7e1a-combined-ca-bundle\") pod \"glance-db-sync-dvzwp\" (UID: \"d20989a7-98b9-4bf2-9f85-d72890ef7e1a\") " pod="openstack/glance-db-sync-dvzwp" Oct 05 21:11:09 crc kubenswrapper[4754]: I1005 21:11:09.272100 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5rfdn\" (UniqueName: \"kubernetes.io/projected/d20989a7-98b9-4bf2-9f85-d72890ef7e1a-kube-api-access-5rfdn\") pod \"glance-db-sync-dvzwp\" (UID: \"d20989a7-98b9-4bf2-9f85-d72890ef7e1a\") " pod="openstack/glance-db-sync-dvzwp" Oct 05 21:11:09 crc kubenswrapper[4754]: I1005 21:11:09.293666 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-dvzwp" Oct 05 21:11:09 crc kubenswrapper[4754]: I1005 21:11:09.340268 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/c0a1bb02-14d5-4596-938d-5db342200716-var-log-ovn\") pod \"ovn-controller-pp728-config-xd2gh\" (UID: \"c0a1bb02-14d5-4596-938d-5db342200716\") " pod="openstack/ovn-controller-pp728-config-xd2gh" Oct 05 21:11:09 crc kubenswrapper[4754]: I1005 21:11:09.340335 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c0a1bb02-14d5-4596-938d-5db342200716-scripts\") pod \"ovn-controller-pp728-config-xd2gh\" (UID: \"c0a1bb02-14d5-4596-938d-5db342200716\") " pod="openstack/ovn-controller-pp728-config-xd2gh" Oct 05 21:11:09 crc kubenswrapper[4754]: I1005 21:11:09.340386 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/c0a1bb02-14d5-4596-938d-5db342200716-additional-scripts\") pod \"ovn-controller-pp728-config-xd2gh\" (UID: \"c0a1bb02-14d5-4596-938d-5db342200716\") " pod="openstack/ovn-controller-pp728-config-xd2gh" Oct 05 21:11:09 crc kubenswrapper[4754]: I1005 21:11:09.340412 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/c0a1bb02-14d5-4596-938d-5db342200716-var-run\") pod \"ovn-controller-pp728-config-xd2gh\" (UID: \"c0a1bb02-14d5-4596-938d-5db342200716\") " pod="openstack/ovn-controller-pp728-config-xd2gh" Oct 05 21:11:09 crc kubenswrapper[4754]: I1005 21:11:09.340442 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/c0a1bb02-14d5-4596-938d-5db342200716-var-run-ovn\") pod \"ovn-controller-pp728-config-xd2gh\" (UID: \"c0a1bb02-14d5-4596-938d-5db342200716\") " pod="openstack/ovn-controller-pp728-config-xd2gh" Oct 05 21:11:09 crc kubenswrapper[4754]: I1005 21:11:09.340461 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mqk4j\" (UniqueName: \"kubernetes.io/projected/c0a1bb02-14d5-4596-938d-5db342200716-kube-api-access-mqk4j\") pod \"ovn-controller-pp728-config-xd2gh\" (UID: \"c0a1bb02-14d5-4596-938d-5db342200716\") " pod="openstack/ovn-controller-pp728-config-xd2gh" Oct 05 21:11:09 crc kubenswrapper[4754]: I1005 21:11:09.341026 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/c0a1bb02-14d5-4596-938d-5db342200716-var-run-ovn\") pod \"ovn-controller-pp728-config-xd2gh\" (UID: \"c0a1bb02-14d5-4596-938d-5db342200716\") " pod="openstack/ovn-controller-pp728-config-xd2gh" Oct 05 21:11:09 crc kubenswrapper[4754]: I1005 21:11:09.341051 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/c0a1bb02-14d5-4596-938d-5db342200716-var-log-ovn\") pod \"ovn-controller-pp728-config-xd2gh\" (UID: \"c0a1bb02-14d5-4596-938d-5db342200716\") " pod="openstack/ovn-controller-pp728-config-xd2gh" Oct 05 21:11:09 crc kubenswrapper[4754]: I1005 21:11:09.341116 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/c0a1bb02-14d5-4596-938d-5db342200716-var-run\") pod \"ovn-controller-pp728-config-xd2gh\" (UID: \"c0a1bb02-14d5-4596-938d-5db342200716\") " pod="openstack/ovn-controller-pp728-config-xd2gh" Oct 05 21:11:09 crc kubenswrapper[4754]: I1005 21:11:09.341473 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/c0a1bb02-14d5-4596-938d-5db342200716-additional-scripts\") pod \"ovn-controller-pp728-config-xd2gh\" (UID: \"c0a1bb02-14d5-4596-938d-5db342200716\") " pod="openstack/ovn-controller-pp728-config-xd2gh" Oct 05 21:11:09 crc kubenswrapper[4754]: I1005 21:11:09.342878 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c0a1bb02-14d5-4596-938d-5db342200716-scripts\") pod \"ovn-controller-pp728-config-xd2gh\" (UID: \"c0a1bb02-14d5-4596-938d-5db342200716\") " pod="openstack/ovn-controller-pp728-config-xd2gh" Oct 05 21:11:09 crc kubenswrapper[4754]: I1005 21:11:09.379271 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mqk4j\" (UniqueName: \"kubernetes.io/projected/c0a1bb02-14d5-4596-938d-5db342200716-kube-api-access-mqk4j\") pod \"ovn-controller-pp728-config-xd2gh\" (UID: \"c0a1bb02-14d5-4596-938d-5db342200716\") " pod="openstack/ovn-controller-pp728-config-xd2gh" Oct 05 21:11:09 crc kubenswrapper[4754]: I1005 21:11:09.417577 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-pp728-config-xd2gh" Oct 05 21:11:09 crc kubenswrapper[4754]: I1005 21:11:09.827395 4754 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="39aff8a0-2905-4ad3-ad2d-5bf6b5de3858" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.100:5671: connect: connection refused" Oct 05 21:11:10 crc kubenswrapper[4754]: I1005 21:11:10.157413 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-pp728-config-xd2gh"] Oct 05 21:11:10 crc kubenswrapper[4754]: W1005 21:11:10.167747 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc0a1bb02_14d5_4596_938d_5db342200716.slice/crio-a685c95f3339a667d4c160bb0810115ea003553d287a038d72b26a22ccf7453c WatchSource:0}: Error finding container a685c95f3339a667d4c160bb0810115ea003553d287a038d72b26a22ccf7453c: Status 404 returned error can't find the container with id a685c95f3339a667d4c160bb0810115ea003553d287a038d72b26a22ccf7453c Oct 05 21:11:10 crc kubenswrapper[4754]: I1005 21:11:10.337518 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-dvzwp"] Oct 05 21:11:10 crc kubenswrapper[4754]: I1005 21:11:10.374716 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Oct 05 21:11:10 crc kubenswrapper[4754]: I1005 21:11:10.666450 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"9af24378-0013-4510-aee6-d2fe2172935c","Type":"ContainerStarted","Data":"c278ef170b53cfda9ad8cf2ee912762eedbbb199c70ee6a655c913f90d59230c"} Oct 05 21:11:10 crc kubenswrapper[4754]: I1005 21:11:10.666826 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"9af24378-0013-4510-aee6-d2fe2172935c","Type":"ContainerStarted","Data":"7a4ae3c98624ed21f8a372244c682e866889fd3d580ef8380a74969c8672cb02"} Oct 05 21:11:10 crc kubenswrapper[4754]: I1005 21:11:10.666837 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"9af24378-0013-4510-aee6-d2fe2172935c","Type":"ContainerStarted","Data":"4b7246439d066722347aaeb3d73246b57e0e35055305016e5f7c728a18d2f584"} Oct 05 21:11:10 crc kubenswrapper[4754]: I1005 21:11:10.666845 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"9af24378-0013-4510-aee6-d2fe2172935c","Type":"ContainerStarted","Data":"7f5ffa64d333abd3b47f153e9fa2d1d8df2fa9e4e94f1181842b90645f7614de"} Oct 05 21:11:10 crc kubenswrapper[4754]: I1005 21:11:10.678148 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-dvzwp" event={"ID":"d20989a7-98b9-4bf2-9f85-d72890ef7e1a","Type":"ContainerStarted","Data":"48aaa2d81fd2975c481ebe8744b0a6f2a38775ce5c0870569e3d451376dbae75"} Oct 05 21:11:10 crc kubenswrapper[4754]: I1005 21:11:10.680016 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-pp728-config-xd2gh" event={"ID":"c0a1bb02-14d5-4596-938d-5db342200716","Type":"ContainerStarted","Data":"a685c95f3339a667d4c160bb0810115ea003553d287a038d72b26a22ccf7453c"} Oct 05 21:11:11 crc kubenswrapper[4754]: I1005 21:11:11.701664 4754 generic.go:334] "Generic (PLEG): container finished" podID="c0a1bb02-14d5-4596-938d-5db342200716" containerID="4eb4ebe30f250163db5a231ad434d1d6d333d163269168a970eb63490f9dac7e" exitCode=0 Oct 05 21:11:11 crc kubenswrapper[4754]: I1005 21:11:11.701910 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-pp728-config-xd2gh" event={"ID":"c0a1bb02-14d5-4596-938d-5db342200716","Type":"ContainerDied","Data":"4eb4ebe30f250163db5a231ad434d1d6d333d163269168a970eb63490f9dac7e"} Oct 05 21:11:11 crc kubenswrapper[4754]: I1005 21:11:11.737666 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"9af24378-0013-4510-aee6-d2fe2172935c","Type":"ContainerStarted","Data":"ca880ae44f7ea266eaa7c105134f88d3c6672ac3d27cf6d6a114dc8379b4e410"} Oct 05 21:11:11 crc kubenswrapper[4754]: I1005 21:11:11.737707 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"9af24378-0013-4510-aee6-d2fe2172935c","Type":"ContainerStarted","Data":"948d4f98e15037de18261286dcb47fc4b8a72ab31598eeeea24f6c7f86b376e3"} Oct 05 21:11:11 crc kubenswrapper[4754]: I1005 21:11:11.737717 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"9af24378-0013-4510-aee6-d2fe2172935c","Type":"ContainerStarted","Data":"965d9cdd1317ee196db17c7f79831956b39b0e232236d244b39acbf6d40b97bb"} Oct 05 21:11:11 crc kubenswrapper[4754]: I1005 21:11:11.805278 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=20.396826124 podStartE2EDuration="26.805259162s" podCreationTimestamp="2025-10-05 21:10:45 +0000 UTC" firstStartedPulling="2025-10-05 21:11:03.201680376 +0000 UTC m=+987.105799086" lastFinishedPulling="2025-10-05 21:11:09.610113414 +0000 UTC m=+993.514232124" observedRunningTime="2025-10-05 21:11:11.799414906 +0000 UTC m=+995.703533606" watchObservedRunningTime="2025-10-05 21:11:11.805259162 +0000 UTC m=+995.709377872" Oct 05 21:11:12 crc kubenswrapper[4754]: I1005 21:11:12.107743 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-7lxjx"] Oct 05 21:11:12 crc kubenswrapper[4754]: I1005 21:11:12.109401 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77585f5f8c-7lxjx" Oct 05 21:11:12 crc kubenswrapper[4754]: I1005 21:11:12.128353 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-7lxjx"] Oct 05 21:11:12 crc kubenswrapper[4754]: I1005 21:11:12.129718 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Oct 05 21:11:12 crc kubenswrapper[4754]: I1005 21:11:12.198663 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7f0ec46e-d8ac-4749-b7d5-736dba5964f9-config\") pod \"dnsmasq-dns-77585f5f8c-7lxjx\" (UID: \"7f0ec46e-d8ac-4749-b7d5-736dba5964f9\") " pod="openstack/dnsmasq-dns-77585f5f8c-7lxjx" Oct 05 21:11:12 crc kubenswrapper[4754]: I1005 21:11:12.198733 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wd6zp\" (UniqueName: \"kubernetes.io/projected/7f0ec46e-d8ac-4749-b7d5-736dba5964f9-kube-api-access-wd6zp\") pod \"dnsmasq-dns-77585f5f8c-7lxjx\" (UID: \"7f0ec46e-d8ac-4749-b7d5-736dba5964f9\") " pod="openstack/dnsmasq-dns-77585f5f8c-7lxjx" Oct 05 21:11:12 crc kubenswrapper[4754]: I1005 21:11:12.198763 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7f0ec46e-d8ac-4749-b7d5-736dba5964f9-ovsdbserver-nb\") pod \"dnsmasq-dns-77585f5f8c-7lxjx\" (UID: \"7f0ec46e-d8ac-4749-b7d5-736dba5964f9\") " pod="openstack/dnsmasq-dns-77585f5f8c-7lxjx" Oct 05 21:11:12 crc kubenswrapper[4754]: I1005 21:11:12.198800 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7f0ec46e-d8ac-4749-b7d5-736dba5964f9-dns-svc\") pod \"dnsmasq-dns-77585f5f8c-7lxjx\" (UID: \"7f0ec46e-d8ac-4749-b7d5-736dba5964f9\") " pod="openstack/dnsmasq-dns-77585f5f8c-7lxjx" Oct 05 21:11:12 crc kubenswrapper[4754]: I1005 21:11:12.198850 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7f0ec46e-d8ac-4749-b7d5-736dba5964f9-ovsdbserver-sb\") pod \"dnsmasq-dns-77585f5f8c-7lxjx\" (UID: \"7f0ec46e-d8ac-4749-b7d5-736dba5964f9\") " pod="openstack/dnsmasq-dns-77585f5f8c-7lxjx" Oct 05 21:11:12 crc kubenswrapper[4754]: I1005 21:11:12.198883 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7f0ec46e-d8ac-4749-b7d5-736dba5964f9-dns-swift-storage-0\") pod \"dnsmasq-dns-77585f5f8c-7lxjx\" (UID: \"7f0ec46e-d8ac-4749-b7d5-736dba5964f9\") " pod="openstack/dnsmasq-dns-77585f5f8c-7lxjx" Oct 05 21:11:12 crc kubenswrapper[4754]: I1005 21:11:12.300099 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7f0ec46e-d8ac-4749-b7d5-736dba5964f9-dns-svc\") pod \"dnsmasq-dns-77585f5f8c-7lxjx\" (UID: \"7f0ec46e-d8ac-4749-b7d5-736dba5964f9\") " pod="openstack/dnsmasq-dns-77585f5f8c-7lxjx" Oct 05 21:11:12 crc kubenswrapper[4754]: I1005 21:11:12.300199 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7f0ec46e-d8ac-4749-b7d5-736dba5964f9-ovsdbserver-sb\") pod \"dnsmasq-dns-77585f5f8c-7lxjx\" (UID: \"7f0ec46e-d8ac-4749-b7d5-736dba5964f9\") " pod="openstack/dnsmasq-dns-77585f5f8c-7lxjx" Oct 05 21:11:12 crc kubenswrapper[4754]: I1005 21:11:12.300237 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7f0ec46e-d8ac-4749-b7d5-736dba5964f9-dns-swift-storage-0\") pod \"dnsmasq-dns-77585f5f8c-7lxjx\" (UID: \"7f0ec46e-d8ac-4749-b7d5-736dba5964f9\") " pod="openstack/dnsmasq-dns-77585f5f8c-7lxjx" Oct 05 21:11:12 crc kubenswrapper[4754]: I1005 21:11:12.300286 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7f0ec46e-d8ac-4749-b7d5-736dba5964f9-config\") pod \"dnsmasq-dns-77585f5f8c-7lxjx\" (UID: \"7f0ec46e-d8ac-4749-b7d5-736dba5964f9\") " pod="openstack/dnsmasq-dns-77585f5f8c-7lxjx" Oct 05 21:11:12 crc kubenswrapper[4754]: I1005 21:11:12.300317 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wd6zp\" (UniqueName: \"kubernetes.io/projected/7f0ec46e-d8ac-4749-b7d5-736dba5964f9-kube-api-access-wd6zp\") pod \"dnsmasq-dns-77585f5f8c-7lxjx\" (UID: \"7f0ec46e-d8ac-4749-b7d5-736dba5964f9\") " pod="openstack/dnsmasq-dns-77585f5f8c-7lxjx" Oct 05 21:11:12 crc kubenswrapper[4754]: I1005 21:11:12.300345 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7f0ec46e-d8ac-4749-b7d5-736dba5964f9-ovsdbserver-nb\") pod \"dnsmasq-dns-77585f5f8c-7lxjx\" (UID: \"7f0ec46e-d8ac-4749-b7d5-736dba5964f9\") " pod="openstack/dnsmasq-dns-77585f5f8c-7lxjx" Oct 05 21:11:12 crc kubenswrapper[4754]: I1005 21:11:12.301022 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7f0ec46e-d8ac-4749-b7d5-736dba5964f9-dns-svc\") pod \"dnsmasq-dns-77585f5f8c-7lxjx\" (UID: \"7f0ec46e-d8ac-4749-b7d5-736dba5964f9\") " pod="openstack/dnsmasq-dns-77585f5f8c-7lxjx" Oct 05 21:11:12 crc kubenswrapper[4754]: I1005 21:11:12.301096 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7f0ec46e-d8ac-4749-b7d5-736dba5964f9-ovsdbserver-sb\") pod \"dnsmasq-dns-77585f5f8c-7lxjx\" (UID: \"7f0ec46e-d8ac-4749-b7d5-736dba5964f9\") " pod="openstack/dnsmasq-dns-77585f5f8c-7lxjx" Oct 05 21:11:12 crc kubenswrapper[4754]: I1005 21:11:12.301670 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7f0ec46e-d8ac-4749-b7d5-736dba5964f9-ovsdbserver-nb\") pod \"dnsmasq-dns-77585f5f8c-7lxjx\" (UID: \"7f0ec46e-d8ac-4749-b7d5-736dba5964f9\") " pod="openstack/dnsmasq-dns-77585f5f8c-7lxjx" Oct 05 21:11:12 crc kubenswrapper[4754]: I1005 21:11:12.302049 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7f0ec46e-d8ac-4749-b7d5-736dba5964f9-dns-swift-storage-0\") pod \"dnsmasq-dns-77585f5f8c-7lxjx\" (UID: \"7f0ec46e-d8ac-4749-b7d5-736dba5964f9\") " pod="openstack/dnsmasq-dns-77585f5f8c-7lxjx" Oct 05 21:11:12 crc kubenswrapper[4754]: I1005 21:11:12.302379 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7f0ec46e-d8ac-4749-b7d5-736dba5964f9-config\") pod \"dnsmasq-dns-77585f5f8c-7lxjx\" (UID: \"7f0ec46e-d8ac-4749-b7d5-736dba5964f9\") " pod="openstack/dnsmasq-dns-77585f5f8c-7lxjx" Oct 05 21:11:12 crc kubenswrapper[4754]: I1005 21:11:12.326776 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wd6zp\" (UniqueName: \"kubernetes.io/projected/7f0ec46e-d8ac-4749-b7d5-736dba5964f9-kube-api-access-wd6zp\") pod \"dnsmasq-dns-77585f5f8c-7lxjx\" (UID: \"7f0ec46e-d8ac-4749-b7d5-736dba5964f9\") " pod="openstack/dnsmasq-dns-77585f5f8c-7lxjx" Oct 05 21:11:12 crc kubenswrapper[4754]: I1005 21:11:12.428291 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77585f5f8c-7lxjx" Oct 05 21:11:12 crc kubenswrapper[4754]: I1005 21:11:12.820549 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-7lxjx"] Oct 05 21:11:13 crc kubenswrapper[4754]: I1005 21:11:13.116027 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-pp728-config-xd2gh" Oct 05 21:11:13 crc kubenswrapper[4754]: I1005 21:11:13.230348 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mqk4j\" (UniqueName: \"kubernetes.io/projected/c0a1bb02-14d5-4596-938d-5db342200716-kube-api-access-mqk4j\") pod \"c0a1bb02-14d5-4596-938d-5db342200716\" (UID: \"c0a1bb02-14d5-4596-938d-5db342200716\") " Oct 05 21:11:13 crc kubenswrapper[4754]: I1005 21:11:13.230440 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/c0a1bb02-14d5-4596-938d-5db342200716-additional-scripts\") pod \"c0a1bb02-14d5-4596-938d-5db342200716\" (UID: \"c0a1bb02-14d5-4596-938d-5db342200716\") " Oct 05 21:11:13 crc kubenswrapper[4754]: I1005 21:11:13.230539 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/c0a1bb02-14d5-4596-938d-5db342200716-var-run\") pod \"c0a1bb02-14d5-4596-938d-5db342200716\" (UID: \"c0a1bb02-14d5-4596-938d-5db342200716\") " Oct 05 21:11:13 crc kubenswrapper[4754]: I1005 21:11:13.230604 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/c0a1bb02-14d5-4596-938d-5db342200716-var-log-ovn\") pod \"c0a1bb02-14d5-4596-938d-5db342200716\" (UID: \"c0a1bb02-14d5-4596-938d-5db342200716\") " Oct 05 21:11:13 crc kubenswrapper[4754]: I1005 21:11:13.230642 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c0a1bb02-14d5-4596-938d-5db342200716-scripts\") pod \"c0a1bb02-14d5-4596-938d-5db342200716\" (UID: \"c0a1bb02-14d5-4596-938d-5db342200716\") " Oct 05 21:11:13 crc kubenswrapper[4754]: I1005 21:11:13.230660 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/c0a1bb02-14d5-4596-938d-5db342200716-var-run-ovn\") pod \"c0a1bb02-14d5-4596-938d-5db342200716\" (UID: \"c0a1bb02-14d5-4596-938d-5db342200716\") " Oct 05 21:11:13 crc kubenswrapper[4754]: I1005 21:11:13.230751 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c0a1bb02-14d5-4596-938d-5db342200716-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "c0a1bb02-14d5-4596-938d-5db342200716" (UID: "c0a1bb02-14d5-4596-938d-5db342200716"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 05 21:11:13 crc kubenswrapper[4754]: I1005 21:11:13.230850 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c0a1bb02-14d5-4596-938d-5db342200716-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "c0a1bb02-14d5-4596-938d-5db342200716" (UID: "c0a1bb02-14d5-4596-938d-5db342200716"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 05 21:11:13 crc kubenswrapper[4754]: I1005 21:11:13.230878 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c0a1bb02-14d5-4596-938d-5db342200716-var-run" (OuterVolumeSpecName: "var-run") pod "c0a1bb02-14d5-4596-938d-5db342200716" (UID: "c0a1bb02-14d5-4596-938d-5db342200716"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 05 21:11:13 crc kubenswrapper[4754]: I1005 21:11:13.231383 4754 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/c0a1bb02-14d5-4596-938d-5db342200716-var-run\") on node \"crc\" DevicePath \"\"" Oct 05 21:11:13 crc kubenswrapper[4754]: I1005 21:11:13.231397 4754 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/c0a1bb02-14d5-4596-938d-5db342200716-var-log-ovn\") on node \"crc\" DevicePath \"\"" Oct 05 21:11:13 crc kubenswrapper[4754]: I1005 21:11:13.231406 4754 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/c0a1bb02-14d5-4596-938d-5db342200716-var-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 05 21:11:13 crc kubenswrapper[4754]: I1005 21:11:13.231579 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c0a1bb02-14d5-4596-938d-5db342200716-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "c0a1bb02-14d5-4596-938d-5db342200716" (UID: "c0a1bb02-14d5-4596-938d-5db342200716"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 21:11:13 crc kubenswrapper[4754]: I1005 21:11:13.231881 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c0a1bb02-14d5-4596-938d-5db342200716-scripts" (OuterVolumeSpecName: "scripts") pod "c0a1bb02-14d5-4596-938d-5db342200716" (UID: "c0a1bb02-14d5-4596-938d-5db342200716"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 21:11:13 crc kubenswrapper[4754]: I1005 21:11:13.237847 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c0a1bb02-14d5-4596-938d-5db342200716-kube-api-access-mqk4j" (OuterVolumeSpecName: "kube-api-access-mqk4j") pod "c0a1bb02-14d5-4596-938d-5db342200716" (UID: "c0a1bb02-14d5-4596-938d-5db342200716"). InnerVolumeSpecName "kube-api-access-mqk4j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:11:13 crc kubenswrapper[4754]: I1005 21:11:13.339468 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mqk4j\" (UniqueName: \"kubernetes.io/projected/c0a1bb02-14d5-4596-938d-5db342200716-kube-api-access-mqk4j\") on node \"crc\" DevicePath \"\"" Oct 05 21:11:13 crc kubenswrapper[4754]: I1005 21:11:13.339521 4754 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/c0a1bb02-14d5-4596-938d-5db342200716-additional-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 21:11:13 crc kubenswrapper[4754]: I1005 21:11:13.339538 4754 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c0a1bb02-14d5-4596-938d-5db342200716-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 21:11:13 crc kubenswrapper[4754]: E1005 21:11:13.686465 4754 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7f0ec46e_d8ac_4749_b7d5_736dba5964f9.slice/crio-1666bb36d8d6aaba6b4ed3941d3eb3397fd3d42047e5eb00b6bec97ac6d0d12c.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7f0ec46e_d8ac_4749_b7d5_736dba5964f9.slice/crio-conmon-1666bb36d8d6aaba6b4ed3941d3eb3397fd3d42047e5eb00b6bec97ac6d0d12c.scope\": RecentStats: unable to find data in memory cache]" Oct 05 21:11:13 crc kubenswrapper[4754]: I1005 21:11:13.748856 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-pp728" Oct 05 21:11:13 crc kubenswrapper[4754]: I1005 21:11:13.773303 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-pp728-config-xd2gh" event={"ID":"c0a1bb02-14d5-4596-938d-5db342200716","Type":"ContainerDied","Data":"a685c95f3339a667d4c160bb0810115ea003553d287a038d72b26a22ccf7453c"} Oct 05 21:11:13 crc kubenswrapper[4754]: I1005 21:11:13.773368 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a685c95f3339a667d4c160bb0810115ea003553d287a038d72b26a22ccf7453c" Oct 05 21:11:13 crc kubenswrapper[4754]: I1005 21:11:13.773449 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-pp728-config-xd2gh" Oct 05 21:11:13 crc kubenswrapper[4754]: I1005 21:11:13.788172 4754 generic.go:334] "Generic (PLEG): container finished" podID="7f0ec46e-d8ac-4749-b7d5-736dba5964f9" containerID="1666bb36d8d6aaba6b4ed3941d3eb3397fd3d42047e5eb00b6bec97ac6d0d12c" exitCode=0 Oct 05 21:11:13 crc kubenswrapper[4754]: I1005 21:11:13.788631 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77585f5f8c-7lxjx" event={"ID":"7f0ec46e-d8ac-4749-b7d5-736dba5964f9","Type":"ContainerDied","Data":"1666bb36d8d6aaba6b4ed3941d3eb3397fd3d42047e5eb00b6bec97ac6d0d12c"} Oct 05 21:11:13 crc kubenswrapper[4754]: I1005 21:11:13.788661 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77585f5f8c-7lxjx" event={"ID":"7f0ec46e-d8ac-4749-b7d5-736dba5964f9","Type":"ContainerStarted","Data":"899f15b73841d1c02c0a24103eeeb8df271d5d86eb9182cde2e7f54080c6af16"} Oct 05 21:11:14 crc kubenswrapper[4754]: I1005 21:11:14.222141 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-pp728-config-xd2gh"] Oct 05 21:11:14 crc kubenswrapper[4754]: I1005 21:11:14.234014 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-pp728-config-xd2gh"] Oct 05 21:11:14 crc kubenswrapper[4754]: I1005 21:11:14.265535 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-pp728-config-l8fc7"] Oct 05 21:11:14 crc kubenswrapper[4754]: E1005 21:11:14.268245 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c0a1bb02-14d5-4596-938d-5db342200716" containerName="ovn-config" Oct 05 21:11:14 crc kubenswrapper[4754]: I1005 21:11:14.268276 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="c0a1bb02-14d5-4596-938d-5db342200716" containerName="ovn-config" Oct 05 21:11:14 crc kubenswrapper[4754]: I1005 21:11:14.268545 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="c0a1bb02-14d5-4596-938d-5db342200716" containerName="ovn-config" Oct 05 21:11:14 crc kubenswrapper[4754]: I1005 21:11:14.269262 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-pp728-config-l8fc7" Oct 05 21:11:14 crc kubenswrapper[4754]: I1005 21:11:14.272137 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Oct 05 21:11:14 crc kubenswrapper[4754]: I1005 21:11:14.289052 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-pp728-config-l8fc7"] Oct 05 21:11:14 crc kubenswrapper[4754]: I1005 21:11:14.392554 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/d39d8e73-69f1-40b3-8025-3c65bb2a50ed-var-run-ovn\") pod \"ovn-controller-pp728-config-l8fc7\" (UID: \"d39d8e73-69f1-40b3-8025-3c65bb2a50ed\") " pod="openstack/ovn-controller-pp728-config-l8fc7" Oct 05 21:11:14 crc kubenswrapper[4754]: I1005 21:11:14.392646 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d39d8e73-69f1-40b3-8025-3c65bb2a50ed-scripts\") pod \"ovn-controller-pp728-config-l8fc7\" (UID: \"d39d8e73-69f1-40b3-8025-3c65bb2a50ed\") " pod="openstack/ovn-controller-pp728-config-l8fc7" Oct 05 21:11:14 crc kubenswrapper[4754]: I1005 21:11:14.392703 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zr2ww\" (UniqueName: \"kubernetes.io/projected/d39d8e73-69f1-40b3-8025-3c65bb2a50ed-kube-api-access-zr2ww\") pod \"ovn-controller-pp728-config-l8fc7\" (UID: \"d39d8e73-69f1-40b3-8025-3c65bb2a50ed\") " pod="openstack/ovn-controller-pp728-config-l8fc7" Oct 05 21:11:14 crc kubenswrapper[4754]: I1005 21:11:14.392723 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/d39d8e73-69f1-40b3-8025-3c65bb2a50ed-additional-scripts\") pod \"ovn-controller-pp728-config-l8fc7\" (UID: \"d39d8e73-69f1-40b3-8025-3c65bb2a50ed\") " pod="openstack/ovn-controller-pp728-config-l8fc7" Oct 05 21:11:14 crc kubenswrapper[4754]: I1005 21:11:14.392768 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/d39d8e73-69f1-40b3-8025-3c65bb2a50ed-var-run\") pod \"ovn-controller-pp728-config-l8fc7\" (UID: \"d39d8e73-69f1-40b3-8025-3c65bb2a50ed\") " pod="openstack/ovn-controller-pp728-config-l8fc7" Oct 05 21:11:14 crc kubenswrapper[4754]: I1005 21:11:14.392799 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/d39d8e73-69f1-40b3-8025-3c65bb2a50ed-var-log-ovn\") pod \"ovn-controller-pp728-config-l8fc7\" (UID: \"d39d8e73-69f1-40b3-8025-3c65bb2a50ed\") " pod="openstack/ovn-controller-pp728-config-l8fc7" Oct 05 21:11:14 crc kubenswrapper[4754]: I1005 21:11:14.494842 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zr2ww\" (UniqueName: \"kubernetes.io/projected/d39d8e73-69f1-40b3-8025-3c65bb2a50ed-kube-api-access-zr2ww\") pod \"ovn-controller-pp728-config-l8fc7\" (UID: \"d39d8e73-69f1-40b3-8025-3c65bb2a50ed\") " pod="openstack/ovn-controller-pp728-config-l8fc7" Oct 05 21:11:14 crc kubenswrapper[4754]: I1005 21:11:14.494894 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/d39d8e73-69f1-40b3-8025-3c65bb2a50ed-additional-scripts\") pod \"ovn-controller-pp728-config-l8fc7\" (UID: \"d39d8e73-69f1-40b3-8025-3c65bb2a50ed\") " pod="openstack/ovn-controller-pp728-config-l8fc7" Oct 05 21:11:14 crc kubenswrapper[4754]: I1005 21:11:14.494966 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/d39d8e73-69f1-40b3-8025-3c65bb2a50ed-var-run\") pod \"ovn-controller-pp728-config-l8fc7\" (UID: \"d39d8e73-69f1-40b3-8025-3c65bb2a50ed\") " pod="openstack/ovn-controller-pp728-config-l8fc7" Oct 05 21:11:14 crc kubenswrapper[4754]: I1005 21:11:14.495005 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/d39d8e73-69f1-40b3-8025-3c65bb2a50ed-var-log-ovn\") pod \"ovn-controller-pp728-config-l8fc7\" (UID: \"d39d8e73-69f1-40b3-8025-3c65bb2a50ed\") " pod="openstack/ovn-controller-pp728-config-l8fc7" Oct 05 21:11:14 crc kubenswrapper[4754]: I1005 21:11:14.495044 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/d39d8e73-69f1-40b3-8025-3c65bb2a50ed-var-run-ovn\") pod \"ovn-controller-pp728-config-l8fc7\" (UID: \"d39d8e73-69f1-40b3-8025-3c65bb2a50ed\") " pod="openstack/ovn-controller-pp728-config-l8fc7" Oct 05 21:11:14 crc kubenswrapper[4754]: I1005 21:11:14.495105 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d39d8e73-69f1-40b3-8025-3c65bb2a50ed-scripts\") pod \"ovn-controller-pp728-config-l8fc7\" (UID: \"d39d8e73-69f1-40b3-8025-3c65bb2a50ed\") " pod="openstack/ovn-controller-pp728-config-l8fc7" Oct 05 21:11:14 crc kubenswrapper[4754]: I1005 21:11:14.495838 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/d39d8e73-69f1-40b3-8025-3c65bb2a50ed-additional-scripts\") pod \"ovn-controller-pp728-config-l8fc7\" (UID: \"d39d8e73-69f1-40b3-8025-3c65bb2a50ed\") " pod="openstack/ovn-controller-pp728-config-l8fc7" Oct 05 21:11:14 crc kubenswrapper[4754]: I1005 21:11:14.495940 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/d39d8e73-69f1-40b3-8025-3c65bb2a50ed-var-run\") pod \"ovn-controller-pp728-config-l8fc7\" (UID: \"d39d8e73-69f1-40b3-8025-3c65bb2a50ed\") " pod="openstack/ovn-controller-pp728-config-l8fc7" Oct 05 21:11:14 crc kubenswrapper[4754]: I1005 21:11:14.495955 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/d39d8e73-69f1-40b3-8025-3c65bb2a50ed-var-log-ovn\") pod \"ovn-controller-pp728-config-l8fc7\" (UID: \"d39d8e73-69f1-40b3-8025-3c65bb2a50ed\") " pod="openstack/ovn-controller-pp728-config-l8fc7" Oct 05 21:11:14 crc kubenswrapper[4754]: I1005 21:11:14.495988 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/d39d8e73-69f1-40b3-8025-3c65bb2a50ed-var-run-ovn\") pod \"ovn-controller-pp728-config-l8fc7\" (UID: \"d39d8e73-69f1-40b3-8025-3c65bb2a50ed\") " pod="openstack/ovn-controller-pp728-config-l8fc7" Oct 05 21:11:14 crc kubenswrapper[4754]: I1005 21:11:14.508762 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d39d8e73-69f1-40b3-8025-3c65bb2a50ed-scripts\") pod \"ovn-controller-pp728-config-l8fc7\" (UID: \"d39d8e73-69f1-40b3-8025-3c65bb2a50ed\") " pod="openstack/ovn-controller-pp728-config-l8fc7" Oct 05 21:11:14 crc kubenswrapper[4754]: I1005 21:11:14.515133 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zr2ww\" (UniqueName: \"kubernetes.io/projected/d39d8e73-69f1-40b3-8025-3c65bb2a50ed-kube-api-access-zr2ww\") pod \"ovn-controller-pp728-config-l8fc7\" (UID: \"d39d8e73-69f1-40b3-8025-3c65bb2a50ed\") " pod="openstack/ovn-controller-pp728-config-l8fc7" Oct 05 21:11:14 crc kubenswrapper[4754]: I1005 21:11:14.601050 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-pp728-config-l8fc7" Oct 05 21:11:14 crc kubenswrapper[4754]: I1005 21:11:14.804423 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77585f5f8c-7lxjx" event={"ID":"7f0ec46e-d8ac-4749-b7d5-736dba5964f9","Type":"ContainerStarted","Data":"3b296eec4386637feee1cf621c596d6555e98cfd40bfd7e08031ea4ee1ac7e03"} Oct 05 21:11:14 crc kubenswrapper[4754]: I1005 21:11:14.804802 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-77585f5f8c-7lxjx" Oct 05 21:11:14 crc kubenswrapper[4754]: I1005 21:11:14.823983 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-77585f5f8c-7lxjx" podStartSLOduration=2.823963556 podStartE2EDuration="2.823963556s" podCreationTimestamp="2025-10-05 21:11:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 21:11:14.823816312 +0000 UTC m=+998.727935022" watchObservedRunningTime="2025-10-05 21:11:14.823963556 +0000 UTC m=+998.728082256" Oct 05 21:11:14 crc kubenswrapper[4754]: I1005 21:11:14.851992 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c0a1bb02-14d5-4596-938d-5db342200716" path="/var/lib/kubelet/pods/c0a1bb02-14d5-4596-938d-5db342200716/volumes" Oct 05 21:11:15 crc kubenswrapper[4754]: I1005 21:11:15.113920 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-pp728-config-l8fc7"] Oct 05 21:11:15 crc kubenswrapper[4754]: I1005 21:11:15.815181 4754 generic.go:334] "Generic (PLEG): container finished" podID="d39d8e73-69f1-40b3-8025-3c65bb2a50ed" containerID="e0274a90d2ea4d614f3f772870b226b425ff9e29d203263f43d6f358f0642c31" exitCode=0 Oct 05 21:11:15 crc kubenswrapper[4754]: I1005 21:11:15.815271 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-pp728-config-l8fc7" event={"ID":"d39d8e73-69f1-40b3-8025-3c65bb2a50ed","Type":"ContainerDied","Data":"e0274a90d2ea4d614f3f772870b226b425ff9e29d203263f43d6f358f0642c31"} Oct 05 21:11:15 crc kubenswrapper[4754]: I1005 21:11:15.815326 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-pp728-config-l8fc7" event={"ID":"d39d8e73-69f1-40b3-8025-3c65bb2a50ed","Type":"ContainerStarted","Data":"761832c5a85634cc263c458a9ebc2f374383e0ccf944934ef4bb936cb21af6ee"} Oct 05 21:11:17 crc kubenswrapper[4754]: I1005 21:11:17.168438 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-pp728-config-l8fc7" Oct 05 21:11:17 crc kubenswrapper[4754]: I1005 21:11:17.345313 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/d39d8e73-69f1-40b3-8025-3c65bb2a50ed-var-run\") pod \"d39d8e73-69f1-40b3-8025-3c65bb2a50ed\" (UID: \"d39d8e73-69f1-40b3-8025-3c65bb2a50ed\") " Oct 05 21:11:17 crc kubenswrapper[4754]: I1005 21:11:17.345388 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d39d8e73-69f1-40b3-8025-3c65bb2a50ed-var-run" (OuterVolumeSpecName: "var-run") pod "d39d8e73-69f1-40b3-8025-3c65bb2a50ed" (UID: "d39d8e73-69f1-40b3-8025-3c65bb2a50ed"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 05 21:11:17 crc kubenswrapper[4754]: I1005 21:11:17.345403 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d39d8e73-69f1-40b3-8025-3c65bb2a50ed-scripts\") pod \"d39d8e73-69f1-40b3-8025-3c65bb2a50ed\" (UID: \"d39d8e73-69f1-40b3-8025-3c65bb2a50ed\") " Oct 05 21:11:17 crc kubenswrapper[4754]: I1005 21:11:17.345432 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/d39d8e73-69f1-40b3-8025-3c65bb2a50ed-var-run-ovn\") pod \"d39d8e73-69f1-40b3-8025-3c65bb2a50ed\" (UID: \"d39d8e73-69f1-40b3-8025-3c65bb2a50ed\") " Oct 05 21:11:17 crc kubenswrapper[4754]: I1005 21:11:17.345518 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/d39d8e73-69f1-40b3-8025-3c65bb2a50ed-additional-scripts\") pod \"d39d8e73-69f1-40b3-8025-3c65bb2a50ed\" (UID: \"d39d8e73-69f1-40b3-8025-3c65bb2a50ed\") " Oct 05 21:11:17 crc kubenswrapper[4754]: I1005 21:11:17.345551 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zr2ww\" (UniqueName: \"kubernetes.io/projected/d39d8e73-69f1-40b3-8025-3c65bb2a50ed-kube-api-access-zr2ww\") pod \"d39d8e73-69f1-40b3-8025-3c65bb2a50ed\" (UID: \"d39d8e73-69f1-40b3-8025-3c65bb2a50ed\") " Oct 05 21:11:17 crc kubenswrapper[4754]: I1005 21:11:17.345576 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d39d8e73-69f1-40b3-8025-3c65bb2a50ed-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "d39d8e73-69f1-40b3-8025-3c65bb2a50ed" (UID: "d39d8e73-69f1-40b3-8025-3c65bb2a50ed"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 05 21:11:17 crc kubenswrapper[4754]: I1005 21:11:17.345618 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/d39d8e73-69f1-40b3-8025-3c65bb2a50ed-var-log-ovn\") pod \"d39d8e73-69f1-40b3-8025-3c65bb2a50ed\" (UID: \"d39d8e73-69f1-40b3-8025-3c65bb2a50ed\") " Oct 05 21:11:17 crc kubenswrapper[4754]: I1005 21:11:17.346056 4754 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/d39d8e73-69f1-40b3-8025-3c65bb2a50ed-var-run\") on node \"crc\" DevicePath \"\"" Oct 05 21:11:17 crc kubenswrapper[4754]: I1005 21:11:17.346075 4754 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/d39d8e73-69f1-40b3-8025-3c65bb2a50ed-var-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 05 21:11:17 crc kubenswrapper[4754]: I1005 21:11:17.346096 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d39d8e73-69f1-40b3-8025-3c65bb2a50ed-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "d39d8e73-69f1-40b3-8025-3c65bb2a50ed" (UID: "d39d8e73-69f1-40b3-8025-3c65bb2a50ed"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 05 21:11:17 crc kubenswrapper[4754]: I1005 21:11:17.346130 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d39d8e73-69f1-40b3-8025-3c65bb2a50ed-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "d39d8e73-69f1-40b3-8025-3c65bb2a50ed" (UID: "d39d8e73-69f1-40b3-8025-3c65bb2a50ed"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 21:11:17 crc kubenswrapper[4754]: I1005 21:11:17.346816 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d39d8e73-69f1-40b3-8025-3c65bb2a50ed-scripts" (OuterVolumeSpecName: "scripts") pod "d39d8e73-69f1-40b3-8025-3c65bb2a50ed" (UID: "d39d8e73-69f1-40b3-8025-3c65bb2a50ed"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 21:11:17 crc kubenswrapper[4754]: I1005 21:11:17.352936 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d39d8e73-69f1-40b3-8025-3c65bb2a50ed-kube-api-access-zr2ww" (OuterVolumeSpecName: "kube-api-access-zr2ww") pod "d39d8e73-69f1-40b3-8025-3c65bb2a50ed" (UID: "d39d8e73-69f1-40b3-8025-3c65bb2a50ed"). InnerVolumeSpecName "kube-api-access-zr2ww". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:11:17 crc kubenswrapper[4754]: I1005 21:11:17.447438 4754 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d39d8e73-69f1-40b3-8025-3c65bb2a50ed-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 21:11:17 crc kubenswrapper[4754]: I1005 21:11:17.447472 4754 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/d39d8e73-69f1-40b3-8025-3c65bb2a50ed-additional-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 21:11:17 crc kubenswrapper[4754]: I1005 21:11:17.447484 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zr2ww\" (UniqueName: \"kubernetes.io/projected/d39d8e73-69f1-40b3-8025-3c65bb2a50ed-kube-api-access-zr2ww\") on node \"crc\" DevicePath \"\"" Oct 05 21:11:17 crc kubenswrapper[4754]: I1005 21:11:17.447542 4754 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/d39d8e73-69f1-40b3-8025-3c65bb2a50ed-var-log-ovn\") on node \"crc\" DevicePath \"\"" Oct 05 21:11:17 crc kubenswrapper[4754]: I1005 21:11:17.832933 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-pp728-config-l8fc7" event={"ID":"d39d8e73-69f1-40b3-8025-3c65bb2a50ed","Type":"ContainerDied","Data":"761832c5a85634cc263c458a9ebc2f374383e0ccf944934ef4bb936cb21af6ee"} Oct 05 21:11:17 crc kubenswrapper[4754]: I1005 21:11:17.832969 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="761832c5a85634cc263c458a9ebc2f374383e0ccf944934ef4bb936cb21af6ee" Oct 05 21:11:17 crc kubenswrapper[4754]: I1005 21:11:17.832967 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-pp728-config-l8fc7" Oct 05 21:11:18 crc kubenswrapper[4754]: I1005 21:11:18.251345 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-pp728-config-l8fc7"] Oct 05 21:11:18 crc kubenswrapper[4754]: I1005 21:11:18.258547 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-pp728-config-l8fc7"] Oct 05 21:11:18 crc kubenswrapper[4754]: I1005 21:11:18.846944 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d39d8e73-69f1-40b3-8025-3c65bb2a50ed" path="/var/lib/kubelet/pods/d39d8e73-69f1-40b3-8025-3c65bb2a50ed/volumes" Oct 05 21:11:19 crc kubenswrapper[4754]: I1005 21:11:19.826573 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Oct 05 21:11:20 crc kubenswrapper[4754]: I1005 21:11:20.194052 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-qpgn5"] Oct 05 21:11:20 crc kubenswrapper[4754]: E1005 21:11:20.194437 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d39d8e73-69f1-40b3-8025-3c65bb2a50ed" containerName="ovn-config" Oct 05 21:11:20 crc kubenswrapper[4754]: I1005 21:11:20.194457 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="d39d8e73-69f1-40b3-8025-3c65bb2a50ed" containerName="ovn-config" Oct 05 21:11:20 crc kubenswrapper[4754]: I1005 21:11:20.194697 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="d39d8e73-69f1-40b3-8025-3c65bb2a50ed" containerName="ovn-config" Oct 05 21:11:20 crc kubenswrapper[4754]: I1005 21:11:20.195301 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-qpgn5" Oct 05 21:11:20 crc kubenswrapper[4754]: I1005 21:11:20.224085 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-qpgn5"] Oct 05 21:11:20 crc kubenswrapper[4754]: I1005 21:11:20.300347 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q9wpv\" (UniqueName: \"kubernetes.io/projected/c0678fc1-b079-4379-8bbc-8848c3848f02-kube-api-access-q9wpv\") pod \"cinder-db-create-qpgn5\" (UID: \"c0678fc1-b079-4379-8bbc-8848c3848f02\") " pod="openstack/cinder-db-create-qpgn5" Oct 05 21:11:20 crc kubenswrapper[4754]: I1005 21:11:20.382755 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-5mdmj"] Oct 05 21:11:20 crc kubenswrapper[4754]: I1005 21:11:20.383780 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-5mdmj" Oct 05 21:11:20 crc kubenswrapper[4754]: I1005 21:11:20.399557 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-5mdmj"] Oct 05 21:11:20 crc kubenswrapper[4754]: I1005 21:11:20.401345 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q9wpv\" (UniqueName: \"kubernetes.io/projected/c0678fc1-b079-4379-8bbc-8848c3848f02-kube-api-access-q9wpv\") pod \"cinder-db-create-qpgn5\" (UID: \"c0678fc1-b079-4379-8bbc-8848c3848f02\") " pod="openstack/cinder-db-create-qpgn5" Oct 05 21:11:20 crc kubenswrapper[4754]: I1005 21:11:20.441452 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q9wpv\" (UniqueName: \"kubernetes.io/projected/c0678fc1-b079-4379-8bbc-8848c3848f02-kube-api-access-q9wpv\") pod \"cinder-db-create-qpgn5\" (UID: \"c0678fc1-b079-4379-8bbc-8848c3848f02\") " pod="openstack/cinder-db-create-qpgn5" Oct 05 21:11:20 crc kubenswrapper[4754]: I1005 21:11:20.504025 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vlnbv\" (UniqueName: \"kubernetes.io/projected/0f34b855-e478-41a0-b7b4-28dcf667b7c2-kube-api-access-vlnbv\") pod \"barbican-db-create-5mdmj\" (UID: \"0f34b855-e478-41a0-b7b4-28dcf667b7c2\") " pod="openstack/barbican-db-create-5mdmj" Oct 05 21:11:20 crc kubenswrapper[4754]: I1005 21:11:20.518749 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-qpgn5" Oct 05 21:11:20 crc kubenswrapper[4754]: I1005 21:11:20.588527 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-679wz"] Oct 05 21:11:20 crc kubenswrapper[4754]: I1005 21:11:20.589483 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-679wz" Oct 05 21:11:20 crc kubenswrapper[4754]: I1005 21:11:20.605407 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-679wz"] Oct 05 21:11:20 crc kubenswrapper[4754]: I1005 21:11:20.605805 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vlnbv\" (UniqueName: \"kubernetes.io/projected/0f34b855-e478-41a0-b7b4-28dcf667b7c2-kube-api-access-vlnbv\") pod \"barbican-db-create-5mdmj\" (UID: \"0f34b855-e478-41a0-b7b4-28dcf667b7c2\") " pod="openstack/barbican-db-create-5mdmj" Oct 05 21:11:20 crc kubenswrapper[4754]: I1005 21:11:20.634930 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vlnbv\" (UniqueName: \"kubernetes.io/projected/0f34b855-e478-41a0-b7b4-28dcf667b7c2-kube-api-access-vlnbv\") pod \"barbican-db-create-5mdmj\" (UID: \"0f34b855-e478-41a0-b7b4-28dcf667b7c2\") " pod="openstack/barbican-db-create-5mdmj" Oct 05 21:11:20 crc kubenswrapper[4754]: I1005 21:11:20.701158 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-5mdmj" Oct 05 21:11:20 crc kubenswrapper[4754]: I1005 21:11:20.707129 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q6gfh\" (UniqueName: \"kubernetes.io/projected/9a72c674-5986-45a3-aae6-7f15de18efed-kube-api-access-q6gfh\") pod \"neutron-db-create-679wz\" (UID: \"9a72c674-5986-45a3-aae6-7f15de18efed\") " pod="openstack/neutron-db-create-679wz" Oct 05 21:11:20 crc kubenswrapper[4754]: I1005 21:11:20.808767 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q6gfh\" (UniqueName: \"kubernetes.io/projected/9a72c674-5986-45a3-aae6-7f15de18efed-kube-api-access-q6gfh\") pod \"neutron-db-create-679wz\" (UID: \"9a72c674-5986-45a3-aae6-7f15de18efed\") " pod="openstack/neutron-db-create-679wz" Oct 05 21:11:20 crc kubenswrapper[4754]: I1005 21:11:20.822755 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q6gfh\" (UniqueName: \"kubernetes.io/projected/9a72c674-5986-45a3-aae6-7f15de18efed-kube-api-access-q6gfh\") pod \"neutron-db-create-679wz\" (UID: \"9a72c674-5986-45a3-aae6-7f15de18efed\") " pod="openstack/neutron-db-create-679wz" Oct 05 21:11:20 crc kubenswrapper[4754]: I1005 21:11:20.917165 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-679wz" Oct 05 21:11:21 crc kubenswrapper[4754]: I1005 21:11:21.137257 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-9fzft"] Oct 05 21:11:21 crc kubenswrapper[4754]: I1005 21:11:21.138600 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-9fzft" Oct 05 21:11:21 crc kubenswrapper[4754]: I1005 21:11:21.140287 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 05 21:11:21 crc kubenswrapper[4754]: I1005 21:11:21.141926 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 05 21:11:21 crc kubenswrapper[4754]: I1005 21:11:21.142094 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 05 21:11:21 crc kubenswrapper[4754]: I1005 21:11:21.142268 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-5846s" Oct 05 21:11:21 crc kubenswrapper[4754]: I1005 21:11:21.156216 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-9fzft"] Oct 05 21:11:21 crc kubenswrapper[4754]: I1005 21:11:21.216870 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87ecff30-d447-417b-9fe5-97cd6fd64c93-combined-ca-bundle\") pod \"keystone-db-sync-9fzft\" (UID: \"87ecff30-d447-417b-9fe5-97cd6fd64c93\") " pod="openstack/keystone-db-sync-9fzft" Oct 05 21:11:21 crc kubenswrapper[4754]: I1005 21:11:21.217256 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6qg4l\" (UniqueName: \"kubernetes.io/projected/87ecff30-d447-417b-9fe5-97cd6fd64c93-kube-api-access-6qg4l\") pod \"keystone-db-sync-9fzft\" (UID: \"87ecff30-d447-417b-9fe5-97cd6fd64c93\") " pod="openstack/keystone-db-sync-9fzft" Oct 05 21:11:21 crc kubenswrapper[4754]: I1005 21:11:21.217349 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87ecff30-d447-417b-9fe5-97cd6fd64c93-config-data\") pod \"keystone-db-sync-9fzft\" (UID: \"87ecff30-d447-417b-9fe5-97cd6fd64c93\") " pod="openstack/keystone-db-sync-9fzft" Oct 05 21:11:21 crc kubenswrapper[4754]: I1005 21:11:21.319253 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87ecff30-d447-417b-9fe5-97cd6fd64c93-combined-ca-bundle\") pod \"keystone-db-sync-9fzft\" (UID: \"87ecff30-d447-417b-9fe5-97cd6fd64c93\") " pod="openstack/keystone-db-sync-9fzft" Oct 05 21:11:21 crc kubenswrapper[4754]: I1005 21:11:21.319329 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6qg4l\" (UniqueName: \"kubernetes.io/projected/87ecff30-d447-417b-9fe5-97cd6fd64c93-kube-api-access-6qg4l\") pod \"keystone-db-sync-9fzft\" (UID: \"87ecff30-d447-417b-9fe5-97cd6fd64c93\") " pod="openstack/keystone-db-sync-9fzft" Oct 05 21:11:21 crc kubenswrapper[4754]: I1005 21:11:21.319467 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87ecff30-d447-417b-9fe5-97cd6fd64c93-config-data\") pod \"keystone-db-sync-9fzft\" (UID: \"87ecff30-d447-417b-9fe5-97cd6fd64c93\") " pod="openstack/keystone-db-sync-9fzft" Oct 05 21:11:21 crc kubenswrapper[4754]: I1005 21:11:21.322763 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87ecff30-d447-417b-9fe5-97cd6fd64c93-combined-ca-bundle\") pod \"keystone-db-sync-9fzft\" (UID: \"87ecff30-d447-417b-9fe5-97cd6fd64c93\") " pod="openstack/keystone-db-sync-9fzft" Oct 05 21:11:21 crc kubenswrapper[4754]: I1005 21:11:21.323268 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87ecff30-d447-417b-9fe5-97cd6fd64c93-config-data\") pod \"keystone-db-sync-9fzft\" (UID: \"87ecff30-d447-417b-9fe5-97cd6fd64c93\") " pod="openstack/keystone-db-sync-9fzft" Oct 05 21:11:21 crc kubenswrapper[4754]: I1005 21:11:21.335628 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6qg4l\" (UniqueName: \"kubernetes.io/projected/87ecff30-d447-417b-9fe5-97cd6fd64c93-kube-api-access-6qg4l\") pod \"keystone-db-sync-9fzft\" (UID: \"87ecff30-d447-417b-9fe5-97cd6fd64c93\") " pod="openstack/keystone-db-sync-9fzft" Oct 05 21:11:21 crc kubenswrapper[4754]: I1005 21:11:21.458740 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-9fzft" Oct 05 21:11:22 crc kubenswrapper[4754]: I1005 21:11:22.430659 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-77585f5f8c-7lxjx" Oct 05 21:11:22 crc kubenswrapper[4754]: I1005 21:11:22.488463 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-hh4gc"] Oct 05 21:11:22 crc kubenswrapper[4754]: I1005 21:11:22.488712 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-698758b865-hh4gc" podUID="c636e2d4-bd61-47ff-ba53-e7da442521ce" containerName="dnsmasq-dns" containerID="cri-o://d487267b338cdf082ec536549e664a2ceda532b59c086ea25c0f0500412b65b9" gracePeriod=10 Oct 05 21:11:22 crc kubenswrapper[4754]: I1005 21:11:22.889050 4754 generic.go:334] "Generic (PLEG): container finished" podID="c636e2d4-bd61-47ff-ba53-e7da442521ce" containerID="d487267b338cdf082ec536549e664a2ceda532b59c086ea25c0f0500412b65b9" exitCode=0 Oct 05 21:11:22 crc kubenswrapper[4754]: I1005 21:11:22.889104 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-hh4gc" event={"ID":"c636e2d4-bd61-47ff-ba53-e7da442521ce","Type":"ContainerDied","Data":"d487267b338cdf082ec536549e664a2ceda532b59c086ea25c0f0500412b65b9"} Oct 05 21:11:25 crc kubenswrapper[4754]: I1005 21:11:25.503381 4754 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-698758b865-hh4gc" podUID="c636e2d4-bd61-47ff-ba53-e7da442521ce" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.114:5353: connect: connection refused" Oct 05 21:11:27 crc kubenswrapper[4754]: I1005 21:11:27.382457 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-hh4gc" Oct 05 21:11:27 crc kubenswrapper[4754]: I1005 21:11:27.535709 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6hvh\" (UniqueName: \"kubernetes.io/projected/c636e2d4-bd61-47ff-ba53-e7da442521ce-kube-api-access-d6hvh\") pod \"c636e2d4-bd61-47ff-ba53-e7da442521ce\" (UID: \"c636e2d4-bd61-47ff-ba53-e7da442521ce\") " Oct 05 21:11:27 crc kubenswrapper[4754]: I1005 21:11:27.535845 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c636e2d4-bd61-47ff-ba53-e7da442521ce-dns-svc\") pod \"c636e2d4-bd61-47ff-ba53-e7da442521ce\" (UID: \"c636e2d4-bd61-47ff-ba53-e7da442521ce\") " Oct 05 21:11:27 crc kubenswrapper[4754]: I1005 21:11:27.535941 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c636e2d4-bd61-47ff-ba53-e7da442521ce-ovsdbserver-nb\") pod \"c636e2d4-bd61-47ff-ba53-e7da442521ce\" (UID: \"c636e2d4-bd61-47ff-ba53-e7da442521ce\") " Oct 05 21:11:27 crc kubenswrapper[4754]: I1005 21:11:27.536007 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c636e2d4-bd61-47ff-ba53-e7da442521ce-ovsdbserver-sb\") pod \"c636e2d4-bd61-47ff-ba53-e7da442521ce\" (UID: \"c636e2d4-bd61-47ff-ba53-e7da442521ce\") " Oct 05 21:11:27 crc kubenswrapper[4754]: I1005 21:11:27.536055 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c636e2d4-bd61-47ff-ba53-e7da442521ce-config\") pod \"c636e2d4-bd61-47ff-ba53-e7da442521ce\" (UID: \"c636e2d4-bd61-47ff-ba53-e7da442521ce\") " Oct 05 21:11:27 crc kubenswrapper[4754]: I1005 21:11:27.540601 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c636e2d4-bd61-47ff-ba53-e7da442521ce-kube-api-access-d6hvh" (OuterVolumeSpecName: "kube-api-access-d6hvh") pod "c636e2d4-bd61-47ff-ba53-e7da442521ce" (UID: "c636e2d4-bd61-47ff-ba53-e7da442521ce"). InnerVolumeSpecName "kube-api-access-d6hvh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:11:27 crc kubenswrapper[4754]: I1005 21:11:27.594162 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c636e2d4-bd61-47ff-ba53-e7da442521ce-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "c636e2d4-bd61-47ff-ba53-e7da442521ce" (UID: "c636e2d4-bd61-47ff-ba53-e7da442521ce"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 21:11:27 crc kubenswrapper[4754]: I1005 21:11:27.594328 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c636e2d4-bd61-47ff-ba53-e7da442521ce-config" (OuterVolumeSpecName: "config") pod "c636e2d4-bd61-47ff-ba53-e7da442521ce" (UID: "c636e2d4-bd61-47ff-ba53-e7da442521ce"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 21:11:27 crc kubenswrapper[4754]: I1005 21:11:27.594676 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c636e2d4-bd61-47ff-ba53-e7da442521ce-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "c636e2d4-bd61-47ff-ba53-e7da442521ce" (UID: "c636e2d4-bd61-47ff-ba53-e7da442521ce"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 21:11:27 crc kubenswrapper[4754]: I1005 21:11:27.605073 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c636e2d4-bd61-47ff-ba53-e7da442521ce-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "c636e2d4-bd61-47ff-ba53-e7da442521ce" (UID: "c636e2d4-bd61-47ff-ba53-e7da442521ce"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 21:11:27 crc kubenswrapper[4754]: I1005 21:11:27.637472 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6hvh\" (UniqueName: \"kubernetes.io/projected/c636e2d4-bd61-47ff-ba53-e7da442521ce-kube-api-access-d6hvh\") on node \"crc\" DevicePath \"\"" Oct 05 21:11:27 crc kubenswrapper[4754]: I1005 21:11:27.638003 4754 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c636e2d4-bd61-47ff-ba53-e7da442521ce-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 05 21:11:27 crc kubenswrapper[4754]: I1005 21:11:27.638017 4754 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c636e2d4-bd61-47ff-ba53-e7da442521ce-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 05 21:11:27 crc kubenswrapper[4754]: I1005 21:11:27.638025 4754 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c636e2d4-bd61-47ff-ba53-e7da442521ce-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 05 21:11:27 crc kubenswrapper[4754]: I1005 21:11:27.638033 4754 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c636e2d4-bd61-47ff-ba53-e7da442521ce-config\") on node \"crc\" DevicePath \"\"" Oct 05 21:11:27 crc kubenswrapper[4754]: I1005 21:11:27.686565 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-679wz"] Oct 05 21:11:27 crc kubenswrapper[4754]: W1005 21:11:27.693605 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9a72c674_5986_45a3_aae6_7f15de18efed.slice/crio-84ab4cc4bcaec2bc885e6e3708a97369cb44dc4592d1ba2198a9eace0eaa0f82 WatchSource:0}: Error finding container 84ab4cc4bcaec2bc885e6e3708a97369cb44dc4592d1ba2198a9eace0eaa0f82: Status 404 returned error can't find the container with id 84ab4cc4bcaec2bc885e6e3708a97369cb44dc4592d1ba2198a9eace0eaa0f82 Oct 05 21:11:27 crc kubenswrapper[4754]: I1005 21:11:27.762587 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-9fzft"] Oct 05 21:11:27 crc kubenswrapper[4754]: W1005 21:11:27.772429 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod87ecff30_d447_417b_9fe5_97cd6fd64c93.slice/crio-61e10ef5c21cbb4e674c4ab0586b82bc9ca40a4baad239b081248a215f23b0c6 WatchSource:0}: Error finding container 61e10ef5c21cbb4e674c4ab0586b82bc9ca40a4baad239b081248a215f23b0c6: Status 404 returned error can't find the container with id 61e10ef5c21cbb4e674c4ab0586b82bc9ca40a4baad239b081248a215f23b0c6 Oct 05 21:11:27 crc kubenswrapper[4754]: I1005 21:11:27.859082 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-5mdmj"] Oct 05 21:11:27 crc kubenswrapper[4754]: I1005 21:11:27.867587 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-qpgn5"] Oct 05 21:11:27 crc kubenswrapper[4754]: W1005 21:11:27.869618 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc0678fc1_b079_4379_8bbc_8848c3848f02.slice/crio-b7084df6d89fe4f7a1cdc8d2c09500dfab20fe6c8a9cfe8b71a0cb9991f19445 WatchSource:0}: Error finding container b7084df6d89fe4f7a1cdc8d2c09500dfab20fe6c8a9cfe8b71a0cb9991f19445: Status 404 returned error can't find the container with id b7084df6d89fe4f7a1cdc8d2c09500dfab20fe6c8a9cfe8b71a0cb9991f19445 Oct 05 21:11:27 crc kubenswrapper[4754]: I1005 21:11:27.938365 4754 generic.go:334] "Generic (PLEG): container finished" podID="9a72c674-5986-45a3-aae6-7f15de18efed" containerID="8667484195cb7e3c97d3a0017d2bb486586073eaa74b701c1abddcc59d96b189" exitCode=0 Oct 05 21:11:27 crc kubenswrapper[4754]: I1005 21:11:27.938434 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-679wz" event={"ID":"9a72c674-5986-45a3-aae6-7f15de18efed","Type":"ContainerDied","Data":"8667484195cb7e3c97d3a0017d2bb486586073eaa74b701c1abddcc59d96b189"} Oct 05 21:11:27 crc kubenswrapper[4754]: I1005 21:11:27.938463 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-679wz" event={"ID":"9a72c674-5986-45a3-aae6-7f15de18efed","Type":"ContainerStarted","Data":"84ab4cc4bcaec2bc885e6e3708a97369cb44dc4592d1ba2198a9eace0eaa0f82"} Oct 05 21:11:27 crc kubenswrapper[4754]: I1005 21:11:27.942830 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-5mdmj" event={"ID":"0f34b855-e478-41a0-b7b4-28dcf667b7c2","Type":"ContainerStarted","Data":"54c4d8c9935cd6e025dc3388a8e4eb6c6188f5f4c4416c251127e11c38d28154"} Oct 05 21:11:27 crc kubenswrapper[4754]: I1005 21:11:27.945006 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-hh4gc" event={"ID":"c636e2d4-bd61-47ff-ba53-e7da442521ce","Type":"ContainerDied","Data":"e92f12d5a4e18bd997af43c419ac2d9e9ae5bd5ce362214f7e6bce07f42cda4b"} Oct 05 21:11:27 crc kubenswrapper[4754]: I1005 21:11:27.945037 4754 scope.go:117] "RemoveContainer" containerID="d487267b338cdf082ec536549e664a2ceda532b59c086ea25c0f0500412b65b9" Oct 05 21:11:27 crc kubenswrapper[4754]: I1005 21:11:27.945036 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-hh4gc" Oct 05 21:11:27 crc kubenswrapper[4754]: I1005 21:11:27.950884 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-qpgn5" event={"ID":"c0678fc1-b079-4379-8bbc-8848c3848f02","Type":"ContainerStarted","Data":"b7084df6d89fe4f7a1cdc8d2c09500dfab20fe6c8a9cfe8b71a0cb9991f19445"} Oct 05 21:11:27 crc kubenswrapper[4754]: I1005 21:11:27.951926 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-9fzft" event={"ID":"87ecff30-d447-417b-9fe5-97cd6fd64c93","Type":"ContainerStarted","Data":"61e10ef5c21cbb4e674c4ab0586b82bc9ca40a4baad239b081248a215f23b0c6"} Oct 05 21:11:27 crc kubenswrapper[4754]: I1005 21:11:27.992229 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-dvzwp" podStartSLOduration=3.108217002 podStartE2EDuration="19.992206436s" podCreationTimestamp="2025-10-05 21:11:08 +0000 UTC" firstStartedPulling="2025-10-05 21:11:10.357467543 +0000 UTC m=+994.261586253" lastFinishedPulling="2025-10-05 21:11:27.241456977 +0000 UTC m=+1011.145575687" observedRunningTime="2025-10-05 21:11:27.980611557 +0000 UTC m=+1011.884730297" watchObservedRunningTime="2025-10-05 21:11:27.992206436 +0000 UTC m=+1011.896325146" Oct 05 21:11:27 crc kubenswrapper[4754]: I1005 21:11:27.996539 4754 scope.go:117] "RemoveContainer" containerID="378fc5007a8d78723eabecf80c7b1dda63d132593907e383e3dae39d94e14ea2" Oct 05 21:11:28 crc kubenswrapper[4754]: I1005 21:11:28.007589 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-hh4gc"] Oct 05 21:11:28 crc kubenswrapper[4754]: I1005 21:11:28.014477 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-698758b865-hh4gc"] Oct 05 21:11:28 crc kubenswrapper[4754]: I1005 21:11:28.863517 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c636e2d4-bd61-47ff-ba53-e7da442521ce" path="/var/lib/kubelet/pods/c636e2d4-bd61-47ff-ba53-e7da442521ce/volumes" Oct 05 21:11:28 crc kubenswrapper[4754]: I1005 21:11:28.962918 4754 generic.go:334] "Generic (PLEG): container finished" podID="0f34b855-e478-41a0-b7b4-28dcf667b7c2" containerID="05effd9954d08807f18e5f05169d35b21126fe1f01e7b273fe9f9107f13be6b4" exitCode=0 Oct 05 21:11:28 crc kubenswrapper[4754]: I1005 21:11:28.962974 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-5mdmj" event={"ID":"0f34b855-e478-41a0-b7b4-28dcf667b7c2","Type":"ContainerDied","Data":"05effd9954d08807f18e5f05169d35b21126fe1f01e7b273fe9f9107f13be6b4"} Oct 05 21:11:28 crc kubenswrapper[4754]: I1005 21:11:28.966819 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-dvzwp" event={"ID":"d20989a7-98b9-4bf2-9f85-d72890ef7e1a","Type":"ContainerStarted","Data":"fdd3e88cffd311578243f0016120a4d7c1b3556ae7bc395b06d22414814fa856"} Oct 05 21:11:28 crc kubenswrapper[4754]: I1005 21:11:28.970586 4754 generic.go:334] "Generic (PLEG): container finished" podID="c0678fc1-b079-4379-8bbc-8848c3848f02" containerID="99792af3e76c6ce02963c39a0a6f464ab5bf88393b412a908230ca3a3b9fac0a" exitCode=0 Oct 05 21:11:28 crc kubenswrapper[4754]: I1005 21:11:28.970849 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-qpgn5" event={"ID":"c0678fc1-b079-4379-8bbc-8848c3848f02","Type":"ContainerDied","Data":"99792af3e76c6ce02963c39a0a6f464ab5bf88393b412a908230ca3a3b9fac0a"} Oct 05 21:11:29 crc kubenswrapper[4754]: I1005 21:11:29.277768 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-679wz" Oct 05 21:11:29 crc kubenswrapper[4754]: I1005 21:11:29.365028 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q6gfh\" (UniqueName: \"kubernetes.io/projected/9a72c674-5986-45a3-aae6-7f15de18efed-kube-api-access-q6gfh\") pod \"9a72c674-5986-45a3-aae6-7f15de18efed\" (UID: \"9a72c674-5986-45a3-aae6-7f15de18efed\") " Oct 05 21:11:29 crc kubenswrapper[4754]: I1005 21:11:29.387396 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9a72c674-5986-45a3-aae6-7f15de18efed-kube-api-access-q6gfh" (OuterVolumeSpecName: "kube-api-access-q6gfh") pod "9a72c674-5986-45a3-aae6-7f15de18efed" (UID: "9a72c674-5986-45a3-aae6-7f15de18efed"). InnerVolumeSpecName "kube-api-access-q6gfh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:11:29 crc kubenswrapper[4754]: I1005 21:11:29.466874 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q6gfh\" (UniqueName: \"kubernetes.io/projected/9a72c674-5986-45a3-aae6-7f15de18efed-kube-api-access-q6gfh\") on node \"crc\" DevicePath \"\"" Oct 05 21:11:29 crc kubenswrapper[4754]: I1005 21:11:29.993518 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-679wz" event={"ID":"9a72c674-5986-45a3-aae6-7f15de18efed","Type":"ContainerDied","Data":"84ab4cc4bcaec2bc885e6e3708a97369cb44dc4592d1ba2198a9eace0eaa0f82"} Oct 05 21:11:29 crc kubenswrapper[4754]: I1005 21:11:29.993580 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="84ab4cc4bcaec2bc885e6e3708a97369cb44dc4592d1ba2198a9eace0eaa0f82" Oct 05 21:11:29 crc kubenswrapper[4754]: I1005 21:11:29.993653 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-679wz" Oct 05 21:11:35 crc kubenswrapper[4754]: I1005 21:11:35.475201 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-qpgn5" Oct 05 21:11:35 crc kubenswrapper[4754]: I1005 21:11:35.500450 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-5mdmj" Oct 05 21:11:35 crc kubenswrapper[4754]: I1005 21:11:35.590846 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q9wpv\" (UniqueName: \"kubernetes.io/projected/c0678fc1-b079-4379-8bbc-8848c3848f02-kube-api-access-q9wpv\") pod \"c0678fc1-b079-4379-8bbc-8848c3848f02\" (UID: \"c0678fc1-b079-4379-8bbc-8848c3848f02\") " Oct 05 21:11:35 crc kubenswrapper[4754]: I1005 21:11:35.591624 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vlnbv\" (UniqueName: \"kubernetes.io/projected/0f34b855-e478-41a0-b7b4-28dcf667b7c2-kube-api-access-vlnbv\") pod \"0f34b855-e478-41a0-b7b4-28dcf667b7c2\" (UID: \"0f34b855-e478-41a0-b7b4-28dcf667b7c2\") " Oct 05 21:11:35 crc kubenswrapper[4754]: I1005 21:11:35.599777 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0f34b855-e478-41a0-b7b4-28dcf667b7c2-kube-api-access-vlnbv" (OuterVolumeSpecName: "kube-api-access-vlnbv") pod "0f34b855-e478-41a0-b7b4-28dcf667b7c2" (UID: "0f34b855-e478-41a0-b7b4-28dcf667b7c2"). InnerVolumeSpecName "kube-api-access-vlnbv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:11:35 crc kubenswrapper[4754]: I1005 21:11:35.600002 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c0678fc1-b079-4379-8bbc-8848c3848f02-kube-api-access-q9wpv" (OuterVolumeSpecName: "kube-api-access-q9wpv") pod "c0678fc1-b079-4379-8bbc-8848c3848f02" (UID: "c0678fc1-b079-4379-8bbc-8848c3848f02"). InnerVolumeSpecName "kube-api-access-q9wpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:11:35 crc kubenswrapper[4754]: I1005 21:11:35.694046 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q9wpv\" (UniqueName: \"kubernetes.io/projected/c0678fc1-b079-4379-8bbc-8848c3848f02-kube-api-access-q9wpv\") on node \"crc\" DevicePath \"\"" Oct 05 21:11:35 crc kubenswrapper[4754]: I1005 21:11:35.694322 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vlnbv\" (UniqueName: \"kubernetes.io/projected/0f34b855-e478-41a0-b7b4-28dcf667b7c2-kube-api-access-vlnbv\") on node \"crc\" DevicePath \"\"" Oct 05 21:11:36 crc kubenswrapper[4754]: I1005 21:11:36.057162 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-9fzft" event={"ID":"87ecff30-d447-417b-9fe5-97cd6fd64c93","Type":"ContainerStarted","Data":"6ec2c24cd529e409522a7d5601c008441f4e50246937630801fbcd282d2978b4"} Oct 05 21:11:36 crc kubenswrapper[4754]: I1005 21:11:36.060866 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-5mdmj" event={"ID":"0f34b855-e478-41a0-b7b4-28dcf667b7c2","Type":"ContainerDied","Data":"54c4d8c9935cd6e025dc3388a8e4eb6c6188f5f4c4416c251127e11c38d28154"} Oct 05 21:11:36 crc kubenswrapper[4754]: I1005 21:11:36.060904 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="54c4d8c9935cd6e025dc3388a8e4eb6c6188f5f4c4416c251127e11c38d28154" Oct 05 21:11:36 crc kubenswrapper[4754]: I1005 21:11:36.060971 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-5mdmj" Oct 05 21:11:36 crc kubenswrapper[4754]: I1005 21:11:36.063899 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-qpgn5" event={"ID":"c0678fc1-b079-4379-8bbc-8848c3848f02","Type":"ContainerDied","Data":"b7084df6d89fe4f7a1cdc8d2c09500dfab20fe6c8a9cfe8b71a0cb9991f19445"} Oct 05 21:11:36 crc kubenswrapper[4754]: I1005 21:11:36.063931 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b7084df6d89fe4f7a1cdc8d2c09500dfab20fe6c8a9cfe8b71a0cb9991f19445" Oct 05 21:11:36 crc kubenswrapper[4754]: I1005 21:11:36.063945 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-qpgn5" Oct 05 21:11:36 crc kubenswrapper[4754]: I1005 21:11:36.092909 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-9fzft" podStartSLOduration=7.610256832 podStartE2EDuration="15.09287155s" podCreationTimestamp="2025-10-05 21:11:21 +0000 UTC" firstStartedPulling="2025-10-05 21:11:27.776730039 +0000 UTC m=+1011.680848749" lastFinishedPulling="2025-10-05 21:11:35.259344747 +0000 UTC m=+1019.163463467" observedRunningTime="2025-10-05 21:11:36.086833329 +0000 UTC m=+1019.990952039" watchObservedRunningTime="2025-10-05 21:11:36.09287155 +0000 UTC m=+1019.996990260" Oct 05 21:11:39 crc kubenswrapper[4754]: I1005 21:11:39.095320 4754 generic.go:334] "Generic (PLEG): container finished" podID="87ecff30-d447-417b-9fe5-97cd6fd64c93" containerID="6ec2c24cd529e409522a7d5601c008441f4e50246937630801fbcd282d2978b4" exitCode=0 Oct 05 21:11:39 crc kubenswrapper[4754]: I1005 21:11:39.095459 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-9fzft" event={"ID":"87ecff30-d447-417b-9fe5-97cd6fd64c93","Type":"ContainerDied","Data":"6ec2c24cd529e409522a7d5601c008441f4e50246937630801fbcd282d2978b4"} Oct 05 21:11:40 crc kubenswrapper[4754]: I1005 21:11:40.107090 4754 generic.go:334] "Generic (PLEG): container finished" podID="d20989a7-98b9-4bf2-9f85-d72890ef7e1a" containerID="fdd3e88cffd311578243f0016120a4d7c1b3556ae7bc395b06d22414814fa856" exitCode=0 Oct 05 21:11:40 crc kubenswrapper[4754]: I1005 21:11:40.107187 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-dvzwp" event={"ID":"d20989a7-98b9-4bf2-9f85-d72890ef7e1a","Type":"ContainerDied","Data":"fdd3e88cffd311578243f0016120a4d7c1b3556ae7bc395b06d22414814fa856"} Oct 05 21:11:40 crc kubenswrapper[4754]: I1005 21:11:40.430928 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-9fzft" Oct 05 21:11:40 crc kubenswrapper[4754]: I1005 21:11:40.482569 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87ecff30-d447-417b-9fe5-97cd6fd64c93-combined-ca-bundle\") pod \"87ecff30-d447-417b-9fe5-97cd6fd64c93\" (UID: \"87ecff30-d447-417b-9fe5-97cd6fd64c93\") " Oct 05 21:11:40 crc kubenswrapper[4754]: I1005 21:11:40.482652 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87ecff30-d447-417b-9fe5-97cd6fd64c93-config-data\") pod \"87ecff30-d447-417b-9fe5-97cd6fd64c93\" (UID: \"87ecff30-d447-417b-9fe5-97cd6fd64c93\") " Oct 05 21:11:40 crc kubenswrapper[4754]: I1005 21:11:40.482700 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6qg4l\" (UniqueName: \"kubernetes.io/projected/87ecff30-d447-417b-9fe5-97cd6fd64c93-kube-api-access-6qg4l\") pod \"87ecff30-d447-417b-9fe5-97cd6fd64c93\" (UID: \"87ecff30-d447-417b-9fe5-97cd6fd64c93\") " Oct 05 21:11:40 crc kubenswrapper[4754]: I1005 21:11:40.499731 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87ecff30-d447-417b-9fe5-97cd6fd64c93-kube-api-access-6qg4l" (OuterVolumeSpecName: "kube-api-access-6qg4l") pod "87ecff30-d447-417b-9fe5-97cd6fd64c93" (UID: "87ecff30-d447-417b-9fe5-97cd6fd64c93"). InnerVolumeSpecName "kube-api-access-6qg4l". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:11:40 crc kubenswrapper[4754]: I1005 21:11:40.515254 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87ecff30-d447-417b-9fe5-97cd6fd64c93-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "87ecff30-d447-417b-9fe5-97cd6fd64c93" (UID: "87ecff30-d447-417b-9fe5-97cd6fd64c93"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:11:40 crc kubenswrapper[4754]: I1005 21:11:40.542624 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-3385-account-create-62sw4"] Oct 05 21:11:40 crc kubenswrapper[4754]: E1005 21:11:40.543308 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f34b855-e478-41a0-b7b4-28dcf667b7c2" containerName="mariadb-database-create" Oct 05 21:11:40 crc kubenswrapper[4754]: I1005 21:11:40.543333 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f34b855-e478-41a0-b7b4-28dcf667b7c2" containerName="mariadb-database-create" Oct 05 21:11:40 crc kubenswrapper[4754]: E1005 21:11:40.543382 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c636e2d4-bd61-47ff-ba53-e7da442521ce" containerName="init" Oct 05 21:11:40 crc kubenswrapper[4754]: I1005 21:11:40.543395 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="c636e2d4-bd61-47ff-ba53-e7da442521ce" containerName="init" Oct 05 21:11:40 crc kubenswrapper[4754]: E1005 21:11:40.543416 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87ecff30-d447-417b-9fe5-97cd6fd64c93" containerName="keystone-db-sync" Oct 05 21:11:40 crc kubenswrapper[4754]: I1005 21:11:40.543430 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="87ecff30-d447-417b-9fe5-97cd6fd64c93" containerName="keystone-db-sync" Oct 05 21:11:40 crc kubenswrapper[4754]: E1005 21:11:40.543455 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c636e2d4-bd61-47ff-ba53-e7da442521ce" containerName="dnsmasq-dns" Oct 05 21:11:40 crc kubenswrapper[4754]: I1005 21:11:40.543468 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="c636e2d4-bd61-47ff-ba53-e7da442521ce" containerName="dnsmasq-dns" Oct 05 21:11:40 crc kubenswrapper[4754]: E1005 21:11:40.543519 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a72c674-5986-45a3-aae6-7f15de18efed" containerName="mariadb-database-create" Oct 05 21:11:40 crc kubenswrapper[4754]: I1005 21:11:40.543533 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a72c674-5986-45a3-aae6-7f15de18efed" containerName="mariadb-database-create" Oct 05 21:11:40 crc kubenswrapper[4754]: E1005 21:11:40.543551 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c0678fc1-b079-4379-8bbc-8848c3848f02" containerName="mariadb-database-create" Oct 05 21:11:40 crc kubenswrapper[4754]: I1005 21:11:40.543564 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="c0678fc1-b079-4379-8bbc-8848c3848f02" containerName="mariadb-database-create" Oct 05 21:11:40 crc kubenswrapper[4754]: I1005 21:11:40.543903 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="c0678fc1-b079-4379-8bbc-8848c3848f02" containerName="mariadb-database-create" Oct 05 21:11:40 crc kubenswrapper[4754]: I1005 21:11:40.543930 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="c636e2d4-bd61-47ff-ba53-e7da442521ce" containerName="dnsmasq-dns" Oct 05 21:11:40 crc kubenswrapper[4754]: I1005 21:11:40.543955 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a72c674-5986-45a3-aae6-7f15de18efed" containerName="mariadb-database-create" Oct 05 21:11:40 crc kubenswrapper[4754]: I1005 21:11:40.543991 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="87ecff30-d447-417b-9fe5-97cd6fd64c93" containerName="keystone-db-sync" Oct 05 21:11:40 crc kubenswrapper[4754]: I1005 21:11:40.544032 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f34b855-e478-41a0-b7b4-28dcf667b7c2" containerName="mariadb-database-create" Oct 05 21:11:40 crc kubenswrapper[4754]: I1005 21:11:40.545044 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-3385-account-create-62sw4" Oct 05 21:11:40 crc kubenswrapper[4754]: I1005 21:11:40.552314 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Oct 05 21:11:40 crc kubenswrapper[4754]: I1005 21:11:40.569471 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-3385-account-create-62sw4"] Oct 05 21:11:40 crc kubenswrapper[4754]: I1005 21:11:40.579469 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87ecff30-d447-417b-9fe5-97cd6fd64c93-config-data" (OuterVolumeSpecName: "config-data") pod "87ecff30-d447-417b-9fe5-97cd6fd64c93" (UID: "87ecff30-d447-417b-9fe5-97cd6fd64c93"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:11:40 crc kubenswrapper[4754]: I1005 21:11:40.584614 4754 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87ecff30-d447-417b-9fe5-97cd6fd64c93-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 21:11:40 crc kubenswrapper[4754]: I1005 21:11:40.584795 4754 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87ecff30-d447-417b-9fe5-97cd6fd64c93-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 21:11:40 crc kubenswrapper[4754]: I1005 21:11:40.584942 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6qg4l\" (UniqueName: \"kubernetes.io/projected/87ecff30-d447-417b-9fe5-97cd6fd64c93-kube-api-access-6qg4l\") on node \"crc\" DevicePath \"\"" Oct 05 21:11:40 crc kubenswrapper[4754]: I1005 21:11:40.675266 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-dfec-account-create-qdgw4"] Oct 05 21:11:40 crc kubenswrapper[4754]: I1005 21:11:40.676658 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dfec-account-create-qdgw4" Oct 05 21:11:40 crc kubenswrapper[4754]: I1005 21:11:40.678895 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Oct 05 21:11:40 crc kubenswrapper[4754]: I1005 21:11:40.687223 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c6nvg\" (UniqueName: \"kubernetes.io/projected/b7ded8c6-2bd5-46ca-b36f-68ccbd92366d-kube-api-access-c6nvg\") pod \"cinder-3385-account-create-62sw4\" (UID: \"b7ded8c6-2bd5-46ca-b36f-68ccbd92366d\") " pod="openstack/cinder-3385-account-create-62sw4" Oct 05 21:11:40 crc kubenswrapper[4754]: I1005 21:11:40.693988 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-dfec-account-create-qdgw4"] Oct 05 21:11:40 crc kubenswrapper[4754]: I1005 21:11:40.789161 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c6nvg\" (UniqueName: \"kubernetes.io/projected/b7ded8c6-2bd5-46ca-b36f-68ccbd92366d-kube-api-access-c6nvg\") pod \"cinder-3385-account-create-62sw4\" (UID: \"b7ded8c6-2bd5-46ca-b36f-68ccbd92366d\") " pod="openstack/cinder-3385-account-create-62sw4" Oct 05 21:11:40 crc kubenswrapper[4754]: I1005 21:11:40.789256 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wkq8n\" (UniqueName: \"kubernetes.io/projected/a454365f-a32c-427d-a2d1-cecff2501134-kube-api-access-wkq8n\") pod \"neutron-dfec-account-create-qdgw4\" (UID: \"a454365f-a32c-427d-a2d1-cecff2501134\") " pod="openstack/neutron-dfec-account-create-qdgw4" Oct 05 21:11:40 crc kubenswrapper[4754]: I1005 21:11:40.810422 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c6nvg\" (UniqueName: \"kubernetes.io/projected/b7ded8c6-2bd5-46ca-b36f-68ccbd92366d-kube-api-access-c6nvg\") pod \"cinder-3385-account-create-62sw4\" (UID: \"b7ded8c6-2bd5-46ca-b36f-68ccbd92366d\") " pod="openstack/cinder-3385-account-create-62sw4" Oct 05 21:11:40 crc kubenswrapper[4754]: I1005 21:11:40.890986 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wkq8n\" (UniqueName: \"kubernetes.io/projected/a454365f-a32c-427d-a2d1-cecff2501134-kube-api-access-wkq8n\") pod \"neutron-dfec-account-create-qdgw4\" (UID: \"a454365f-a32c-427d-a2d1-cecff2501134\") " pod="openstack/neutron-dfec-account-create-qdgw4" Oct 05 21:11:40 crc kubenswrapper[4754]: I1005 21:11:40.916460 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-3385-account-create-62sw4" Oct 05 21:11:40 crc kubenswrapper[4754]: I1005 21:11:40.921227 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wkq8n\" (UniqueName: \"kubernetes.io/projected/a454365f-a32c-427d-a2d1-cecff2501134-kube-api-access-wkq8n\") pod \"neutron-dfec-account-create-qdgw4\" (UID: \"a454365f-a32c-427d-a2d1-cecff2501134\") " pod="openstack/neutron-dfec-account-create-qdgw4" Oct 05 21:11:40 crc kubenswrapper[4754]: I1005 21:11:40.991487 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dfec-account-create-qdgw4" Oct 05 21:11:41 crc kubenswrapper[4754]: I1005 21:11:41.125041 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-9fzft" Oct 05 21:11:41 crc kubenswrapper[4754]: I1005 21:11:41.125028 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-9fzft" event={"ID":"87ecff30-d447-417b-9fe5-97cd6fd64c93","Type":"ContainerDied","Data":"61e10ef5c21cbb4e674c4ab0586b82bc9ca40a4baad239b081248a215f23b0c6"} Oct 05 21:11:41 crc kubenswrapper[4754]: I1005 21:11:41.125104 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="61e10ef5c21cbb4e674c4ab0586b82bc9ca40a4baad239b081248a215f23b0c6" Oct 05 21:11:41 crc kubenswrapper[4754]: I1005 21:11:41.422242 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-55fff446b9-nvb2b"] Oct 05 21:11:41 crc kubenswrapper[4754]: I1005 21:11:41.433215 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55fff446b9-nvb2b" Oct 05 21:11:41 crc kubenswrapper[4754]: I1005 21:11:41.460744 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55fff446b9-nvb2b"] Oct 05 21:11:41 crc kubenswrapper[4754]: I1005 21:11:41.492586 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-fvkbb"] Oct 05 21:11:41 crc kubenswrapper[4754]: I1005 21:11:41.493750 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-fvkbb" Oct 05 21:11:41 crc kubenswrapper[4754]: I1005 21:11:41.504550 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 05 21:11:41 crc kubenswrapper[4754]: I1005 21:11:41.504833 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 05 21:11:41 crc kubenswrapper[4754]: I1005 21:11:41.505452 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4752f446-876a-4d46-8e36-b4b03143c9db-dns-svc\") pod \"dnsmasq-dns-55fff446b9-nvb2b\" (UID: \"4752f446-876a-4d46-8e36-b4b03143c9db\") " pod="openstack/dnsmasq-dns-55fff446b9-nvb2b" Oct 05 21:11:41 crc kubenswrapper[4754]: I1005 21:11:41.505520 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4752f446-876a-4d46-8e36-b4b03143c9db-ovsdbserver-nb\") pod \"dnsmasq-dns-55fff446b9-nvb2b\" (UID: \"4752f446-876a-4d46-8e36-b4b03143c9db\") " pod="openstack/dnsmasq-dns-55fff446b9-nvb2b" Oct 05 21:11:41 crc kubenswrapper[4754]: I1005 21:11:41.505592 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4752f446-876a-4d46-8e36-b4b03143c9db-config\") pod \"dnsmasq-dns-55fff446b9-nvb2b\" (UID: \"4752f446-876a-4d46-8e36-b4b03143c9db\") " pod="openstack/dnsmasq-dns-55fff446b9-nvb2b" Oct 05 21:11:41 crc kubenswrapper[4754]: I1005 21:11:41.505623 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4752f446-876a-4d46-8e36-b4b03143c9db-ovsdbserver-sb\") pod \"dnsmasq-dns-55fff446b9-nvb2b\" (UID: \"4752f446-876a-4d46-8e36-b4b03143c9db\") " pod="openstack/dnsmasq-dns-55fff446b9-nvb2b" Oct 05 21:11:41 crc kubenswrapper[4754]: I1005 21:11:41.505679 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4752f446-876a-4d46-8e36-b4b03143c9db-dns-swift-storage-0\") pod \"dnsmasq-dns-55fff446b9-nvb2b\" (UID: \"4752f446-876a-4d46-8e36-b4b03143c9db\") " pod="openstack/dnsmasq-dns-55fff446b9-nvb2b" Oct 05 21:11:41 crc kubenswrapper[4754]: I1005 21:11:41.505746 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-95xqp\" (UniqueName: \"kubernetes.io/projected/4752f446-876a-4d46-8e36-b4b03143c9db-kube-api-access-95xqp\") pod \"dnsmasq-dns-55fff446b9-nvb2b\" (UID: \"4752f446-876a-4d46-8e36-b4b03143c9db\") " pod="openstack/dnsmasq-dns-55fff446b9-nvb2b" Oct 05 21:11:41 crc kubenswrapper[4754]: I1005 21:11:41.511689 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 05 21:11:41 crc kubenswrapper[4754]: I1005 21:11:41.512905 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-5846s" Oct 05 21:11:41 crc kubenswrapper[4754]: I1005 21:11:41.563953 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-3385-account-create-62sw4"] Oct 05 21:11:41 crc kubenswrapper[4754]: I1005 21:11:41.607214 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cbf1bb4f-3997-4581-ba76-64b10ba3f686-config-data\") pod \"keystone-bootstrap-fvkbb\" (UID: \"cbf1bb4f-3997-4581-ba76-64b10ba3f686\") " pod="openstack/keystone-bootstrap-fvkbb" Oct 05 21:11:41 crc kubenswrapper[4754]: I1005 21:11:41.607272 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4752f446-876a-4d46-8e36-b4b03143c9db-dns-svc\") pod \"dnsmasq-dns-55fff446b9-nvb2b\" (UID: \"4752f446-876a-4d46-8e36-b4b03143c9db\") " pod="openstack/dnsmasq-dns-55fff446b9-nvb2b" Oct 05 21:11:41 crc kubenswrapper[4754]: I1005 21:11:41.607298 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbf1bb4f-3997-4581-ba76-64b10ba3f686-combined-ca-bundle\") pod \"keystone-bootstrap-fvkbb\" (UID: \"cbf1bb4f-3997-4581-ba76-64b10ba3f686\") " pod="openstack/keystone-bootstrap-fvkbb" Oct 05 21:11:41 crc kubenswrapper[4754]: I1005 21:11:41.607324 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4752f446-876a-4d46-8e36-b4b03143c9db-ovsdbserver-nb\") pod \"dnsmasq-dns-55fff446b9-nvb2b\" (UID: \"4752f446-876a-4d46-8e36-b4b03143c9db\") " pod="openstack/dnsmasq-dns-55fff446b9-nvb2b" Oct 05 21:11:41 crc kubenswrapper[4754]: I1005 21:11:41.607352 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cbf1bb4f-3997-4581-ba76-64b10ba3f686-scripts\") pod \"keystone-bootstrap-fvkbb\" (UID: \"cbf1bb4f-3997-4581-ba76-64b10ba3f686\") " pod="openstack/keystone-bootstrap-fvkbb" Oct 05 21:11:41 crc kubenswrapper[4754]: I1005 21:11:41.609617 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4752f446-876a-4d46-8e36-b4b03143c9db-config\") pod \"dnsmasq-dns-55fff446b9-nvb2b\" (UID: \"4752f446-876a-4d46-8e36-b4b03143c9db\") " pod="openstack/dnsmasq-dns-55fff446b9-nvb2b" Oct 05 21:11:41 crc kubenswrapper[4754]: I1005 21:11:41.609648 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4752f446-876a-4d46-8e36-b4b03143c9db-ovsdbserver-sb\") pod \"dnsmasq-dns-55fff446b9-nvb2b\" (UID: \"4752f446-876a-4d46-8e36-b4b03143c9db\") " pod="openstack/dnsmasq-dns-55fff446b9-nvb2b" Oct 05 21:11:41 crc kubenswrapper[4754]: I1005 21:11:41.609693 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4752f446-876a-4d46-8e36-b4b03143c9db-dns-swift-storage-0\") pod \"dnsmasq-dns-55fff446b9-nvb2b\" (UID: \"4752f446-876a-4d46-8e36-b4b03143c9db\") " pod="openstack/dnsmasq-dns-55fff446b9-nvb2b" Oct 05 21:11:41 crc kubenswrapper[4754]: I1005 21:11:41.609724 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/cbf1bb4f-3997-4581-ba76-64b10ba3f686-credential-keys\") pod \"keystone-bootstrap-fvkbb\" (UID: \"cbf1bb4f-3997-4581-ba76-64b10ba3f686\") " pod="openstack/keystone-bootstrap-fvkbb" Oct 05 21:11:41 crc kubenswrapper[4754]: I1005 21:11:41.609778 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-95xqp\" (UniqueName: \"kubernetes.io/projected/4752f446-876a-4d46-8e36-b4b03143c9db-kube-api-access-95xqp\") pod \"dnsmasq-dns-55fff446b9-nvb2b\" (UID: \"4752f446-876a-4d46-8e36-b4b03143c9db\") " pod="openstack/dnsmasq-dns-55fff446b9-nvb2b" Oct 05 21:11:41 crc kubenswrapper[4754]: I1005 21:11:41.609801 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/cbf1bb4f-3997-4581-ba76-64b10ba3f686-fernet-keys\") pod \"keystone-bootstrap-fvkbb\" (UID: \"cbf1bb4f-3997-4581-ba76-64b10ba3f686\") " pod="openstack/keystone-bootstrap-fvkbb" Oct 05 21:11:41 crc kubenswrapper[4754]: I1005 21:11:41.609827 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2zgjt\" (UniqueName: \"kubernetes.io/projected/cbf1bb4f-3997-4581-ba76-64b10ba3f686-kube-api-access-2zgjt\") pod \"keystone-bootstrap-fvkbb\" (UID: \"cbf1bb4f-3997-4581-ba76-64b10ba3f686\") " pod="openstack/keystone-bootstrap-fvkbb" Oct 05 21:11:41 crc kubenswrapper[4754]: I1005 21:11:41.615419 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4752f446-876a-4d46-8e36-b4b03143c9db-dns-svc\") pod \"dnsmasq-dns-55fff446b9-nvb2b\" (UID: \"4752f446-876a-4d46-8e36-b4b03143c9db\") " pod="openstack/dnsmasq-dns-55fff446b9-nvb2b" Oct 05 21:11:41 crc kubenswrapper[4754]: I1005 21:11:41.615991 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4752f446-876a-4d46-8e36-b4b03143c9db-ovsdbserver-nb\") pod \"dnsmasq-dns-55fff446b9-nvb2b\" (UID: \"4752f446-876a-4d46-8e36-b4b03143c9db\") " pod="openstack/dnsmasq-dns-55fff446b9-nvb2b" Oct 05 21:11:41 crc kubenswrapper[4754]: I1005 21:11:41.616529 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4752f446-876a-4d46-8e36-b4b03143c9db-config\") pod \"dnsmasq-dns-55fff446b9-nvb2b\" (UID: \"4752f446-876a-4d46-8e36-b4b03143c9db\") " pod="openstack/dnsmasq-dns-55fff446b9-nvb2b" Oct 05 21:11:41 crc kubenswrapper[4754]: I1005 21:11:41.624408 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4752f446-876a-4d46-8e36-b4b03143c9db-ovsdbserver-sb\") pod \"dnsmasq-dns-55fff446b9-nvb2b\" (UID: \"4752f446-876a-4d46-8e36-b4b03143c9db\") " pod="openstack/dnsmasq-dns-55fff446b9-nvb2b" Oct 05 21:11:41 crc kubenswrapper[4754]: I1005 21:11:41.629107 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4752f446-876a-4d46-8e36-b4b03143c9db-dns-swift-storage-0\") pod \"dnsmasq-dns-55fff446b9-nvb2b\" (UID: \"4752f446-876a-4d46-8e36-b4b03143c9db\") " pod="openstack/dnsmasq-dns-55fff446b9-nvb2b" Oct 05 21:11:41 crc kubenswrapper[4754]: I1005 21:11:41.694783 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-fvkbb"] Oct 05 21:11:41 crc kubenswrapper[4754]: I1005 21:11:41.735721 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/cbf1bb4f-3997-4581-ba76-64b10ba3f686-fernet-keys\") pod \"keystone-bootstrap-fvkbb\" (UID: \"cbf1bb4f-3997-4581-ba76-64b10ba3f686\") " pod="openstack/keystone-bootstrap-fvkbb" Oct 05 21:11:41 crc kubenswrapper[4754]: I1005 21:11:41.735858 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2zgjt\" (UniqueName: \"kubernetes.io/projected/cbf1bb4f-3997-4581-ba76-64b10ba3f686-kube-api-access-2zgjt\") pod \"keystone-bootstrap-fvkbb\" (UID: \"cbf1bb4f-3997-4581-ba76-64b10ba3f686\") " pod="openstack/keystone-bootstrap-fvkbb" Oct 05 21:11:41 crc kubenswrapper[4754]: I1005 21:11:41.747412 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cbf1bb4f-3997-4581-ba76-64b10ba3f686-config-data\") pod \"keystone-bootstrap-fvkbb\" (UID: \"cbf1bb4f-3997-4581-ba76-64b10ba3f686\") " pod="openstack/keystone-bootstrap-fvkbb" Oct 05 21:11:41 crc kubenswrapper[4754]: I1005 21:11:41.747583 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbf1bb4f-3997-4581-ba76-64b10ba3f686-combined-ca-bundle\") pod \"keystone-bootstrap-fvkbb\" (UID: \"cbf1bb4f-3997-4581-ba76-64b10ba3f686\") " pod="openstack/keystone-bootstrap-fvkbb" Oct 05 21:11:41 crc kubenswrapper[4754]: I1005 21:11:41.747652 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cbf1bb4f-3997-4581-ba76-64b10ba3f686-scripts\") pod \"keystone-bootstrap-fvkbb\" (UID: \"cbf1bb4f-3997-4581-ba76-64b10ba3f686\") " pod="openstack/keystone-bootstrap-fvkbb" Oct 05 21:11:41 crc kubenswrapper[4754]: I1005 21:11:41.747966 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/cbf1bb4f-3997-4581-ba76-64b10ba3f686-credential-keys\") pod \"keystone-bootstrap-fvkbb\" (UID: \"cbf1bb4f-3997-4581-ba76-64b10ba3f686\") " pod="openstack/keystone-bootstrap-fvkbb" Oct 05 21:11:41 crc kubenswrapper[4754]: I1005 21:11:41.752215 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/cbf1bb4f-3997-4581-ba76-64b10ba3f686-fernet-keys\") pod \"keystone-bootstrap-fvkbb\" (UID: \"cbf1bb4f-3997-4581-ba76-64b10ba3f686\") " pod="openstack/keystone-bootstrap-fvkbb" Oct 05 21:11:41 crc kubenswrapper[4754]: I1005 21:11:41.753665 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-95xqp\" (UniqueName: \"kubernetes.io/projected/4752f446-876a-4d46-8e36-b4b03143c9db-kube-api-access-95xqp\") pod \"dnsmasq-dns-55fff446b9-nvb2b\" (UID: \"4752f446-876a-4d46-8e36-b4b03143c9db\") " pod="openstack/dnsmasq-dns-55fff446b9-nvb2b" Oct 05 21:11:41 crc kubenswrapper[4754]: I1005 21:11:41.757023 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbf1bb4f-3997-4581-ba76-64b10ba3f686-combined-ca-bundle\") pod \"keystone-bootstrap-fvkbb\" (UID: \"cbf1bb4f-3997-4581-ba76-64b10ba3f686\") " pod="openstack/keystone-bootstrap-fvkbb" Oct 05 21:11:41 crc kubenswrapper[4754]: I1005 21:11:41.757853 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cbf1bb4f-3997-4581-ba76-64b10ba3f686-config-data\") pod \"keystone-bootstrap-fvkbb\" (UID: \"cbf1bb4f-3997-4581-ba76-64b10ba3f686\") " pod="openstack/keystone-bootstrap-fvkbb" Oct 05 21:11:41 crc kubenswrapper[4754]: I1005 21:11:41.766235 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cbf1bb4f-3997-4581-ba76-64b10ba3f686-scripts\") pod \"keystone-bootstrap-fvkbb\" (UID: \"cbf1bb4f-3997-4581-ba76-64b10ba3f686\") " pod="openstack/keystone-bootstrap-fvkbb" Oct 05 21:11:41 crc kubenswrapper[4754]: I1005 21:11:41.771864 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55fff446b9-nvb2b" Oct 05 21:11:41 crc kubenswrapper[4754]: I1005 21:11:41.809609 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2zgjt\" (UniqueName: \"kubernetes.io/projected/cbf1bb4f-3997-4581-ba76-64b10ba3f686-kube-api-access-2zgjt\") pod \"keystone-bootstrap-fvkbb\" (UID: \"cbf1bb4f-3997-4581-ba76-64b10ba3f686\") " pod="openstack/keystone-bootstrap-fvkbb" Oct 05 21:11:41 crc kubenswrapper[4754]: I1005 21:11:41.851304 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/cbf1bb4f-3997-4581-ba76-64b10ba3f686-credential-keys\") pod \"keystone-bootstrap-fvkbb\" (UID: \"cbf1bb4f-3997-4581-ba76-64b10ba3f686\") " pod="openstack/keystone-bootstrap-fvkbb" Oct 05 21:11:41 crc kubenswrapper[4754]: I1005 21:11:41.883119 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-dfec-account-create-qdgw4"] Oct 05 21:11:41 crc kubenswrapper[4754]: I1005 21:11:41.906053 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-fvkbb" Oct 05 21:11:41 crc kubenswrapper[4754]: I1005 21:11:41.911027 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-689b67764f-n4l6h"] Oct 05 21:11:41 crc kubenswrapper[4754]: I1005 21:11:41.913355 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-689b67764f-n4l6h" Oct 05 21:11:41 crc kubenswrapper[4754]: I1005 21:11:41.932097 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-nzt44" Oct 05 21:11:41 crc kubenswrapper[4754]: I1005 21:11:41.932307 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-689b67764f-n4l6h"] Oct 05 21:11:41 crc kubenswrapper[4754]: I1005 21:11:41.945659 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-xlw68"] Oct 05 21:11:41 crc kubenswrapper[4754]: I1005 21:11:41.946893 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-xlw68" Oct 05 21:11:41 crc kubenswrapper[4754]: I1005 21:11:41.977162 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-dvzwp" Oct 05 21:11:41 crc kubenswrapper[4754]: I1005 21:11:41.978555 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Oct 05 21:11:41 crc kubenswrapper[4754]: I1005 21:11:41.979087 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Oct 05 21:11:41 crc kubenswrapper[4754]: I1005 21:11:41.979481 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Oct 05 21:11:41 crc kubenswrapper[4754]: I1005 21:11:41.980144 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Oct 05 21:11:41 crc kubenswrapper[4754]: I1005 21:11:41.980190 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Oct 05 21:11:41 crc kubenswrapper[4754]: I1005 21:11:41.980333 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-8l4rk" Oct 05 21:11:42 crc kubenswrapper[4754]: I1005 21:11:42.018486 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55fff446b9-nvb2b"] Oct 05 21:11:42 crc kubenswrapper[4754]: I1005 21:11:42.057630 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d20989a7-98b9-4bf2-9f85-d72890ef7e1a-combined-ca-bundle\") pod \"d20989a7-98b9-4bf2-9f85-d72890ef7e1a\" (UID: \"d20989a7-98b9-4bf2-9f85-d72890ef7e1a\") " Oct 05 21:11:42 crc kubenswrapper[4754]: I1005 21:11:42.057688 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5rfdn\" (UniqueName: \"kubernetes.io/projected/d20989a7-98b9-4bf2-9f85-d72890ef7e1a-kube-api-access-5rfdn\") pod \"d20989a7-98b9-4bf2-9f85-d72890ef7e1a\" (UID: \"d20989a7-98b9-4bf2-9f85-d72890ef7e1a\") " Oct 05 21:11:42 crc kubenswrapper[4754]: I1005 21:11:42.057759 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d20989a7-98b9-4bf2-9f85-d72890ef7e1a-db-sync-config-data\") pod \"d20989a7-98b9-4bf2-9f85-d72890ef7e1a\" (UID: \"d20989a7-98b9-4bf2-9f85-d72890ef7e1a\") " Oct 05 21:11:42 crc kubenswrapper[4754]: I1005 21:11:42.057787 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d20989a7-98b9-4bf2-9f85-d72890ef7e1a-config-data\") pod \"d20989a7-98b9-4bf2-9f85-d72890ef7e1a\" (UID: \"d20989a7-98b9-4bf2-9f85-d72890ef7e1a\") " Oct 05 21:11:42 crc kubenswrapper[4754]: I1005 21:11:42.057955 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/94c73d9b-e224-4788-995d-be11dcee7b3c-logs\") pod \"placement-db-sync-xlw68\" (UID: \"94c73d9b-e224-4788-995d-be11dcee7b3c\") " pod="openstack/placement-db-sync-xlw68" Oct 05 21:11:42 crc kubenswrapper[4754]: I1005 21:11:42.057975 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/94c73d9b-e224-4788-995d-be11dcee7b3c-config-data\") pod \"placement-db-sync-xlw68\" (UID: \"94c73d9b-e224-4788-995d-be11dcee7b3c\") " pod="openstack/placement-db-sync-xlw68" Oct 05 21:11:42 crc kubenswrapper[4754]: I1005 21:11:42.057992 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/eea94483-ef0a-4518-8258-587ee1112608-logs\") pod \"horizon-689b67764f-n4l6h\" (UID: \"eea94483-ef0a-4518-8258-587ee1112608\") " pod="openstack/horizon-689b67764f-n4l6h" Oct 05 21:11:42 crc kubenswrapper[4754]: I1005 21:11:42.058008 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94c73d9b-e224-4788-995d-be11dcee7b3c-combined-ca-bundle\") pod \"placement-db-sync-xlw68\" (UID: \"94c73d9b-e224-4788-995d-be11dcee7b3c\") " pod="openstack/placement-db-sync-xlw68" Oct 05 21:11:42 crc kubenswrapper[4754]: I1005 21:11:42.058052 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/eea94483-ef0a-4518-8258-587ee1112608-horizon-secret-key\") pod \"horizon-689b67764f-n4l6h\" (UID: \"eea94483-ef0a-4518-8258-587ee1112608\") " pod="openstack/horizon-689b67764f-n4l6h" Oct 05 21:11:42 crc kubenswrapper[4754]: I1005 21:11:42.058073 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hrgbh\" (UniqueName: \"kubernetes.io/projected/eea94483-ef0a-4518-8258-587ee1112608-kube-api-access-hrgbh\") pod \"horizon-689b67764f-n4l6h\" (UID: \"eea94483-ef0a-4518-8258-587ee1112608\") " pod="openstack/horizon-689b67764f-n4l6h" Oct 05 21:11:42 crc kubenswrapper[4754]: I1005 21:11:42.058110 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/eea94483-ef0a-4518-8258-587ee1112608-config-data\") pod \"horizon-689b67764f-n4l6h\" (UID: \"eea94483-ef0a-4518-8258-587ee1112608\") " pod="openstack/horizon-689b67764f-n4l6h" Oct 05 21:11:42 crc kubenswrapper[4754]: I1005 21:11:42.058135 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v775j\" (UniqueName: \"kubernetes.io/projected/94c73d9b-e224-4788-995d-be11dcee7b3c-kube-api-access-v775j\") pod \"placement-db-sync-xlw68\" (UID: \"94c73d9b-e224-4788-995d-be11dcee7b3c\") " pod="openstack/placement-db-sync-xlw68" Oct 05 21:11:42 crc kubenswrapper[4754]: I1005 21:11:42.058176 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/eea94483-ef0a-4518-8258-587ee1112608-scripts\") pod \"horizon-689b67764f-n4l6h\" (UID: \"eea94483-ef0a-4518-8258-587ee1112608\") " pod="openstack/horizon-689b67764f-n4l6h" Oct 05 21:11:42 crc kubenswrapper[4754]: I1005 21:11:42.058200 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/94c73d9b-e224-4788-995d-be11dcee7b3c-scripts\") pod \"placement-db-sync-xlw68\" (UID: \"94c73d9b-e224-4788-995d-be11dcee7b3c\") " pod="openstack/placement-db-sync-xlw68" Oct 05 21:11:42 crc kubenswrapper[4754]: I1005 21:11:42.074271 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d20989a7-98b9-4bf2-9f85-d72890ef7e1a-kube-api-access-5rfdn" (OuterVolumeSpecName: "kube-api-access-5rfdn") pod "d20989a7-98b9-4bf2-9f85-d72890ef7e1a" (UID: "d20989a7-98b9-4bf2-9f85-d72890ef7e1a"). InnerVolumeSpecName "kube-api-access-5rfdn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:11:42 crc kubenswrapper[4754]: I1005 21:11:42.084136 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-xlw68"] Oct 05 21:11:42 crc kubenswrapper[4754]: I1005 21:11:42.085016 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d20989a7-98b9-4bf2-9f85-d72890ef7e1a-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "d20989a7-98b9-4bf2-9f85-d72890ef7e1a" (UID: "d20989a7-98b9-4bf2-9f85-d72890ef7e1a"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:11:42 crc kubenswrapper[4754]: I1005 21:11:42.142582 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-76fcf4b695-ptgrq"] Oct 05 21:11:42 crc kubenswrapper[4754]: E1005 21:11:42.143298 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d20989a7-98b9-4bf2-9f85-d72890ef7e1a" containerName="glance-db-sync" Oct 05 21:11:42 crc kubenswrapper[4754]: I1005 21:11:42.143311 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="d20989a7-98b9-4bf2-9f85-d72890ef7e1a" containerName="glance-db-sync" Oct 05 21:11:42 crc kubenswrapper[4754]: I1005 21:11:42.143527 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="d20989a7-98b9-4bf2-9f85-d72890ef7e1a" containerName="glance-db-sync" Oct 05 21:11:42 crc kubenswrapper[4754]: I1005 21:11:42.146779 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-76fcf4b695-ptgrq" Oct 05 21:11:42 crc kubenswrapper[4754]: I1005 21:11:42.166128 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/eea94483-ef0a-4518-8258-587ee1112608-scripts\") pod \"horizon-689b67764f-n4l6h\" (UID: \"eea94483-ef0a-4518-8258-587ee1112608\") " pod="openstack/horizon-689b67764f-n4l6h" Oct 05 21:11:42 crc kubenswrapper[4754]: I1005 21:11:42.166179 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/94c73d9b-e224-4788-995d-be11dcee7b3c-scripts\") pod \"placement-db-sync-xlw68\" (UID: \"94c73d9b-e224-4788-995d-be11dcee7b3c\") " pod="openstack/placement-db-sync-xlw68" Oct 05 21:11:42 crc kubenswrapper[4754]: I1005 21:11:42.166212 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/94c73d9b-e224-4788-995d-be11dcee7b3c-logs\") pod \"placement-db-sync-xlw68\" (UID: \"94c73d9b-e224-4788-995d-be11dcee7b3c\") " pod="openstack/placement-db-sync-xlw68" Oct 05 21:11:42 crc kubenswrapper[4754]: I1005 21:11:42.166233 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/94c73d9b-e224-4788-995d-be11dcee7b3c-config-data\") pod \"placement-db-sync-xlw68\" (UID: \"94c73d9b-e224-4788-995d-be11dcee7b3c\") " pod="openstack/placement-db-sync-xlw68" Oct 05 21:11:42 crc kubenswrapper[4754]: I1005 21:11:42.166250 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/eea94483-ef0a-4518-8258-587ee1112608-logs\") pod \"horizon-689b67764f-n4l6h\" (UID: \"eea94483-ef0a-4518-8258-587ee1112608\") " pod="openstack/horizon-689b67764f-n4l6h" Oct 05 21:11:42 crc kubenswrapper[4754]: I1005 21:11:42.166268 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94c73d9b-e224-4788-995d-be11dcee7b3c-combined-ca-bundle\") pod \"placement-db-sync-xlw68\" (UID: \"94c73d9b-e224-4788-995d-be11dcee7b3c\") " pod="openstack/placement-db-sync-xlw68" Oct 05 21:11:42 crc kubenswrapper[4754]: I1005 21:11:42.166324 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/eea94483-ef0a-4518-8258-587ee1112608-horizon-secret-key\") pod \"horizon-689b67764f-n4l6h\" (UID: \"eea94483-ef0a-4518-8258-587ee1112608\") " pod="openstack/horizon-689b67764f-n4l6h" Oct 05 21:11:42 crc kubenswrapper[4754]: I1005 21:11:42.166351 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hrgbh\" (UniqueName: \"kubernetes.io/projected/eea94483-ef0a-4518-8258-587ee1112608-kube-api-access-hrgbh\") pod \"horizon-689b67764f-n4l6h\" (UID: \"eea94483-ef0a-4518-8258-587ee1112608\") " pod="openstack/horizon-689b67764f-n4l6h" Oct 05 21:11:42 crc kubenswrapper[4754]: I1005 21:11:42.166393 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/eea94483-ef0a-4518-8258-587ee1112608-config-data\") pod \"horizon-689b67764f-n4l6h\" (UID: \"eea94483-ef0a-4518-8258-587ee1112608\") " pod="openstack/horizon-689b67764f-n4l6h" Oct 05 21:11:42 crc kubenswrapper[4754]: I1005 21:11:42.166422 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v775j\" (UniqueName: \"kubernetes.io/projected/94c73d9b-e224-4788-995d-be11dcee7b3c-kube-api-access-v775j\") pod \"placement-db-sync-xlw68\" (UID: \"94c73d9b-e224-4788-995d-be11dcee7b3c\") " pod="openstack/placement-db-sync-xlw68" Oct 05 21:11:42 crc kubenswrapper[4754]: I1005 21:11:42.166477 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5rfdn\" (UniqueName: \"kubernetes.io/projected/d20989a7-98b9-4bf2-9f85-d72890ef7e1a-kube-api-access-5rfdn\") on node \"crc\" DevicePath \"\"" Oct 05 21:11:42 crc kubenswrapper[4754]: I1005 21:11:42.166503 4754 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d20989a7-98b9-4bf2-9f85-d72890ef7e1a-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 21:11:42 crc kubenswrapper[4754]: I1005 21:11:42.167475 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/eea94483-ef0a-4518-8258-587ee1112608-scripts\") pod \"horizon-689b67764f-n4l6h\" (UID: \"eea94483-ef0a-4518-8258-587ee1112608\") " pod="openstack/horizon-689b67764f-n4l6h" Oct 05 21:11:42 crc kubenswrapper[4754]: I1005 21:11:42.174778 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/eea94483-ef0a-4518-8258-587ee1112608-logs\") pod \"horizon-689b67764f-n4l6h\" (UID: \"eea94483-ef0a-4518-8258-587ee1112608\") " pod="openstack/horizon-689b67764f-n4l6h" Oct 05 21:11:42 crc kubenswrapper[4754]: I1005 21:11:42.175024 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/94c73d9b-e224-4788-995d-be11dcee7b3c-logs\") pod \"placement-db-sync-xlw68\" (UID: \"94c73d9b-e224-4788-995d-be11dcee7b3c\") " pod="openstack/placement-db-sync-xlw68" Oct 05 21:11:42 crc kubenswrapper[4754]: I1005 21:11:42.184421 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/eea94483-ef0a-4518-8258-587ee1112608-config-data\") pod \"horizon-689b67764f-n4l6h\" (UID: \"eea94483-ef0a-4518-8258-587ee1112608\") " pod="openstack/horizon-689b67764f-n4l6h" Oct 05 21:11:42 crc kubenswrapper[4754]: I1005 21:11:42.184651 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dfec-account-create-qdgw4" event={"ID":"a454365f-a32c-427d-a2d1-cecff2501134","Type":"ContainerStarted","Data":"82a8703e58d341988c6e36bb9f54c30a61d28a9bb36326d380cdd46e8556c52a"} Oct 05 21:11:42 crc kubenswrapper[4754]: I1005 21:11:42.188865 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d20989a7-98b9-4bf2-9f85-d72890ef7e1a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d20989a7-98b9-4bf2-9f85-d72890ef7e1a" (UID: "d20989a7-98b9-4bf2-9f85-d72890ef7e1a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:11:42 crc kubenswrapper[4754]: I1005 21:11:42.225081 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-3385-account-create-62sw4" event={"ID":"b7ded8c6-2bd5-46ca-b36f-68ccbd92366d","Type":"ContainerStarted","Data":"d15b0f28fe749bd27bc1345b47de59518787999b5d97a6b04a0efc953b93aa6c"} Oct 05 21:11:42 crc kubenswrapper[4754]: I1005 21:11:42.240505 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-6d595fb7c-nnfgp"] Oct 05 21:11:42 crc kubenswrapper[4754]: I1005 21:11:42.241986 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6d595fb7c-nnfgp" Oct 05 21:11:42 crc kubenswrapper[4754]: I1005 21:11:42.251802 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/eea94483-ef0a-4518-8258-587ee1112608-horizon-secret-key\") pod \"horizon-689b67764f-n4l6h\" (UID: \"eea94483-ef0a-4518-8258-587ee1112608\") " pod="openstack/horizon-689b67764f-n4l6h" Oct 05 21:11:42 crc kubenswrapper[4754]: I1005 21:11:42.252003 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-dvzwp" event={"ID":"d20989a7-98b9-4bf2-9f85-d72890ef7e1a","Type":"ContainerDied","Data":"48aaa2d81fd2975c481ebe8744b0a6f2a38775ce5c0870569e3d451376dbae75"} Oct 05 21:11:42 crc kubenswrapper[4754]: I1005 21:11:42.252027 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="48aaa2d81fd2975c481ebe8744b0a6f2a38775ce5c0870569e3d451376dbae75" Oct 05 21:11:42 crc kubenswrapper[4754]: I1005 21:11:42.252076 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-dvzwp" Oct 05 21:11:42 crc kubenswrapper[4754]: I1005 21:11:42.253007 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/94c73d9b-e224-4788-995d-be11dcee7b3c-config-data\") pod \"placement-db-sync-xlw68\" (UID: \"94c73d9b-e224-4788-995d-be11dcee7b3c\") " pod="openstack/placement-db-sync-xlw68" Oct 05 21:11:42 crc kubenswrapper[4754]: I1005 21:11:42.253026 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94c73d9b-e224-4788-995d-be11dcee7b3c-combined-ca-bundle\") pod \"placement-db-sync-xlw68\" (UID: \"94c73d9b-e224-4788-995d-be11dcee7b3c\") " pod="openstack/placement-db-sync-xlw68" Oct 05 21:11:42 crc kubenswrapper[4754]: I1005 21:11:42.258904 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-76fcf4b695-ptgrq"] Oct 05 21:11:42 crc kubenswrapper[4754]: I1005 21:11:42.259421 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/94c73d9b-e224-4788-995d-be11dcee7b3c-scripts\") pod \"placement-db-sync-xlw68\" (UID: \"94c73d9b-e224-4788-995d-be11dcee7b3c\") " pod="openstack/placement-db-sync-xlw68" Oct 05 21:11:42 crc kubenswrapper[4754]: I1005 21:11:42.273279 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6d595fb7c-nnfgp"] Oct 05 21:11:42 crc kubenswrapper[4754]: I1005 21:11:42.274202 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-chwgt\" (UniqueName: \"kubernetes.io/projected/6795f4e9-da0e-4ff6-8447-3fe8668e4db1-kube-api-access-chwgt\") pod \"dnsmasq-dns-76fcf4b695-ptgrq\" (UID: \"6795f4e9-da0e-4ff6-8447-3fe8668e4db1\") " pod="openstack/dnsmasq-dns-76fcf4b695-ptgrq" Oct 05 21:11:42 crc kubenswrapper[4754]: I1005 21:11:42.274249 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6795f4e9-da0e-4ff6-8447-3fe8668e4db1-dns-swift-storage-0\") pod \"dnsmasq-dns-76fcf4b695-ptgrq\" (UID: \"6795f4e9-da0e-4ff6-8447-3fe8668e4db1\") " pod="openstack/dnsmasq-dns-76fcf4b695-ptgrq" Oct 05 21:11:42 crc kubenswrapper[4754]: I1005 21:11:42.274285 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6795f4e9-da0e-4ff6-8447-3fe8668e4db1-ovsdbserver-nb\") pod \"dnsmasq-dns-76fcf4b695-ptgrq\" (UID: \"6795f4e9-da0e-4ff6-8447-3fe8668e4db1\") " pod="openstack/dnsmasq-dns-76fcf4b695-ptgrq" Oct 05 21:11:42 crc kubenswrapper[4754]: I1005 21:11:42.274302 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6795f4e9-da0e-4ff6-8447-3fe8668e4db1-dns-svc\") pod \"dnsmasq-dns-76fcf4b695-ptgrq\" (UID: \"6795f4e9-da0e-4ff6-8447-3fe8668e4db1\") " pod="openstack/dnsmasq-dns-76fcf4b695-ptgrq" Oct 05 21:11:42 crc kubenswrapper[4754]: I1005 21:11:42.274320 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6795f4e9-da0e-4ff6-8447-3fe8668e4db1-ovsdbserver-sb\") pod \"dnsmasq-dns-76fcf4b695-ptgrq\" (UID: \"6795f4e9-da0e-4ff6-8447-3fe8668e4db1\") " pod="openstack/dnsmasq-dns-76fcf4b695-ptgrq" Oct 05 21:11:42 crc kubenswrapper[4754]: I1005 21:11:42.274338 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6795f4e9-da0e-4ff6-8447-3fe8668e4db1-config\") pod \"dnsmasq-dns-76fcf4b695-ptgrq\" (UID: \"6795f4e9-da0e-4ff6-8447-3fe8668e4db1\") " pod="openstack/dnsmasq-dns-76fcf4b695-ptgrq" Oct 05 21:11:42 crc kubenswrapper[4754]: I1005 21:11:42.274375 4754 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d20989a7-98b9-4bf2-9f85-d72890ef7e1a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 21:11:43 crc kubenswrapper[4754]: I1005 21:11:42.309434 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v775j\" (UniqueName: \"kubernetes.io/projected/94c73d9b-e224-4788-995d-be11dcee7b3c-kube-api-access-v775j\") pod \"placement-db-sync-xlw68\" (UID: \"94c73d9b-e224-4788-995d-be11dcee7b3c\") " pod="openstack/placement-db-sync-xlw68" Oct 05 21:11:43 crc kubenswrapper[4754]: I1005 21:11:42.317321 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hrgbh\" (UniqueName: \"kubernetes.io/projected/eea94483-ef0a-4518-8258-587ee1112608-kube-api-access-hrgbh\") pod \"horizon-689b67764f-n4l6h\" (UID: \"eea94483-ef0a-4518-8258-587ee1112608\") " pod="openstack/horizon-689b67764f-n4l6h" Oct 05 21:11:43 crc kubenswrapper[4754]: I1005 21:11:42.329943 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-689b67764f-n4l6h" Oct 05 21:11:43 crc kubenswrapper[4754]: I1005 21:11:42.342263 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 05 21:11:43 crc kubenswrapper[4754]: I1005 21:11:42.380125 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-xlw68" Oct 05 21:11:43 crc kubenswrapper[4754]: I1005 21:11:42.380862 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 05 21:11:43 crc kubenswrapper[4754]: I1005 21:11:42.394023 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 05 21:11:43 crc kubenswrapper[4754]: I1005 21:11:42.394878 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 05 21:11:43 crc kubenswrapper[4754]: I1005 21:11:42.396640 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6795f4e9-da0e-4ff6-8447-3fe8668e4db1-dns-swift-storage-0\") pod \"dnsmasq-dns-76fcf4b695-ptgrq\" (UID: \"6795f4e9-da0e-4ff6-8447-3fe8668e4db1\") " pod="openstack/dnsmasq-dns-76fcf4b695-ptgrq" Oct 05 21:11:43 crc kubenswrapper[4754]: I1005 21:11:42.396683 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6f47b134-d55b-454d-8084-7676eb153105-scripts\") pod \"horizon-6d595fb7c-nnfgp\" (UID: \"6f47b134-d55b-454d-8084-7676eb153105\") " pod="openstack/horizon-6d595fb7c-nnfgp" Oct 05 21:11:43 crc kubenswrapper[4754]: I1005 21:11:42.396741 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6795f4e9-da0e-4ff6-8447-3fe8668e4db1-ovsdbserver-nb\") pod \"dnsmasq-dns-76fcf4b695-ptgrq\" (UID: \"6795f4e9-da0e-4ff6-8447-3fe8668e4db1\") " pod="openstack/dnsmasq-dns-76fcf4b695-ptgrq" Oct 05 21:11:43 crc kubenswrapper[4754]: I1005 21:11:42.396762 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6795f4e9-da0e-4ff6-8447-3fe8668e4db1-dns-svc\") pod \"dnsmasq-dns-76fcf4b695-ptgrq\" (UID: \"6795f4e9-da0e-4ff6-8447-3fe8668e4db1\") " pod="openstack/dnsmasq-dns-76fcf4b695-ptgrq" Oct 05 21:11:43 crc kubenswrapper[4754]: I1005 21:11:42.396784 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6795f4e9-da0e-4ff6-8447-3fe8668e4db1-ovsdbserver-sb\") pod \"dnsmasq-dns-76fcf4b695-ptgrq\" (UID: \"6795f4e9-da0e-4ff6-8447-3fe8668e4db1\") " pod="openstack/dnsmasq-dns-76fcf4b695-ptgrq" Oct 05 21:11:43 crc kubenswrapper[4754]: I1005 21:11:42.396804 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6795f4e9-da0e-4ff6-8447-3fe8668e4db1-config\") pod \"dnsmasq-dns-76fcf4b695-ptgrq\" (UID: \"6795f4e9-da0e-4ff6-8447-3fe8668e4db1\") " pod="openstack/dnsmasq-dns-76fcf4b695-ptgrq" Oct 05 21:11:43 crc kubenswrapper[4754]: I1005 21:11:42.396828 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/6f47b134-d55b-454d-8084-7676eb153105-horizon-secret-key\") pod \"horizon-6d595fb7c-nnfgp\" (UID: \"6f47b134-d55b-454d-8084-7676eb153105\") " pod="openstack/horizon-6d595fb7c-nnfgp" Oct 05 21:11:43 crc kubenswrapper[4754]: I1005 21:11:42.396853 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6f47b134-d55b-454d-8084-7676eb153105-config-data\") pod \"horizon-6d595fb7c-nnfgp\" (UID: \"6f47b134-d55b-454d-8084-7676eb153105\") " pod="openstack/horizon-6d595fb7c-nnfgp" Oct 05 21:11:43 crc kubenswrapper[4754]: I1005 21:11:42.396962 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-chwgt\" (UniqueName: \"kubernetes.io/projected/6795f4e9-da0e-4ff6-8447-3fe8668e4db1-kube-api-access-chwgt\") pod \"dnsmasq-dns-76fcf4b695-ptgrq\" (UID: \"6795f4e9-da0e-4ff6-8447-3fe8668e4db1\") " pod="openstack/dnsmasq-dns-76fcf4b695-ptgrq" Oct 05 21:11:43 crc kubenswrapper[4754]: I1005 21:11:42.396986 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6f47b134-d55b-454d-8084-7676eb153105-logs\") pod \"horizon-6d595fb7c-nnfgp\" (UID: \"6f47b134-d55b-454d-8084-7676eb153105\") " pod="openstack/horizon-6d595fb7c-nnfgp" Oct 05 21:11:43 crc kubenswrapper[4754]: I1005 21:11:42.397004 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n2bwx\" (UniqueName: \"kubernetes.io/projected/6f47b134-d55b-454d-8084-7676eb153105-kube-api-access-n2bwx\") pod \"horizon-6d595fb7c-nnfgp\" (UID: \"6f47b134-d55b-454d-8084-7676eb153105\") " pod="openstack/horizon-6d595fb7c-nnfgp" Oct 05 21:11:43 crc kubenswrapper[4754]: I1005 21:11:42.398183 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6795f4e9-da0e-4ff6-8447-3fe8668e4db1-dns-swift-storage-0\") pod \"dnsmasq-dns-76fcf4b695-ptgrq\" (UID: \"6795f4e9-da0e-4ff6-8447-3fe8668e4db1\") " pod="openstack/dnsmasq-dns-76fcf4b695-ptgrq" Oct 05 21:11:43 crc kubenswrapper[4754]: I1005 21:11:42.399116 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6795f4e9-da0e-4ff6-8447-3fe8668e4db1-ovsdbserver-sb\") pod \"dnsmasq-dns-76fcf4b695-ptgrq\" (UID: \"6795f4e9-da0e-4ff6-8447-3fe8668e4db1\") " pod="openstack/dnsmasq-dns-76fcf4b695-ptgrq" Oct 05 21:11:43 crc kubenswrapper[4754]: I1005 21:11:42.399478 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6795f4e9-da0e-4ff6-8447-3fe8668e4db1-dns-svc\") pod \"dnsmasq-dns-76fcf4b695-ptgrq\" (UID: \"6795f4e9-da0e-4ff6-8447-3fe8668e4db1\") " pod="openstack/dnsmasq-dns-76fcf4b695-ptgrq" Oct 05 21:11:43 crc kubenswrapper[4754]: I1005 21:11:42.400077 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6795f4e9-da0e-4ff6-8447-3fe8668e4db1-config\") pod \"dnsmasq-dns-76fcf4b695-ptgrq\" (UID: \"6795f4e9-da0e-4ff6-8447-3fe8668e4db1\") " pod="openstack/dnsmasq-dns-76fcf4b695-ptgrq" Oct 05 21:11:43 crc kubenswrapper[4754]: I1005 21:11:42.409763 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6795f4e9-da0e-4ff6-8447-3fe8668e4db1-ovsdbserver-nb\") pod \"dnsmasq-dns-76fcf4b695-ptgrq\" (UID: \"6795f4e9-da0e-4ff6-8447-3fe8668e4db1\") " pod="openstack/dnsmasq-dns-76fcf4b695-ptgrq" Oct 05 21:11:43 crc kubenswrapper[4754]: I1005 21:11:42.446618 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 05 21:11:43 crc kubenswrapper[4754]: I1005 21:11:42.482592 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-chwgt\" (UniqueName: \"kubernetes.io/projected/6795f4e9-da0e-4ff6-8447-3fe8668e4db1-kube-api-access-chwgt\") pod \"dnsmasq-dns-76fcf4b695-ptgrq\" (UID: \"6795f4e9-da0e-4ff6-8447-3fe8668e4db1\") " pod="openstack/dnsmasq-dns-76fcf4b695-ptgrq" Oct 05 21:11:43 crc kubenswrapper[4754]: I1005 21:11:42.501274 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6f47b134-d55b-454d-8084-7676eb153105-logs\") pod \"horizon-6d595fb7c-nnfgp\" (UID: \"6f47b134-d55b-454d-8084-7676eb153105\") " pod="openstack/horizon-6d595fb7c-nnfgp" Oct 05 21:11:43 crc kubenswrapper[4754]: I1005 21:11:42.509936 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n2bwx\" (UniqueName: \"kubernetes.io/projected/6f47b134-d55b-454d-8084-7676eb153105-kube-api-access-n2bwx\") pod \"horizon-6d595fb7c-nnfgp\" (UID: \"6f47b134-d55b-454d-8084-7676eb153105\") " pod="openstack/horizon-6d595fb7c-nnfgp" Oct 05 21:11:43 crc kubenswrapper[4754]: I1005 21:11:42.509999 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a10eb04-c1a7-447a-9400-87b39b1befdf-config-data\") pod \"ceilometer-0\" (UID: \"0a10eb04-c1a7-447a-9400-87b39b1befdf\") " pod="openstack/ceilometer-0" Oct 05 21:11:43 crc kubenswrapper[4754]: I1005 21:11:42.510059 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0a10eb04-c1a7-447a-9400-87b39b1befdf-scripts\") pod \"ceilometer-0\" (UID: \"0a10eb04-c1a7-447a-9400-87b39b1befdf\") " pod="openstack/ceilometer-0" Oct 05 21:11:43 crc kubenswrapper[4754]: I1005 21:11:42.510136 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6f47b134-d55b-454d-8084-7676eb153105-scripts\") pod \"horizon-6d595fb7c-nnfgp\" (UID: \"6f47b134-d55b-454d-8084-7676eb153105\") " pod="openstack/horizon-6d595fb7c-nnfgp" Oct 05 21:11:43 crc kubenswrapper[4754]: I1005 21:11:42.510436 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0a10eb04-c1a7-447a-9400-87b39b1befdf-run-httpd\") pod \"ceilometer-0\" (UID: \"0a10eb04-c1a7-447a-9400-87b39b1befdf\") " pod="openstack/ceilometer-0" Oct 05 21:11:43 crc kubenswrapper[4754]: I1005 21:11:42.510514 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/6f47b134-d55b-454d-8084-7676eb153105-horizon-secret-key\") pod \"horizon-6d595fb7c-nnfgp\" (UID: \"6f47b134-d55b-454d-8084-7676eb153105\") " pod="openstack/horizon-6d595fb7c-nnfgp" Oct 05 21:11:43 crc kubenswrapper[4754]: I1005 21:11:42.510547 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a10eb04-c1a7-447a-9400-87b39b1befdf-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0a10eb04-c1a7-447a-9400-87b39b1befdf\") " pod="openstack/ceilometer-0" Oct 05 21:11:43 crc kubenswrapper[4754]: I1005 21:11:42.510576 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6f47b134-d55b-454d-8084-7676eb153105-config-data\") pod \"horizon-6d595fb7c-nnfgp\" (UID: \"6f47b134-d55b-454d-8084-7676eb153105\") " pod="openstack/horizon-6d595fb7c-nnfgp" Oct 05 21:11:43 crc kubenswrapper[4754]: I1005 21:11:42.510688 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l7k9j\" (UniqueName: \"kubernetes.io/projected/0a10eb04-c1a7-447a-9400-87b39b1befdf-kube-api-access-l7k9j\") pod \"ceilometer-0\" (UID: \"0a10eb04-c1a7-447a-9400-87b39b1befdf\") " pod="openstack/ceilometer-0" Oct 05 21:11:43 crc kubenswrapper[4754]: I1005 21:11:42.510769 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0a10eb04-c1a7-447a-9400-87b39b1befdf-log-httpd\") pod \"ceilometer-0\" (UID: \"0a10eb04-c1a7-447a-9400-87b39b1befdf\") " pod="openstack/ceilometer-0" Oct 05 21:11:43 crc kubenswrapper[4754]: I1005 21:11:42.510859 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0a10eb04-c1a7-447a-9400-87b39b1befdf-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0a10eb04-c1a7-447a-9400-87b39b1befdf\") " pod="openstack/ceilometer-0" Oct 05 21:11:43 crc kubenswrapper[4754]: I1005 21:11:42.508186 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6f47b134-d55b-454d-8084-7676eb153105-logs\") pod \"horizon-6d595fb7c-nnfgp\" (UID: \"6f47b134-d55b-454d-8084-7676eb153105\") " pod="openstack/horizon-6d595fb7c-nnfgp" Oct 05 21:11:43 crc kubenswrapper[4754]: I1005 21:11:42.512235 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6f47b134-d55b-454d-8084-7676eb153105-scripts\") pod \"horizon-6d595fb7c-nnfgp\" (UID: \"6f47b134-d55b-454d-8084-7676eb153105\") " pod="openstack/horizon-6d595fb7c-nnfgp" Oct 05 21:11:43 crc kubenswrapper[4754]: I1005 21:11:42.530093 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6f47b134-d55b-454d-8084-7676eb153105-config-data\") pod \"horizon-6d595fb7c-nnfgp\" (UID: \"6f47b134-d55b-454d-8084-7676eb153105\") " pod="openstack/horizon-6d595fb7c-nnfgp" Oct 05 21:11:43 crc kubenswrapper[4754]: I1005 21:11:42.530732 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-76fcf4b695-ptgrq" Oct 05 21:11:43 crc kubenswrapper[4754]: I1005 21:11:42.566259 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n2bwx\" (UniqueName: \"kubernetes.io/projected/6f47b134-d55b-454d-8084-7676eb153105-kube-api-access-n2bwx\") pod \"horizon-6d595fb7c-nnfgp\" (UID: \"6f47b134-d55b-454d-8084-7676eb153105\") " pod="openstack/horizon-6d595fb7c-nnfgp" Oct 05 21:11:43 crc kubenswrapper[4754]: I1005 21:11:42.601559 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/6f47b134-d55b-454d-8084-7676eb153105-horizon-secret-key\") pod \"horizon-6d595fb7c-nnfgp\" (UID: \"6f47b134-d55b-454d-8084-7676eb153105\") " pod="openstack/horizon-6d595fb7c-nnfgp" Oct 05 21:11:43 crc kubenswrapper[4754]: I1005 21:11:42.609156 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6d595fb7c-nnfgp" Oct 05 21:11:43 crc kubenswrapper[4754]: I1005 21:11:42.612277 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d20989a7-98b9-4bf2-9f85-d72890ef7e1a-config-data" (OuterVolumeSpecName: "config-data") pod "d20989a7-98b9-4bf2-9f85-d72890ef7e1a" (UID: "d20989a7-98b9-4bf2-9f85-d72890ef7e1a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:11:43 crc kubenswrapper[4754]: I1005 21:11:42.613338 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0a10eb04-c1a7-447a-9400-87b39b1befdf-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0a10eb04-c1a7-447a-9400-87b39b1befdf\") " pod="openstack/ceilometer-0" Oct 05 21:11:43 crc kubenswrapper[4754]: I1005 21:11:42.613394 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a10eb04-c1a7-447a-9400-87b39b1befdf-config-data\") pod \"ceilometer-0\" (UID: \"0a10eb04-c1a7-447a-9400-87b39b1befdf\") " pod="openstack/ceilometer-0" Oct 05 21:11:43 crc kubenswrapper[4754]: I1005 21:11:42.613419 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0a10eb04-c1a7-447a-9400-87b39b1befdf-scripts\") pod \"ceilometer-0\" (UID: \"0a10eb04-c1a7-447a-9400-87b39b1befdf\") " pod="openstack/ceilometer-0" Oct 05 21:11:43 crc kubenswrapper[4754]: I1005 21:11:42.613464 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0a10eb04-c1a7-447a-9400-87b39b1befdf-run-httpd\") pod \"ceilometer-0\" (UID: \"0a10eb04-c1a7-447a-9400-87b39b1befdf\") " pod="openstack/ceilometer-0" Oct 05 21:11:43 crc kubenswrapper[4754]: I1005 21:11:42.613506 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a10eb04-c1a7-447a-9400-87b39b1befdf-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0a10eb04-c1a7-447a-9400-87b39b1befdf\") " pod="openstack/ceilometer-0" Oct 05 21:11:43 crc kubenswrapper[4754]: I1005 21:11:42.613548 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l7k9j\" (UniqueName: \"kubernetes.io/projected/0a10eb04-c1a7-447a-9400-87b39b1befdf-kube-api-access-l7k9j\") pod \"ceilometer-0\" (UID: \"0a10eb04-c1a7-447a-9400-87b39b1befdf\") " pod="openstack/ceilometer-0" Oct 05 21:11:43 crc kubenswrapper[4754]: I1005 21:11:42.613576 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0a10eb04-c1a7-447a-9400-87b39b1befdf-log-httpd\") pod \"ceilometer-0\" (UID: \"0a10eb04-c1a7-447a-9400-87b39b1befdf\") " pod="openstack/ceilometer-0" Oct 05 21:11:43 crc kubenswrapper[4754]: I1005 21:11:42.613623 4754 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d20989a7-98b9-4bf2-9f85-d72890ef7e1a-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 21:11:43 crc kubenswrapper[4754]: I1005 21:11:42.613975 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0a10eb04-c1a7-447a-9400-87b39b1befdf-log-httpd\") pod \"ceilometer-0\" (UID: \"0a10eb04-c1a7-447a-9400-87b39b1befdf\") " pod="openstack/ceilometer-0" Oct 05 21:11:43 crc kubenswrapper[4754]: I1005 21:11:42.626529 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0a10eb04-c1a7-447a-9400-87b39b1befdf-run-httpd\") pod \"ceilometer-0\" (UID: \"0a10eb04-c1a7-447a-9400-87b39b1befdf\") " pod="openstack/ceilometer-0" Oct 05 21:11:43 crc kubenswrapper[4754]: I1005 21:11:42.632165 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0a10eb04-c1a7-447a-9400-87b39b1befdf-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0a10eb04-c1a7-447a-9400-87b39b1befdf\") " pod="openstack/ceilometer-0" Oct 05 21:11:43 crc kubenswrapper[4754]: I1005 21:11:42.658439 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a10eb04-c1a7-447a-9400-87b39b1befdf-config-data\") pod \"ceilometer-0\" (UID: \"0a10eb04-c1a7-447a-9400-87b39b1befdf\") " pod="openstack/ceilometer-0" Oct 05 21:11:43 crc kubenswrapper[4754]: I1005 21:11:42.660036 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a10eb04-c1a7-447a-9400-87b39b1befdf-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0a10eb04-c1a7-447a-9400-87b39b1befdf\") " pod="openstack/ceilometer-0" Oct 05 21:11:43 crc kubenswrapper[4754]: I1005 21:11:42.669252 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0a10eb04-c1a7-447a-9400-87b39b1befdf-scripts\") pod \"ceilometer-0\" (UID: \"0a10eb04-c1a7-447a-9400-87b39b1befdf\") " pod="openstack/ceilometer-0" Oct 05 21:11:43 crc kubenswrapper[4754]: I1005 21:11:42.724412 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l7k9j\" (UniqueName: \"kubernetes.io/projected/0a10eb04-c1a7-447a-9400-87b39b1befdf-kube-api-access-l7k9j\") pod \"ceilometer-0\" (UID: \"0a10eb04-c1a7-447a-9400-87b39b1befdf\") " pod="openstack/ceilometer-0" Oct 05 21:11:43 crc kubenswrapper[4754]: I1005 21:11:42.765666 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 05 21:11:43 crc kubenswrapper[4754]: I1005 21:11:43.112214 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-76fcf4b695-ptgrq"] Oct 05 21:11:43 crc kubenswrapper[4754]: I1005 21:11:43.187472 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-lsj2f"] Oct 05 21:11:43 crc kubenswrapper[4754]: I1005 21:11:43.192615 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8b5c85b87-lsj2f" Oct 05 21:11:43 crc kubenswrapper[4754]: I1005 21:11:43.200365 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-lsj2f"] Oct 05 21:11:43 crc kubenswrapper[4754]: I1005 21:11:43.314410 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-3385-account-create-62sw4" event={"ID":"b7ded8c6-2bd5-46ca-b36f-68ccbd92366d","Type":"ContainerStarted","Data":"41a7d5b3018f8f69bfd9504da8a203204ddacb0282e477631de02511aadb7446"} Oct 05 21:11:43 crc kubenswrapper[4754]: I1005 21:11:43.329211 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dfec-account-create-qdgw4" event={"ID":"a454365f-a32c-427d-a2d1-cecff2501134","Type":"ContainerStarted","Data":"50d818cfdab3f39a595898fe80cb53f7246ed57225037a25fb76b9787eeb1f6a"} Oct 05 21:11:43 crc kubenswrapper[4754]: I1005 21:11:43.348301 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f8012368-f6e9-43a6-aaa2-b228f05b79e3-dns-svc\") pod \"dnsmasq-dns-8b5c85b87-lsj2f\" (UID: \"f8012368-f6e9-43a6-aaa2-b228f05b79e3\") " pod="openstack/dnsmasq-dns-8b5c85b87-lsj2f" Oct 05 21:11:43 crc kubenswrapper[4754]: I1005 21:11:43.348409 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f8012368-f6e9-43a6-aaa2-b228f05b79e3-ovsdbserver-sb\") pod \"dnsmasq-dns-8b5c85b87-lsj2f\" (UID: \"f8012368-f6e9-43a6-aaa2-b228f05b79e3\") " pod="openstack/dnsmasq-dns-8b5c85b87-lsj2f" Oct 05 21:11:43 crc kubenswrapper[4754]: I1005 21:11:43.348551 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f8012368-f6e9-43a6-aaa2-b228f05b79e3-dns-swift-storage-0\") pod \"dnsmasq-dns-8b5c85b87-lsj2f\" (UID: \"f8012368-f6e9-43a6-aaa2-b228f05b79e3\") " pod="openstack/dnsmasq-dns-8b5c85b87-lsj2f" Oct 05 21:11:43 crc kubenswrapper[4754]: I1005 21:11:43.348579 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s5hkv\" (UniqueName: \"kubernetes.io/projected/f8012368-f6e9-43a6-aaa2-b228f05b79e3-kube-api-access-s5hkv\") pod \"dnsmasq-dns-8b5c85b87-lsj2f\" (UID: \"f8012368-f6e9-43a6-aaa2-b228f05b79e3\") " pod="openstack/dnsmasq-dns-8b5c85b87-lsj2f" Oct 05 21:11:43 crc kubenswrapper[4754]: I1005 21:11:43.348620 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f8012368-f6e9-43a6-aaa2-b228f05b79e3-ovsdbserver-nb\") pod \"dnsmasq-dns-8b5c85b87-lsj2f\" (UID: \"f8012368-f6e9-43a6-aaa2-b228f05b79e3\") " pod="openstack/dnsmasq-dns-8b5c85b87-lsj2f" Oct 05 21:11:43 crc kubenswrapper[4754]: I1005 21:11:43.348648 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f8012368-f6e9-43a6-aaa2-b228f05b79e3-config\") pod \"dnsmasq-dns-8b5c85b87-lsj2f\" (UID: \"f8012368-f6e9-43a6-aaa2-b228f05b79e3\") " pod="openstack/dnsmasq-dns-8b5c85b87-lsj2f" Oct 05 21:11:43 crc kubenswrapper[4754]: I1005 21:11:43.418640 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-dfec-account-create-qdgw4" podStartSLOduration=3.418621211 podStartE2EDuration="3.418621211s" podCreationTimestamp="2025-10-05 21:11:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 21:11:43.408617265 +0000 UTC m=+1027.312735975" watchObservedRunningTime="2025-10-05 21:11:43.418621211 +0000 UTC m=+1027.322739921" Oct 05 21:11:43 crc kubenswrapper[4754]: I1005 21:11:43.450355 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f8012368-f6e9-43a6-aaa2-b228f05b79e3-ovsdbserver-sb\") pod \"dnsmasq-dns-8b5c85b87-lsj2f\" (UID: \"f8012368-f6e9-43a6-aaa2-b228f05b79e3\") " pod="openstack/dnsmasq-dns-8b5c85b87-lsj2f" Oct 05 21:11:43 crc kubenswrapper[4754]: I1005 21:11:43.450460 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f8012368-f6e9-43a6-aaa2-b228f05b79e3-dns-swift-storage-0\") pod \"dnsmasq-dns-8b5c85b87-lsj2f\" (UID: \"f8012368-f6e9-43a6-aaa2-b228f05b79e3\") " pod="openstack/dnsmasq-dns-8b5c85b87-lsj2f" Oct 05 21:11:43 crc kubenswrapper[4754]: I1005 21:11:43.450481 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s5hkv\" (UniqueName: \"kubernetes.io/projected/f8012368-f6e9-43a6-aaa2-b228f05b79e3-kube-api-access-s5hkv\") pod \"dnsmasq-dns-8b5c85b87-lsj2f\" (UID: \"f8012368-f6e9-43a6-aaa2-b228f05b79e3\") " pod="openstack/dnsmasq-dns-8b5c85b87-lsj2f" Oct 05 21:11:43 crc kubenswrapper[4754]: I1005 21:11:43.450521 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f8012368-f6e9-43a6-aaa2-b228f05b79e3-ovsdbserver-nb\") pod \"dnsmasq-dns-8b5c85b87-lsj2f\" (UID: \"f8012368-f6e9-43a6-aaa2-b228f05b79e3\") " pod="openstack/dnsmasq-dns-8b5c85b87-lsj2f" Oct 05 21:11:43 crc kubenswrapper[4754]: I1005 21:11:43.450544 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f8012368-f6e9-43a6-aaa2-b228f05b79e3-config\") pod \"dnsmasq-dns-8b5c85b87-lsj2f\" (UID: \"f8012368-f6e9-43a6-aaa2-b228f05b79e3\") " pod="openstack/dnsmasq-dns-8b5c85b87-lsj2f" Oct 05 21:11:43 crc kubenswrapper[4754]: I1005 21:11:43.450631 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f8012368-f6e9-43a6-aaa2-b228f05b79e3-dns-svc\") pod \"dnsmasq-dns-8b5c85b87-lsj2f\" (UID: \"f8012368-f6e9-43a6-aaa2-b228f05b79e3\") " pod="openstack/dnsmasq-dns-8b5c85b87-lsj2f" Oct 05 21:11:43 crc kubenswrapper[4754]: I1005 21:11:43.451640 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f8012368-f6e9-43a6-aaa2-b228f05b79e3-dns-svc\") pod \"dnsmasq-dns-8b5c85b87-lsj2f\" (UID: \"f8012368-f6e9-43a6-aaa2-b228f05b79e3\") " pod="openstack/dnsmasq-dns-8b5c85b87-lsj2f" Oct 05 21:11:43 crc kubenswrapper[4754]: I1005 21:11:43.451828 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f8012368-f6e9-43a6-aaa2-b228f05b79e3-dns-swift-storage-0\") pod \"dnsmasq-dns-8b5c85b87-lsj2f\" (UID: \"f8012368-f6e9-43a6-aaa2-b228f05b79e3\") " pod="openstack/dnsmasq-dns-8b5c85b87-lsj2f" Oct 05 21:11:43 crc kubenswrapper[4754]: I1005 21:11:43.451939 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f8012368-f6e9-43a6-aaa2-b228f05b79e3-ovsdbserver-sb\") pod \"dnsmasq-dns-8b5c85b87-lsj2f\" (UID: \"f8012368-f6e9-43a6-aaa2-b228f05b79e3\") " pod="openstack/dnsmasq-dns-8b5c85b87-lsj2f" Oct 05 21:11:43 crc kubenswrapper[4754]: I1005 21:11:43.452419 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f8012368-f6e9-43a6-aaa2-b228f05b79e3-config\") pod \"dnsmasq-dns-8b5c85b87-lsj2f\" (UID: \"f8012368-f6e9-43a6-aaa2-b228f05b79e3\") " pod="openstack/dnsmasq-dns-8b5c85b87-lsj2f" Oct 05 21:11:43 crc kubenswrapper[4754]: I1005 21:11:43.452688 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f8012368-f6e9-43a6-aaa2-b228f05b79e3-ovsdbserver-nb\") pod \"dnsmasq-dns-8b5c85b87-lsj2f\" (UID: \"f8012368-f6e9-43a6-aaa2-b228f05b79e3\") " pod="openstack/dnsmasq-dns-8b5c85b87-lsj2f" Oct 05 21:11:43 crc kubenswrapper[4754]: I1005 21:11:43.477401 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s5hkv\" (UniqueName: \"kubernetes.io/projected/f8012368-f6e9-43a6-aaa2-b228f05b79e3-kube-api-access-s5hkv\") pod \"dnsmasq-dns-8b5c85b87-lsj2f\" (UID: \"f8012368-f6e9-43a6-aaa2-b228f05b79e3\") " pod="openstack/dnsmasq-dns-8b5c85b87-lsj2f" Oct 05 21:11:43 crc kubenswrapper[4754]: I1005 21:11:43.528665 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8b5c85b87-lsj2f" Oct 05 21:11:43 crc kubenswrapper[4754]: I1005 21:11:43.983947 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 05 21:11:43 crc kubenswrapper[4754]: I1005 21:11:43.985809 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 05 21:11:43 crc kubenswrapper[4754]: I1005 21:11:43.988859 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Oct 05 21:11:43 crc kubenswrapper[4754]: I1005 21:11:43.989521 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-gfwbt" Oct 05 21:11:43 crc kubenswrapper[4754]: I1005 21:11:43.989985 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 05 21:11:44 crc kubenswrapper[4754]: I1005 21:11:44.006725 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 05 21:11:44 crc kubenswrapper[4754]: I1005 21:11:44.060190 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-xlw68"] Oct 05 21:11:44 crc kubenswrapper[4754]: I1005 21:11:44.075916 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ce6c227-efce-459e-be29-9276329560ea-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"4ce6c227-efce-459e-be29-9276329560ea\") " pod="openstack/glance-default-external-api-0" Oct 05 21:11:44 crc kubenswrapper[4754]: I1005 21:11:44.076398 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4ce6c227-efce-459e-be29-9276329560ea-logs\") pod \"glance-default-external-api-0\" (UID: \"4ce6c227-efce-459e-be29-9276329560ea\") " pod="openstack/glance-default-external-api-0" Oct 05 21:11:44 crc kubenswrapper[4754]: I1005 21:11:44.076432 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4ce6c227-efce-459e-be29-9276329560ea-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"4ce6c227-efce-459e-be29-9276329560ea\") " pod="openstack/glance-default-external-api-0" Oct 05 21:11:44 crc kubenswrapper[4754]: I1005 21:11:44.076478 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2mfbh\" (UniqueName: \"kubernetes.io/projected/4ce6c227-efce-459e-be29-9276329560ea-kube-api-access-2mfbh\") pod \"glance-default-external-api-0\" (UID: \"4ce6c227-efce-459e-be29-9276329560ea\") " pod="openstack/glance-default-external-api-0" Oct 05 21:11:44 crc kubenswrapper[4754]: I1005 21:11:44.076566 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"4ce6c227-efce-459e-be29-9276329560ea\") " pod="openstack/glance-default-external-api-0" Oct 05 21:11:44 crc kubenswrapper[4754]: I1005 21:11:44.076602 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4ce6c227-efce-459e-be29-9276329560ea-scripts\") pod \"glance-default-external-api-0\" (UID: \"4ce6c227-efce-459e-be29-9276329560ea\") " pod="openstack/glance-default-external-api-0" Oct 05 21:11:44 crc kubenswrapper[4754]: I1005 21:11:44.076659 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ce6c227-efce-459e-be29-9276329560ea-config-data\") pod \"glance-default-external-api-0\" (UID: \"4ce6c227-efce-459e-be29-9276329560ea\") " pod="openstack/glance-default-external-api-0" Oct 05 21:11:44 crc kubenswrapper[4754]: I1005 21:11:44.116661 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-689b67764f-n4l6h"] Oct 05 21:11:44 crc kubenswrapper[4754]: W1005 21:11:44.137207 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podeea94483_ef0a_4518_8258_587ee1112608.slice/crio-8949f2011a09fc96d23dd205200b33c72f26c7eec65ceea8683e2441f452b32c WatchSource:0}: Error finding container 8949f2011a09fc96d23dd205200b33c72f26c7eec65ceea8683e2441f452b32c: Status 404 returned error can't find the container with id 8949f2011a09fc96d23dd205200b33c72f26c7eec65ceea8683e2441f452b32c Oct 05 21:11:44 crc kubenswrapper[4754]: I1005 21:11:44.169115 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55fff446b9-nvb2b"] Oct 05 21:11:44 crc kubenswrapper[4754]: I1005 21:11:44.179411 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ce6c227-efce-459e-be29-9276329560ea-config-data\") pod \"glance-default-external-api-0\" (UID: \"4ce6c227-efce-459e-be29-9276329560ea\") " pod="openstack/glance-default-external-api-0" Oct 05 21:11:44 crc kubenswrapper[4754]: I1005 21:11:44.179502 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ce6c227-efce-459e-be29-9276329560ea-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"4ce6c227-efce-459e-be29-9276329560ea\") " pod="openstack/glance-default-external-api-0" Oct 05 21:11:44 crc kubenswrapper[4754]: I1005 21:11:44.179553 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4ce6c227-efce-459e-be29-9276329560ea-logs\") pod \"glance-default-external-api-0\" (UID: \"4ce6c227-efce-459e-be29-9276329560ea\") " pod="openstack/glance-default-external-api-0" Oct 05 21:11:44 crc kubenswrapper[4754]: I1005 21:11:44.179581 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4ce6c227-efce-459e-be29-9276329560ea-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"4ce6c227-efce-459e-be29-9276329560ea\") " pod="openstack/glance-default-external-api-0" Oct 05 21:11:44 crc kubenswrapper[4754]: I1005 21:11:44.180033 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2mfbh\" (UniqueName: \"kubernetes.io/projected/4ce6c227-efce-459e-be29-9276329560ea-kube-api-access-2mfbh\") pod \"glance-default-external-api-0\" (UID: \"4ce6c227-efce-459e-be29-9276329560ea\") " pod="openstack/glance-default-external-api-0" Oct 05 21:11:44 crc kubenswrapper[4754]: I1005 21:11:44.180083 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"4ce6c227-efce-459e-be29-9276329560ea\") " pod="openstack/glance-default-external-api-0" Oct 05 21:11:44 crc kubenswrapper[4754]: I1005 21:11:44.180115 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4ce6c227-efce-459e-be29-9276329560ea-scripts\") pod \"glance-default-external-api-0\" (UID: \"4ce6c227-efce-459e-be29-9276329560ea\") " pod="openstack/glance-default-external-api-0" Oct 05 21:11:44 crc kubenswrapper[4754]: I1005 21:11:44.186366 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-76fcf4b695-ptgrq"] Oct 05 21:11:44 crc kubenswrapper[4754]: I1005 21:11:44.187632 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4ce6c227-efce-459e-be29-9276329560ea-logs\") pod \"glance-default-external-api-0\" (UID: \"4ce6c227-efce-459e-be29-9276329560ea\") " pod="openstack/glance-default-external-api-0" Oct 05 21:11:44 crc kubenswrapper[4754]: I1005 21:11:44.187684 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4ce6c227-efce-459e-be29-9276329560ea-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"4ce6c227-efce-459e-be29-9276329560ea\") " pod="openstack/glance-default-external-api-0" Oct 05 21:11:44 crc kubenswrapper[4754]: I1005 21:11:44.190682 4754 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"4ce6c227-efce-459e-be29-9276329560ea\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/glance-default-external-api-0" Oct 05 21:11:44 crc kubenswrapper[4754]: I1005 21:11:44.201737 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4ce6c227-efce-459e-be29-9276329560ea-scripts\") pod \"glance-default-external-api-0\" (UID: \"4ce6c227-efce-459e-be29-9276329560ea\") " pod="openstack/glance-default-external-api-0" Oct 05 21:11:44 crc kubenswrapper[4754]: I1005 21:11:44.202084 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ce6c227-efce-459e-be29-9276329560ea-config-data\") pod \"glance-default-external-api-0\" (UID: \"4ce6c227-efce-459e-be29-9276329560ea\") " pod="openstack/glance-default-external-api-0" Oct 05 21:11:44 crc kubenswrapper[4754]: I1005 21:11:44.205639 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ce6c227-efce-459e-be29-9276329560ea-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"4ce6c227-efce-459e-be29-9276329560ea\") " pod="openstack/glance-default-external-api-0" Oct 05 21:11:44 crc kubenswrapper[4754]: I1005 21:11:44.206917 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-fvkbb"] Oct 05 21:11:44 crc kubenswrapper[4754]: I1005 21:11:44.211645 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2mfbh\" (UniqueName: \"kubernetes.io/projected/4ce6c227-efce-459e-be29-9276329560ea-kube-api-access-2mfbh\") pod \"glance-default-external-api-0\" (UID: \"4ce6c227-efce-459e-be29-9276329560ea\") " pod="openstack/glance-default-external-api-0" Oct 05 21:11:44 crc kubenswrapper[4754]: I1005 21:11:44.232774 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 05 21:11:44 crc kubenswrapper[4754]: W1005 21:11:44.246028 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf8012368_f6e9_43a6_aaa2_b228f05b79e3.slice/crio-0d4460efe4cb5abf2ebf32ff09520226a7c4892dcc794f577c86d187e041c4c1 WatchSource:0}: Error finding container 0d4460efe4cb5abf2ebf32ff09520226a7c4892dcc794f577c86d187e041c4c1: Status 404 returned error can't find the container with id 0d4460efe4cb5abf2ebf32ff09520226a7c4892dcc794f577c86d187e041c4c1 Oct 05 21:11:44 crc kubenswrapper[4754]: I1005 21:11:44.252393 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6d595fb7c-nnfgp"] Oct 05 21:11:44 crc kubenswrapper[4754]: I1005 21:11:44.272447 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-lsj2f"] Oct 05 21:11:44 crc kubenswrapper[4754]: I1005 21:11:44.304303 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"4ce6c227-efce-459e-be29-9276329560ea\") " pod="openstack/glance-default-external-api-0" Oct 05 21:11:44 crc kubenswrapper[4754]: I1005 21:11:44.361869 4754 generic.go:334] "Generic (PLEG): container finished" podID="a454365f-a32c-427d-a2d1-cecff2501134" containerID="50d818cfdab3f39a595898fe80cb53f7246ed57225037a25fb76b9787eeb1f6a" exitCode=0 Oct 05 21:11:44 crc kubenswrapper[4754]: I1005 21:11:44.361931 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dfec-account-create-qdgw4" event={"ID":"a454365f-a32c-427d-a2d1-cecff2501134","Type":"ContainerDied","Data":"50d818cfdab3f39a595898fe80cb53f7246ed57225037a25fb76b9787eeb1f6a"} Oct 05 21:11:44 crc kubenswrapper[4754]: I1005 21:11:44.364636 4754 generic.go:334] "Generic (PLEG): container finished" podID="b7ded8c6-2bd5-46ca-b36f-68ccbd92366d" containerID="41a7d5b3018f8f69bfd9504da8a203204ddacb0282e477631de02511aadb7446" exitCode=0 Oct 05 21:11:44 crc kubenswrapper[4754]: I1005 21:11:44.364672 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-3385-account-create-62sw4" event={"ID":"b7ded8c6-2bd5-46ca-b36f-68ccbd92366d","Type":"ContainerDied","Data":"41a7d5b3018f8f69bfd9504da8a203204ddacb0282e477631de02511aadb7446"} Oct 05 21:11:44 crc kubenswrapper[4754]: I1005 21:11:44.369697 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 05 21:11:44 crc kubenswrapper[4754]: I1005 21:11:44.371414 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 05 21:11:44 crc kubenswrapper[4754]: I1005 21:11:44.380145 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 05 21:11:44 crc kubenswrapper[4754]: I1005 21:11:44.389019 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0a10eb04-c1a7-447a-9400-87b39b1befdf","Type":"ContainerStarted","Data":"656f92069277b501ccc8771c65b67ba371b0110cf87ab9be74b51aa3eacf78b6"} Oct 05 21:11:44 crc kubenswrapper[4754]: I1005 21:11:44.391145 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6d595fb7c-nnfgp" event={"ID":"6f47b134-d55b-454d-8084-7676eb153105","Type":"ContainerStarted","Data":"72ca63d3e5da90693a056fdb5f8a671f27ace95c9927768e78c24b40b7af13d3"} Oct 05 21:11:44 crc kubenswrapper[4754]: I1005 21:11:44.392902 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-fvkbb" event={"ID":"cbf1bb4f-3997-4581-ba76-64b10ba3f686","Type":"ContainerStarted","Data":"f681d7e7b627b29f4e56fe9964d7e02052e49864eb30180b1eeab9cdd31695dd"} Oct 05 21:11:44 crc kubenswrapper[4754]: I1005 21:11:44.398015 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-689b67764f-n4l6h" event={"ID":"eea94483-ef0a-4518-8258-587ee1112608","Type":"ContainerStarted","Data":"8949f2011a09fc96d23dd205200b33c72f26c7eec65ceea8683e2441f452b32c"} Oct 05 21:11:44 crc kubenswrapper[4754]: I1005 21:11:44.399555 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76fcf4b695-ptgrq" event={"ID":"6795f4e9-da0e-4ff6-8447-3fe8668e4db1","Type":"ContainerStarted","Data":"2348383e4f763ace20c52b449987101e708e39be7e7f4fdca87cae727406bc27"} Oct 05 21:11:44 crc kubenswrapper[4754]: I1005 21:11:44.400285 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55fff446b9-nvb2b" event={"ID":"4752f446-876a-4d46-8e36-b4b03143c9db","Type":"ContainerStarted","Data":"98d6355ea1f6745e1d508b56e29695fb5d34b19bb4ed647b0449a4135daa7b72"} Oct 05 21:11:44 crc kubenswrapper[4754]: I1005 21:11:44.401180 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-xlw68" event={"ID":"94c73d9b-e224-4788-995d-be11dcee7b3c","Type":"ContainerStarted","Data":"33fe11d04fe5e87797e874bbaac02c954c408a1820e8306bb640a1ce1251b93e"} Oct 05 21:11:44 crc kubenswrapper[4754]: I1005 21:11:44.401980 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-lsj2f" event={"ID":"f8012368-f6e9-43a6-aaa2-b228f05b79e3","Type":"ContainerStarted","Data":"0d4460efe4cb5abf2ebf32ff09520226a7c4892dcc794f577c86d187e041c4c1"} Oct 05 21:11:44 crc kubenswrapper[4754]: I1005 21:11:44.457458 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 05 21:11:44 crc kubenswrapper[4754]: I1005 21:11:44.484163 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/81cb813f-8f33-44b6-8e43-62ee912608ce-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"81cb813f-8f33-44b6-8e43-62ee912608ce\") " pod="openstack/glance-default-internal-api-0" Oct 05 21:11:44 crc kubenswrapper[4754]: I1005 21:11:44.484295 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81cb813f-8f33-44b6-8e43-62ee912608ce-config-data\") pod \"glance-default-internal-api-0\" (UID: \"81cb813f-8f33-44b6-8e43-62ee912608ce\") " pod="openstack/glance-default-internal-api-0" Oct 05 21:11:44 crc kubenswrapper[4754]: I1005 21:11:44.484358 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h2jht\" (UniqueName: \"kubernetes.io/projected/81cb813f-8f33-44b6-8e43-62ee912608ce-kube-api-access-h2jht\") pod \"glance-default-internal-api-0\" (UID: \"81cb813f-8f33-44b6-8e43-62ee912608ce\") " pod="openstack/glance-default-internal-api-0" Oct 05 21:11:44 crc kubenswrapper[4754]: I1005 21:11:44.484389 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81cb813f-8f33-44b6-8e43-62ee912608ce-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"81cb813f-8f33-44b6-8e43-62ee912608ce\") " pod="openstack/glance-default-internal-api-0" Oct 05 21:11:44 crc kubenswrapper[4754]: I1005 21:11:44.484436 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-internal-api-0\" (UID: \"81cb813f-8f33-44b6-8e43-62ee912608ce\") " pod="openstack/glance-default-internal-api-0" Oct 05 21:11:44 crc kubenswrapper[4754]: I1005 21:11:44.484483 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/81cb813f-8f33-44b6-8e43-62ee912608ce-scripts\") pod \"glance-default-internal-api-0\" (UID: \"81cb813f-8f33-44b6-8e43-62ee912608ce\") " pod="openstack/glance-default-internal-api-0" Oct 05 21:11:44 crc kubenswrapper[4754]: I1005 21:11:44.484538 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/81cb813f-8f33-44b6-8e43-62ee912608ce-logs\") pod \"glance-default-internal-api-0\" (UID: \"81cb813f-8f33-44b6-8e43-62ee912608ce\") " pod="openstack/glance-default-internal-api-0" Oct 05 21:11:44 crc kubenswrapper[4754]: I1005 21:11:44.590072 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/81cb813f-8f33-44b6-8e43-62ee912608ce-scripts\") pod \"glance-default-internal-api-0\" (UID: \"81cb813f-8f33-44b6-8e43-62ee912608ce\") " pod="openstack/glance-default-internal-api-0" Oct 05 21:11:44 crc kubenswrapper[4754]: I1005 21:11:44.590689 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/81cb813f-8f33-44b6-8e43-62ee912608ce-logs\") pod \"glance-default-internal-api-0\" (UID: \"81cb813f-8f33-44b6-8e43-62ee912608ce\") " pod="openstack/glance-default-internal-api-0" Oct 05 21:11:44 crc kubenswrapper[4754]: I1005 21:11:44.590726 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/81cb813f-8f33-44b6-8e43-62ee912608ce-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"81cb813f-8f33-44b6-8e43-62ee912608ce\") " pod="openstack/glance-default-internal-api-0" Oct 05 21:11:44 crc kubenswrapper[4754]: I1005 21:11:44.590765 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81cb813f-8f33-44b6-8e43-62ee912608ce-config-data\") pod \"glance-default-internal-api-0\" (UID: \"81cb813f-8f33-44b6-8e43-62ee912608ce\") " pod="openstack/glance-default-internal-api-0" Oct 05 21:11:44 crc kubenswrapper[4754]: I1005 21:11:44.590809 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h2jht\" (UniqueName: \"kubernetes.io/projected/81cb813f-8f33-44b6-8e43-62ee912608ce-kube-api-access-h2jht\") pod \"glance-default-internal-api-0\" (UID: \"81cb813f-8f33-44b6-8e43-62ee912608ce\") " pod="openstack/glance-default-internal-api-0" Oct 05 21:11:44 crc kubenswrapper[4754]: I1005 21:11:44.590845 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81cb813f-8f33-44b6-8e43-62ee912608ce-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"81cb813f-8f33-44b6-8e43-62ee912608ce\") " pod="openstack/glance-default-internal-api-0" Oct 05 21:11:44 crc kubenswrapper[4754]: I1005 21:11:44.590875 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-internal-api-0\" (UID: \"81cb813f-8f33-44b6-8e43-62ee912608ce\") " pod="openstack/glance-default-internal-api-0" Oct 05 21:11:44 crc kubenswrapper[4754]: I1005 21:11:44.591304 4754 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-internal-api-0\" (UID: \"81cb813f-8f33-44b6-8e43-62ee912608ce\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/glance-default-internal-api-0" Oct 05 21:11:44 crc kubenswrapper[4754]: I1005 21:11:44.594209 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/81cb813f-8f33-44b6-8e43-62ee912608ce-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"81cb813f-8f33-44b6-8e43-62ee912608ce\") " pod="openstack/glance-default-internal-api-0" Oct 05 21:11:44 crc kubenswrapper[4754]: I1005 21:11:44.608604 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 05 21:11:44 crc kubenswrapper[4754]: I1005 21:11:44.613242 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/81cb813f-8f33-44b6-8e43-62ee912608ce-scripts\") pod \"glance-default-internal-api-0\" (UID: \"81cb813f-8f33-44b6-8e43-62ee912608ce\") " pod="openstack/glance-default-internal-api-0" Oct 05 21:11:44 crc kubenswrapper[4754]: I1005 21:11:44.653098 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/81cb813f-8f33-44b6-8e43-62ee912608ce-logs\") pod \"glance-default-internal-api-0\" (UID: \"81cb813f-8f33-44b6-8e43-62ee912608ce\") " pod="openstack/glance-default-internal-api-0" Oct 05 21:11:44 crc kubenswrapper[4754]: I1005 21:11:44.653930 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81cb813f-8f33-44b6-8e43-62ee912608ce-config-data\") pod \"glance-default-internal-api-0\" (UID: \"81cb813f-8f33-44b6-8e43-62ee912608ce\") " pod="openstack/glance-default-internal-api-0" Oct 05 21:11:44 crc kubenswrapper[4754]: I1005 21:11:44.654546 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81cb813f-8f33-44b6-8e43-62ee912608ce-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"81cb813f-8f33-44b6-8e43-62ee912608ce\") " pod="openstack/glance-default-internal-api-0" Oct 05 21:11:44 crc kubenswrapper[4754]: I1005 21:11:44.676955 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h2jht\" (UniqueName: \"kubernetes.io/projected/81cb813f-8f33-44b6-8e43-62ee912608ce-kube-api-access-h2jht\") pod \"glance-default-internal-api-0\" (UID: \"81cb813f-8f33-44b6-8e43-62ee912608ce\") " pod="openstack/glance-default-internal-api-0" Oct 05 21:11:44 crc kubenswrapper[4754]: I1005 21:11:44.680793 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-internal-api-0\" (UID: \"81cb813f-8f33-44b6-8e43-62ee912608ce\") " pod="openstack/glance-default-internal-api-0" Oct 05 21:11:44 crc kubenswrapper[4754]: I1005 21:11:44.845747 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 05 21:11:44 crc kubenswrapper[4754]: I1005 21:11:44.944251 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 05 21:11:44 crc kubenswrapper[4754]: I1005 21:11:44.991752 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-3385-account-create-62sw4" Oct 05 21:11:45 crc kubenswrapper[4754]: I1005 21:11:45.005580 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-689b67764f-n4l6h"] Oct 05 21:11:45 crc kubenswrapper[4754]: I1005 21:11:45.041259 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 05 21:11:45 crc kubenswrapper[4754]: I1005 21:11:45.057733 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-5f98f885c7-9ckfb"] Oct 05 21:11:45 crc kubenswrapper[4754]: E1005 21:11:45.058193 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7ded8c6-2bd5-46ca-b36f-68ccbd92366d" containerName="mariadb-account-create" Oct 05 21:11:45 crc kubenswrapper[4754]: I1005 21:11:45.058211 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7ded8c6-2bd5-46ca-b36f-68ccbd92366d" containerName="mariadb-account-create" Oct 05 21:11:45 crc kubenswrapper[4754]: I1005 21:11:45.058384 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="b7ded8c6-2bd5-46ca-b36f-68ccbd92366d" containerName="mariadb-account-create" Oct 05 21:11:45 crc kubenswrapper[4754]: I1005 21:11:45.066406 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5f98f885c7-9ckfb" Oct 05 21:11:45 crc kubenswrapper[4754]: I1005 21:11:45.092802 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5f98f885c7-9ckfb"] Oct 05 21:11:45 crc kubenswrapper[4754]: I1005 21:11:45.119524 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 05 21:11:45 crc kubenswrapper[4754]: I1005 21:11:45.125841 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c6nvg\" (UniqueName: \"kubernetes.io/projected/b7ded8c6-2bd5-46ca-b36f-68ccbd92366d-kube-api-access-c6nvg\") pod \"b7ded8c6-2bd5-46ca-b36f-68ccbd92366d\" (UID: \"b7ded8c6-2bd5-46ca-b36f-68ccbd92366d\") " Oct 05 21:11:45 crc kubenswrapper[4754]: I1005 21:11:45.139191 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b7ded8c6-2bd5-46ca-b36f-68ccbd92366d-kube-api-access-c6nvg" (OuterVolumeSpecName: "kube-api-access-c6nvg") pod "b7ded8c6-2bd5-46ca-b36f-68ccbd92366d" (UID: "b7ded8c6-2bd5-46ca-b36f-68ccbd92366d"). InnerVolumeSpecName "kube-api-access-c6nvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:11:45 crc kubenswrapper[4754]: I1005 21:11:45.230462 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/146ce51d-4eb0-4aae-98ad-08bc8a8f9728-scripts\") pod \"horizon-5f98f885c7-9ckfb\" (UID: \"146ce51d-4eb0-4aae-98ad-08bc8a8f9728\") " pod="openstack/horizon-5f98f885c7-9ckfb" Oct 05 21:11:45 crc kubenswrapper[4754]: I1005 21:11:45.230855 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/146ce51d-4eb0-4aae-98ad-08bc8a8f9728-horizon-secret-key\") pod \"horizon-5f98f885c7-9ckfb\" (UID: \"146ce51d-4eb0-4aae-98ad-08bc8a8f9728\") " pod="openstack/horizon-5f98f885c7-9ckfb" Oct 05 21:11:45 crc kubenswrapper[4754]: I1005 21:11:45.230890 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/146ce51d-4eb0-4aae-98ad-08bc8a8f9728-config-data\") pod \"horizon-5f98f885c7-9ckfb\" (UID: \"146ce51d-4eb0-4aae-98ad-08bc8a8f9728\") " pod="openstack/horizon-5f98f885c7-9ckfb" Oct 05 21:11:45 crc kubenswrapper[4754]: I1005 21:11:45.230917 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/146ce51d-4eb0-4aae-98ad-08bc8a8f9728-logs\") pod \"horizon-5f98f885c7-9ckfb\" (UID: \"146ce51d-4eb0-4aae-98ad-08bc8a8f9728\") " pod="openstack/horizon-5f98f885c7-9ckfb" Oct 05 21:11:45 crc kubenswrapper[4754]: I1005 21:11:45.230951 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4gzct\" (UniqueName: \"kubernetes.io/projected/146ce51d-4eb0-4aae-98ad-08bc8a8f9728-kube-api-access-4gzct\") pod \"horizon-5f98f885c7-9ckfb\" (UID: \"146ce51d-4eb0-4aae-98ad-08bc8a8f9728\") " pod="openstack/horizon-5f98f885c7-9ckfb" Oct 05 21:11:45 crc kubenswrapper[4754]: I1005 21:11:45.231007 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c6nvg\" (UniqueName: \"kubernetes.io/projected/b7ded8c6-2bd5-46ca-b36f-68ccbd92366d-kube-api-access-c6nvg\") on node \"crc\" DevicePath \"\"" Oct 05 21:11:45 crc kubenswrapper[4754]: I1005 21:11:45.333139 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/146ce51d-4eb0-4aae-98ad-08bc8a8f9728-horizon-secret-key\") pod \"horizon-5f98f885c7-9ckfb\" (UID: \"146ce51d-4eb0-4aae-98ad-08bc8a8f9728\") " pod="openstack/horizon-5f98f885c7-9ckfb" Oct 05 21:11:45 crc kubenswrapper[4754]: I1005 21:11:45.333221 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/146ce51d-4eb0-4aae-98ad-08bc8a8f9728-config-data\") pod \"horizon-5f98f885c7-9ckfb\" (UID: \"146ce51d-4eb0-4aae-98ad-08bc8a8f9728\") " pod="openstack/horizon-5f98f885c7-9ckfb" Oct 05 21:11:45 crc kubenswrapper[4754]: I1005 21:11:45.333252 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/146ce51d-4eb0-4aae-98ad-08bc8a8f9728-logs\") pod \"horizon-5f98f885c7-9ckfb\" (UID: \"146ce51d-4eb0-4aae-98ad-08bc8a8f9728\") " pod="openstack/horizon-5f98f885c7-9ckfb" Oct 05 21:11:45 crc kubenswrapper[4754]: I1005 21:11:45.333298 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4gzct\" (UniqueName: \"kubernetes.io/projected/146ce51d-4eb0-4aae-98ad-08bc8a8f9728-kube-api-access-4gzct\") pod \"horizon-5f98f885c7-9ckfb\" (UID: \"146ce51d-4eb0-4aae-98ad-08bc8a8f9728\") " pod="openstack/horizon-5f98f885c7-9ckfb" Oct 05 21:11:45 crc kubenswrapper[4754]: I1005 21:11:45.333358 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/146ce51d-4eb0-4aae-98ad-08bc8a8f9728-scripts\") pod \"horizon-5f98f885c7-9ckfb\" (UID: \"146ce51d-4eb0-4aae-98ad-08bc8a8f9728\") " pod="openstack/horizon-5f98f885c7-9ckfb" Oct 05 21:11:45 crc kubenswrapper[4754]: I1005 21:11:45.334524 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/146ce51d-4eb0-4aae-98ad-08bc8a8f9728-scripts\") pod \"horizon-5f98f885c7-9ckfb\" (UID: \"146ce51d-4eb0-4aae-98ad-08bc8a8f9728\") " pod="openstack/horizon-5f98f885c7-9ckfb" Oct 05 21:11:45 crc kubenswrapper[4754]: I1005 21:11:45.335437 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/146ce51d-4eb0-4aae-98ad-08bc8a8f9728-config-data\") pod \"horizon-5f98f885c7-9ckfb\" (UID: \"146ce51d-4eb0-4aae-98ad-08bc8a8f9728\") " pod="openstack/horizon-5f98f885c7-9ckfb" Oct 05 21:11:45 crc kubenswrapper[4754]: I1005 21:11:45.337072 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/146ce51d-4eb0-4aae-98ad-08bc8a8f9728-logs\") pod \"horizon-5f98f885c7-9ckfb\" (UID: \"146ce51d-4eb0-4aae-98ad-08bc8a8f9728\") " pod="openstack/horizon-5f98f885c7-9ckfb" Oct 05 21:11:45 crc kubenswrapper[4754]: I1005 21:11:45.344871 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/146ce51d-4eb0-4aae-98ad-08bc8a8f9728-horizon-secret-key\") pod \"horizon-5f98f885c7-9ckfb\" (UID: \"146ce51d-4eb0-4aae-98ad-08bc8a8f9728\") " pod="openstack/horizon-5f98f885c7-9ckfb" Oct 05 21:11:45 crc kubenswrapper[4754]: I1005 21:11:45.378896 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4gzct\" (UniqueName: \"kubernetes.io/projected/146ce51d-4eb0-4aae-98ad-08bc8a8f9728-kube-api-access-4gzct\") pod \"horizon-5f98f885c7-9ckfb\" (UID: \"146ce51d-4eb0-4aae-98ad-08bc8a8f9728\") " pod="openstack/horizon-5f98f885c7-9ckfb" Oct 05 21:11:45 crc kubenswrapper[4754]: I1005 21:11:45.428600 4754 generic.go:334] "Generic (PLEG): container finished" podID="f8012368-f6e9-43a6-aaa2-b228f05b79e3" containerID="540f320a309cdb6e6ec64a15c4d6c93e9e2b9eae2b0d9eef21cf099c1c26fd88" exitCode=0 Oct 05 21:11:45 crc kubenswrapper[4754]: I1005 21:11:45.428687 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-lsj2f" event={"ID":"f8012368-f6e9-43a6-aaa2-b228f05b79e3","Type":"ContainerDied","Data":"540f320a309cdb6e6ec64a15c4d6c93e9e2b9eae2b0d9eef21cf099c1c26fd88"} Oct 05 21:11:45 crc kubenswrapper[4754]: I1005 21:11:45.444920 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-fvkbb" event={"ID":"cbf1bb4f-3997-4581-ba76-64b10ba3f686","Type":"ContainerStarted","Data":"70f654feef08dc667d5d46a2f784bbef7d1df0ec788d6b6e2bcab20afb46ee6b"} Oct 05 21:11:45 crc kubenswrapper[4754]: I1005 21:11:45.454680 4754 generic.go:334] "Generic (PLEG): container finished" podID="6795f4e9-da0e-4ff6-8447-3fe8668e4db1" containerID="edd8ecfd83e0778ed71c1fc60c7e379dee445591ca059693b1dccb7ee1c2698c" exitCode=0 Oct 05 21:11:45 crc kubenswrapper[4754]: I1005 21:11:45.454748 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76fcf4b695-ptgrq" event={"ID":"6795f4e9-da0e-4ff6-8447-3fe8668e4db1","Type":"ContainerDied","Data":"edd8ecfd83e0778ed71c1fc60c7e379dee445591ca059693b1dccb7ee1c2698c"} Oct 05 21:11:45 crc kubenswrapper[4754]: I1005 21:11:45.500877 4754 generic.go:334] "Generic (PLEG): container finished" podID="4752f446-876a-4d46-8e36-b4b03143c9db" containerID="7311acd6fdac969f780b1e3eb6e47fa2be7c28e4a1a5d8341fd87edeaebbfcf8" exitCode=0 Oct 05 21:11:45 crc kubenswrapper[4754]: I1005 21:11:45.500956 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55fff446b9-nvb2b" event={"ID":"4752f446-876a-4d46-8e36-b4b03143c9db","Type":"ContainerDied","Data":"7311acd6fdac969f780b1e3eb6e47fa2be7c28e4a1a5d8341fd87edeaebbfcf8"} Oct 05 21:11:45 crc kubenswrapper[4754]: I1005 21:11:45.516730 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-fvkbb" podStartSLOduration=4.516713204 podStartE2EDuration="4.516713204s" podCreationTimestamp="2025-10-05 21:11:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 21:11:45.512856711 +0000 UTC m=+1029.416975421" watchObservedRunningTime="2025-10-05 21:11:45.516713204 +0000 UTC m=+1029.420831914" Oct 05 21:11:45 crc kubenswrapper[4754]: I1005 21:11:45.525573 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-3385-account-create-62sw4" Oct 05 21:11:45 crc kubenswrapper[4754]: I1005 21:11:45.525984 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-3385-account-create-62sw4" event={"ID":"b7ded8c6-2bd5-46ca-b36f-68ccbd92366d","Type":"ContainerDied","Data":"d15b0f28fe749bd27bc1345b47de59518787999b5d97a6b04a0efc953b93aa6c"} Oct 05 21:11:45 crc kubenswrapper[4754]: I1005 21:11:45.526412 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d15b0f28fe749bd27bc1345b47de59518787999b5d97a6b04a0efc953b93aa6c" Oct 05 21:11:45 crc kubenswrapper[4754]: I1005 21:11:45.531340 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5f98f885c7-9ckfb" Oct 05 21:11:45 crc kubenswrapper[4754]: I1005 21:11:45.842448 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 05 21:11:46 crc kubenswrapper[4754]: I1005 21:11:46.145097 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dfec-account-create-qdgw4" Oct 05 21:11:46 crc kubenswrapper[4754]: I1005 21:11:46.277989 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wkq8n\" (UniqueName: \"kubernetes.io/projected/a454365f-a32c-427d-a2d1-cecff2501134-kube-api-access-wkq8n\") pod \"a454365f-a32c-427d-a2d1-cecff2501134\" (UID: \"a454365f-a32c-427d-a2d1-cecff2501134\") " Oct 05 21:11:46 crc kubenswrapper[4754]: I1005 21:11:46.283158 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-76fcf4b695-ptgrq" Oct 05 21:11:46 crc kubenswrapper[4754]: I1005 21:11:46.286371 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a454365f-a32c-427d-a2d1-cecff2501134-kube-api-access-wkq8n" (OuterVolumeSpecName: "kube-api-access-wkq8n") pod "a454365f-a32c-427d-a2d1-cecff2501134" (UID: "a454365f-a32c-427d-a2d1-cecff2501134"). InnerVolumeSpecName "kube-api-access-wkq8n". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:11:46 crc kubenswrapper[4754]: I1005 21:11:46.382885 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6795f4e9-da0e-4ff6-8447-3fe8668e4db1-ovsdbserver-sb\") pod \"6795f4e9-da0e-4ff6-8447-3fe8668e4db1\" (UID: \"6795f4e9-da0e-4ff6-8447-3fe8668e4db1\") " Oct 05 21:11:46 crc kubenswrapper[4754]: I1005 21:11:46.382931 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-chwgt\" (UniqueName: \"kubernetes.io/projected/6795f4e9-da0e-4ff6-8447-3fe8668e4db1-kube-api-access-chwgt\") pod \"6795f4e9-da0e-4ff6-8447-3fe8668e4db1\" (UID: \"6795f4e9-da0e-4ff6-8447-3fe8668e4db1\") " Oct 05 21:11:46 crc kubenswrapper[4754]: I1005 21:11:46.383015 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6795f4e9-da0e-4ff6-8447-3fe8668e4db1-ovsdbserver-nb\") pod \"6795f4e9-da0e-4ff6-8447-3fe8668e4db1\" (UID: \"6795f4e9-da0e-4ff6-8447-3fe8668e4db1\") " Oct 05 21:11:46 crc kubenswrapper[4754]: I1005 21:11:46.383083 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6795f4e9-da0e-4ff6-8447-3fe8668e4db1-dns-svc\") pod \"6795f4e9-da0e-4ff6-8447-3fe8668e4db1\" (UID: \"6795f4e9-da0e-4ff6-8447-3fe8668e4db1\") " Oct 05 21:11:46 crc kubenswrapper[4754]: I1005 21:11:46.383162 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6795f4e9-da0e-4ff6-8447-3fe8668e4db1-config\") pod \"6795f4e9-da0e-4ff6-8447-3fe8668e4db1\" (UID: \"6795f4e9-da0e-4ff6-8447-3fe8668e4db1\") " Oct 05 21:11:46 crc kubenswrapper[4754]: I1005 21:11:46.383197 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6795f4e9-da0e-4ff6-8447-3fe8668e4db1-dns-swift-storage-0\") pod \"6795f4e9-da0e-4ff6-8447-3fe8668e4db1\" (UID: \"6795f4e9-da0e-4ff6-8447-3fe8668e4db1\") " Oct 05 21:11:46 crc kubenswrapper[4754]: I1005 21:11:46.383582 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wkq8n\" (UniqueName: \"kubernetes.io/projected/a454365f-a32c-427d-a2d1-cecff2501134-kube-api-access-wkq8n\") on node \"crc\" DevicePath \"\"" Oct 05 21:11:46 crc kubenswrapper[4754]: I1005 21:11:46.396207 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6795f4e9-da0e-4ff6-8447-3fe8668e4db1-kube-api-access-chwgt" (OuterVolumeSpecName: "kube-api-access-chwgt") pod "6795f4e9-da0e-4ff6-8447-3fe8668e4db1" (UID: "6795f4e9-da0e-4ff6-8447-3fe8668e4db1"). InnerVolumeSpecName "kube-api-access-chwgt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:11:46 crc kubenswrapper[4754]: I1005 21:11:46.440103 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6795f4e9-da0e-4ff6-8447-3fe8668e4db1-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "6795f4e9-da0e-4ff6-8447-3fe8668e4db1" (UID: "6795f4e9-da0e-4ff6-8447-3fe8668e4db1"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 21:11:46 crc kubenswrapper[4754]: I1005 21:11:46.445196 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6795f4e9-da0e-4ff6-8447-3fe8668e4db1-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "6795f4e9-da0e-4ff6-8447-3fe8668e4db1" (UID: "6795f4e9-da0e-4ff6-8447-3fe8668e4db1"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 21:11:46 crc kubenswrapper[4754]: I1005 21:11:46.461676 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6795f4e9-da0e-4ff6-8447-3fe8668e4db1-config" (OuterVolumeSpecName: "config") pod "6795f4e9-da0e-4ff6-8447-3fe8668e4db1" (UID: "6795f4e9-da0e-4ff6-8447-3fe8668e4db1"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 21:11:46 crc kubenswrapper[4754]: I1005 21:11:46.470902 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6795f4e9-da0e-4ff6-8447-3fe8668e4db1-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "6795f4e9-da0e-4ff6-8447-3fe8668e4db1" (UID: "6795f4e9-da0e-4ff6-8447-3fe8668e4db1"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 21:11:46 crc kubenswrapper[4754]: I1005 21:11:46.477643 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5f98f885c7-9ckfb"] Oct 05 21:11:46 crc kubenswrapper[4754]: W1005 21:11:46.477839 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod146ce51d_4eb0_4aae_98ad_08bc8a8f9728.slice/crio-cab648bee655ca280f477f5718ea20e53b388bab4b68ad37032e5beed7eed19f WatchSource:0}: Error finding container cab648bee655ca280f477f5718ea20e53b388bab4b68ad37032e5beed7eed19f: Status 404 returned error can't find the container with id cab648bee655ca280f477f5718ea20e53b388bab4b68ad37032e5beed7eed19f Oct 05 21:11:46 crc kubenswrapper[4754]: I1005 21:11:46.484107 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6795f4e9-da0e-4ff6-8447-3fe8668e4db1-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "6795f4e9-da0e-4ff6-8447-3fe8668e4db1" (UID: "6795f4e9-da0e-4ff6-8447-3fe8668e4db1"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 21:11:46 crc kubenswrapper[4754]: I1005 21:11:46.484444 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6795f4e9-da0e-4ff6-8447-3fe8668e4db1-dns-svc\") pod \"6795f4e9-da0e-4ff6-8447-3fe8668e4db1\" (UID: \"6795f4e9-da0e-4ff6-8447-3fe8668e4db1\") " Oct 05 21:11:46 crc kubenswrapper[4754]: I1005 21:11:46.484824 4754 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6795f4e9-da0e-4ff6-8447-3fe8668e4db1-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 05 21:11:46 crc kubenswrapper[4754]: I1005 21:11:46.484841 4754 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6795f4e9-da0e-4ff6-8447-3fe8668e4db1-config\") on node \"crc\" DevicePath \"\"" Oct 05 21:11:46 crc kubenswrapper[4754]: I1005 21:11:46.484850 4754 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6795f4e9-da0e-4ff6-8447-3fe8668e4db1-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 05 21:11:46 crc kubenswrapper[4754]: I1005 21:11:46.484860 4754 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6795f4e9-da0e-4ff6-8447-3fe8668e4db1-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 05 21:11:46 crc kubenswrapper[4754]: I1005 21:11:46.484870 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-chwgt\" (UniqueName: \"kubernetes.io/projected/6795f4e9-da0e-4ff6-8447-3fe8668e4db1-kube-api-access-chwgt\") on node \"crc\" DevicePath \"\"" Oct 05 21:11:46 crc kubenswrapper[4754]: W1005 21:11:46.484938 4754 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/6795f4e9-da0e-4ff6-8447-3fe8668e4db1/volumes/kubernetes.io~configmap/dns-svc Oct 05 21:11:46 crc kubenswrapper[4754]: I1005 21:11:46.484948 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6795f4e9-da0e-4ff6-8447-3fe8668e4db1-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "6795f4e9-da0e-4ff6-8447-3fe8668e4db1" (UID: "6795f4e9-da0e-4ff6-8447-3fe8668e4db1"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 21:11:46 crc kubenswrapper[4754]: I1005 21:11:46.558965 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 05 21:11:46 crc kubenswrapper[4754]: I1005 21:11:46.561446 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55fff446b9-nvb2b" Oct 05 21:11:46 crc kubenswrapper[4754]: I1005 21:11:46.561895 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55fff446b9-nvb2b" event={"ID":"4752f446-876a-4d46-8e36-b4b03143c9db","Type":"ContainerDied","Data":"98d6355ea1f6745e1d508b56e29695fb5d34b19bb4ed647b0449a4135daa7b72"} Oct 05 21:11:46 crc kubenswrapper[4754]: I1005 21:11:46.561933 4754 scope.go:117] "RemoveContainer" containerID="7311acd6fdac969f780b1e3eb6e47fa2be7c28e4a1a5d8341fd87edeaebbfcf8" Oct 05 21:11:46 crc kubenswrapper[4754]: W1005 21:11:46.567261 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4ce6c227_efce_459e_be29_9276329560ea.slice/crio-ae13d386ea4170e09acc7de99487262694163c237998f9629ec91b16b520f3db WatchSource:0}: Error finding container ae13d386ea4170e09acc7de99487262694163c237998f9629ec91b16b520f3db: Status 404 returned error can't find the container with id ae13d386ea4170e09acc7de99487262694163c237998f9629ec91b16b520f3db Oct 05 21:11:46 crc kubenswrapper[4754]: I1005 21:11:46.567509 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dfec-account-create-qdgw4" Oct 05 21:11:46 crc kubenswrapper[4754]: I1005 21:11:46.567548 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dfec-account-create-qdgw4" event={"ID":"a454365f-a32c-427d-a2d1-cecff2501134","Type":"ContainerDied","Data":"82a8703e58d341988c6e36bb9f54c30a61d28a9bb36326d380cdd46e8556c52a"} Oct 05 21:11:46 crc kubenswrapper[4754]: I1005 21:11:46.567606 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="82a8703e58d341988c6e36bb9f54c30a61d28a9bb36326d380cdd46e8556c52a" Oct 05 21:11:46 crc kubenswrapper[4754]: I1005 21:11:46.572000 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-lsj2f" event={"ID":"f8012368-f6e9-43a6-aaa2-b228f05b79e3","Type":"ContainerStarted","Data":"662c7f7244687f194dbaef3262108de7752198aac4491f90932c3b7f614ab2a3"} Oct 05 21:11:46 crc kubenswrapper[4754]: I1005 21:11:46.573245 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-8b5c85b87-lsj2f" Oct 05 21:11:46 crc kubenswrapper[4754]: I1005 21:11:46.583375 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"81cb813f-8f33-44b6-8e43-62ee912608ce","Type":"ContainerStarted","Data":"f9477fc2456936555beee0d94a22d56142582fc2dcb9e5ebd2fef64825c59b21"} Oct 05 21:11:46 crc kubenswrapper[4754]: I1005 21:11:46.585204 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5f98f885c7-9ckfb" event={"ID":"146ce51d-4eb0-4aae-98ad-08bc8a8f9728","Type":"ContainerStarted","Data":"cab648bee655ca280f477f5718ea20e53b388bab4b68ad37032e5beed7eed19f"} Oct 05 21:11:46 crc kubenswrapper[4754]: I1005 21:11:46.586018 4754 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6795f4e9-da0e-4ff6-8447-3fe8668e4db1-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 05 21:11:46 crc kubenswrapper[4754]: I1005 21:11:46.587657 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76fcf4b695-ptgrq" event={"ID":"6795f4e9-da0e-4ff6-8447-3fe8668e4db1","Type":"ContainerDied","Data":"2348383e4f763ace20c52b449987101e708e39be7e7f4fdca87cae727406bc27"} Oct 05 21:11:46 crc kubenswrapper[4754]: I1005 21:11:46.587773 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-76fcf4b695-ptgrq" Oct 05 21:11:46 crc kubenswrapper[4754]: I1005 21:11:46.615161 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-8b5c85b87-lsj2f" podStartSLOduration=3.615145401 podStartE2EDuration="3.615145401s" podCreationTimestamp="2025-10-05 21:11:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 21:11:46.599657858 +0000 UTC m=+1030.503776568" watchObservedRunningTime="2025-10-05 21:11:46.615145401 +0000 UTC m=+1030.519264111" Oct 05 21:11:46 crc kubenswrapper[4754]: I1005 21:11:46.686932 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4752f446-876a-4d46-8e36-b4b03143c9db-config\") pod \"4752f446-876a-4d46-8e36-b4b03143c9db\" (UID: \"4752f446-876a-4d46-8e36-b4b03143c9db\") " Oct 05 21:11:46 crc kubenswrapper[4754]: I1005 21:11:46.686993 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4752f446-876a-4d46-8e36-b4b03143c9db-dns-swift-storage-0\") pod \"4752f446-876a-4d46-8e36-b4b03143c9db\" (UID: \"4752f446-876a-4d46-8e36-b4b03143c9db\") " Oct 05 21:11:46 crc kubenswrapper[4754]: I1005 21:11:46.687030 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4752f446-876a-4d46-8e36-b4b03143c9db-ovsdbserver-sb\") pod \"4752f446-876a-4d46-8e36-b4b03143c9db\" (UID: \"4752f446-876a-4d46-8e36-b4b03143c9db\") " Oct 05 21:11:46 crc kubenswrapper[4754]: I1005 21:11:46.687158 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4752f446-876a-4d46-8e36-b4b03143c9db-ovsdbserver-nb\") pod \"4752f446-876a-4d46-8e36-b4b03143c9db\" (UID: \"4752f446-876a-4d46-8e36-b4b03143c9db\") " Oct 05 21:11:46 crc kubenswrapper[4754]: I1005 21:11:46.687197 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-95xqp\" (UniqueName: \"kubernetes.io/projected/4752f446-876a-4d46-8e36-b4b03143c9db-kube-api-access-95xqp\") pod \"4752f446-876a-4d46-8e36-b4b03143c9db\" (UID: \"4752f446-876a-4d46-8e36-b4b03143c9db\") " Oct 05 21:11:46 crc kubenswrapper[4754]: I1005 21:11:46.687218 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4752f446-876a-4d46-8e36-b4b03143c9db-dns-svc\") pod \"4752f446-876a-4d46-8e36-b4b03143c9db\" (UID: \"4752f446-876a-4d46-8e36-b4b03143c9db\") " Oct 05 21:11:46 crc kubenswrapper[4754]: I1005 21:11:46.702595 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-76fcf4b695-ptgrq"] Oct 05 21:11:46 crc kubenswrapper[4754]: I1005 21:11:46.703119 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4752f446-876a-4d46-8e36-b4b03143c9db-kube-api-access-95xqp" (OuterVolumeSpecName: "kube-api-access-95xqp") pod "4752f446-876a-4d46-8e36-b4b03143c9db" (UID: "4752f446-876a-4d46-8e36-b4b03143c9db"). InnerVolumeSpecName "kube-api-access-95xqp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:11:46 crc kubenswrapper[4754]: I1005 21:11:46.727879 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4752f446-876a-4d46-8e36-b4b03143c9db-config" (OuterVolumeSpecName: "config") pod "4752f446-876a-4d46-8e36-b4b03143c9db" (UID: "4752f446-876a-4d46-8e36-b4b03143c9db"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 21:11:46 crc kubenswrapper[4754]: I1005 21:11:46.742802 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-76fcf4b695-ptgrq"] Oct 05 21:11:46 crc kubenswrapper[4754]: I1005 21:11:46.765012 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4752f446-876a-4d46-8e36-b4b03143c9db-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "4752f446-876a-4d46-8e36-b4b03143c9db" (UID: "4752f446-876a-4d46-8e36-b4b03143c9db"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 21:11:46 crc kubenswrapper[4754]: I1005 21:11:46.768066 4754 scope.go:117] "RemoveContainer" containerID="edd8ecfd83e0778ed71c1fc60c7e379dee445591ca059693b1dccb7ee1c2698c" Oct 05 21:11:46 crc kubenswrapper[4754]: I1005 21:11:46.786359 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4752f446-876a-4d46-8e36-b4b03143c9db-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "4752f446-876a-4d46-8e36-b4b03143c9db" (UID: "4752f446-876a-4d46-8e36-b4b03143c9db"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 21:11:46 crc kubenswrapper[4754]: I1005 21:11:46.790783 4754 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4752f446-876a-4d46-8e36-b4b03143c9db-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 05 21:11:46 crc kubenswrapper[4754]: I1005 21:11:46.790808 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-95xqp\" (UniqueName: \"kubernetes.io/projected/4752f446-876a-4d46-8e36-b4b03143c9db-kube-api-access-95xqp\") on node \"crc\" DevicePath \"\"" Oct 05 21:11:46 crc kubenswrapper[4754]: I1005 21:11:46.790822 4754 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4752f446-876a-4d46-8e36-b4b03143c9db-config\") on node \"crc\" DevicePath \"\"" Oct 05 21:11:46 crc kubenswrapper[4754]: I1005 21:11:46.790833 4754 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4752f446-876a-4d46-8e36-b4b03143c9db-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 05 21:11:46 crc kubenswrapper[4754]: I1005 21:11:46.794256 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4752f446-876a-4d46-8e36-b4b03143c9db-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "4752f446-876a-4d46-8e36-b4b03143c9db" (UID: "4752f446-876a-4d46-8e36-b4b03143c9db"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 21:11:46 crc kubenswrapper[4754]: I1005 21:11:46.813075 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4752f446-876a-4d46-8e36-b4b03143c9db-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "4752f446-876a-4d46-8e36-b4b03143c9db" (UID: "4752f446-876a-4d46-8e36-b4b03143c9db"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 21:11:46 crc kubenswrapper[4754]: I1005 21:11:46.870127 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6795f4e9-da0e-4ff6-8447-3fe8668e4db1" path="/var/lib/kubelet/pods/6795f4e9-da0e-4ff6-8447-3fe8668e4db1/volumes" Oct 05 21:11:46 crc kubenswrapper[4754]: I1005 21:11:46.896560 4754 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4752f446-876a-4d46-8e36-b4b03143c9db-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 05 21:11:46 crc kubenswrapper[4754]: I1005 21:11:46.896588 4754 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4752f446-876a-4d46-8e36-b4b03143c9db-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 05 21:11:47 crc kubenswrapper[4754]: I1005 21:11:47.637270 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"81cb813f-8f33-44b6-8e43-62ee912608ce","Type":"ContainerStarted","Data":"05a2fef72ece381d89a223af19f4818d6d2b4308ccededf2975de60136b2855c"} Oct 05 21:11:47 crc kubenswrapper[4754]: I1005 21:11:47.650852 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55fff446b9-nvb2b" Oct 05 21:11:47 crc kubenswrapper[4754]: I1005 21:11:47.677349 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"4ce6c227-efce-459e-be29-9276329560ea","Type":"ContainerStarted","Data":"fadc39cb4613473d324e7ee89a7fa105b4a1c1e8c30a775530ef8b396c30ad78"} Oct 05 21:11:47 crc kubenswrapper[4754]: I1005 21:11:47.677387 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"4ce6c227-efce-459e-be29-9276329560ea","Type":"ContainerStarted","Data":"ae13d386ea4170e09acc7de99487262694163c237998f9629ec91b16b520f3db"} Oct 05 21:11:47 crc kubenswrapper[4754]: I1005 21:11:47.736458 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55fff446b9-nvb2b"] Oct 05 21:11:47 crc kubenswrapper[4754]: I1005 21:11:47.749454 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-55fff446b9-nvb2b"] Oct 05 21:11:48 crc kubenswrapper[4754]: I1005 21:11:48.701502 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"81cb813f-8f33-44b6-8e43-62ee912608ce","Type":"ContainerStarted","Data":"9d1b636f33f2a4ee25b0ee5a822723b84ba632b207a15b251d5d962b9f42bf61"} Oct 05 21:11:48 crc kubenswrapper[4754]: I1005 21:11:48.701933 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="81cb813f-8f33-44b6-8e43-62ee912608ce" containerName="glance-httpd" containerID="cri-o://9d1b636f33f2a4ee25b0ee5a822723b84ba632b207a15b251d5d962b9f42bf61" gracePeriod=30 Oct 05 21:11:48 crc kubenswrapper[4754]: I1005 21:11:48.702119 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="81cb813f-8f33-44b6-8e43-62ee912608ce" containerName="glance-log" containerID="cri-o://05a2fef72ece381d89a223af19f4818d6d2b4308ccededf2975de60136b2855c" gracePeriod=30 Oct 05 21:11:48 crc kubenswrapper[4754]: I1005 21:11:48.861105 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4752f446-876a-4d46-8e36-b4b03143c9db" path="/var/lib/kubelet/pods/4752f446-876a-4d46-8e36-b4b03143c9db/volumes" Oct 05 21:11:49 crc kubenswrapper[4754]: I1005 21:11:49.711588 4754 generic.go:334] "Generic (PLEG): container finished" podID="81cb813f-8f33-44b6-8e43-62ee912608ce" containerID="9d1b636f33f2a4ee25b0ee5a822723b84ba632b207a15b251d5d962b9f42bf61" exitCode=0 Oct 05 21:11:49 crc kubenswrapper[4754]: I1005 21:11:49.711897 4754 generic.go:334] "Generic (PLEG): container finished" podID="81cb813f-8f33-44b6-8e43-62ee912608ce" containerID="05a2fef72ece381d89a223af19f4818d6d2b4308ccededf2975de60136b2855c" exitCode=143 Oct 05 21:11:49 crc kubenswrapper[4754]: I1005 21:11:49.711630 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"81cb813f-8f33-44b6-8e43-62ee912608ce","Type":"ContainerDied","Data":"9d1b636f33f2a4ee25b0ee5a822723b84ba632b207a15b251d5d962b9f42bf61"} Oct 05 21:11:49 crc kubenswrapper[4754]: I1005 21:11:49.712029 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"81cb813f-8f33-44b6-8e43-62ee912608ce","Type":"ContainerDied","Data":"05a2fef72ece381d89a223af19f4818d6d2b4308ccededf2975de60136b2855c"} Oct 05 21:11:49 crc kubenswrapper[4754]: I1005 21:11:49.715050 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"4ce6c227-efce-459e-be29-9276329560ea","Type":"ContainerStarted","Data":"faffd538d0e452dc90ce161248c8b227666300ba17506311f3e003f5bdeea6e4"} Oct 05 21:11:49 crc kubenswrapper[4754]: I1005 21:11:49.715178 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="4ce6c227-efce-459e-be29-9276329560ea" containerName="glance-log" containerID="cri-o://fadc39cb4613473d324e7ee89a7fa105b4a1c1e8c30a775530ef8b396c30ad78" gracePeriod=30 Oct 05 21:11:49 crc kubenswrapper[4754]: I1005 21:11:49.715263 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="4ce6c227-efce-459e-be29-9276329560ea" containerName="glance-httpd" containerID="cri-o://faffd538d0e452dc90ce161248c8b227666300ba17506311f3e003f5bdeea6e4" gracePeriod=30 Oct 05 21:11:49 crc kubenswrapper[4754]: I1005 21:11:49.734074 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=6.734058833 podStartE2EDuration="6.734058833s" podCreationTimestamp="2025-10-05 21:11:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 21:11:48.761604251 +0000 UTC m=+1032.665722981" watchObservedRunningTime="2025-10-05 21:11:49.734058833 +0000 UTC m=+1033.638177543" Oct 05 21:11:49 crc kubenswrapper[4754]: I1005 21:11:49.740654 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=7.740636488 podStartE2EDuration="7.740636488s" podCreationTimestamp="2025-10-05 21:11:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 21:11:49.731895725 +0000 UTC m=+1033.636014435" watchObservedRunningTime="2025-10-05 21:11:49.740636488 +0000 UTC m=+1033.644755198" Oct 05 21:11:50 crc kubenswrapper[4754]: I1005 21:11:50.577865 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-3d83-account-create-hq42j"] Oct 05 21:11:50 crc kubenswrapper[4754]: E1005 21:11:50.578234 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6795f4e9-da0e-4ff6-8447-3fe8668e4db1" containerName="init" Oct 05 21:11:50 crc kubenswrapper[4754]: I1005 21:11:50.578250 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="6795f4e9-da0e-4ff6-8447-3fe8668e4db1" containerName="init" Oct 05 21:11:50 crc kubenswrapper[4754]: E1005 21:11:50.578268 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4752f446-876a-4d46-8e36-b4b03143c9db" containerName="init" Oct 05 21:11:50 crc kubenswrapper[4754]: I1005 21:11:50.578274 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="4752f446-876a-4d46-8e36-b4b03143c9db" containerName="init" Oct 05 21:11:50 crc kubenswrapper[4754]: E1005 21:11:50.578302 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a454365f-a32c-427d-a2d1-cecff2501134" containerName="mariadb-account-create" Oct 05 21:11:50 crc kubenswrapper[4754]: I1005 21:11:50.578308 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="a454365f-a32c-427d-a2d1-cecff2501134" containerName="mariadb-account-create" Oct 05 21:11:50 crc kubenswrapper[4754]: I1005 21:11:50.578452 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="a454365f-a32c-427d-a2d1-cecff2501134" containerName="mariadb-account-create" Oct 05 21:11:50 crc kubenswrapper[4754]: I1005 21:11:50.578472 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="4752f446-876a-4d46-8e36-b4b03143c9db" containerName="init" Oct 05 21:11:50 crc kubenswrapper[4754]: I1005 21:11:50.578537 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="6795f4e9-da0e-4ff6-8447-3fe8668e4db1" containerName="init" Oct 05 21:11:50 crc kubenswrapper[4754]: I1005 21:11:50.579142 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-3d83-account-create-hq42j" Oct 05 21:11:50 crc kubenswrapper[4754]: I1005 21:11:50.586159 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Oct 05 21:11:50 crc kubenswrapper[4754]: I1005 21:11:50.597760 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-3d83-account-create-hq42j"] Oct 05 21:11:50 crc kubenswrapper[4754]: I1005 21:11:50.624715 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qws7z\" (UniqueName: \"kubernetes.io/projected/be464a13-b373-46f0-bcfa-fcf9f491652e-kube-api-access-qws7z\") pod \"barbican-3d83-account-create-hq42j\" (UID: \"be464a13-b373-46f0-bcfa-fcf9f491652e\") " pod="openstack/barbican-3d83-account-create-hq42j" Oct 05 21:11:50 crc kubenswrapper[4754]: I1005 21:11:50.723138 4754 generic.go:334] "Generic (PLEG): container finished" podID="cbf1bb4f-3997-4581-ba76-64b10ba3f686" containerID="70f654feef08dc667d5d46a2f784bbef7d1df0ec788d6b6e2bcab20afb46ee6b" exitCode=0 Oct 05 21:11:50 crc kubenswrapper[4754]: I1005 21:11:50.723229 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-fvkbb" event={"ID":"cbf1bb4f-3997-4581-ba76-64b10ba3f686","Type":"ContainerDied","Data":"70f654feef08dc667d5d46a2f784bbef7d1df0ec788d6b6e2bcab20afb46ee6b"} Oct 05 21:11:50 crc kubenswrapper[4754]: I1005 21:11:50.726675 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qws7z\" (UniqueName: \"kubernetes.io/projected/be464a13-b373-46f0-bcfa-fcf9f491652e-kube-api-access-qws7z\") pod \"barbican-3d83-account-create-hq42j\" (UID: \"be464a13-b373-46f0-bcfa-fcf9f491652e\") " pod="openstack/barbican-3d83-account-create-hq42j" Oct 05 21:11:50 crc kubenswrapper[4754]: I1005 21:11:50.732760 4754 generic.go:334] "Generic (PLEG): container finished" podID="4ce6c227-efce-459e-be29-9276329560ea" containerID="faffd538d0e452dc90ce161248c8b227666300ba17506311f3e003f5bdeea6e4" exitCode=0 Oct 05 21:11:50 crc kubenswrapper[4754]: I1005 21:11:50.732783 4754 generic.go:334] "Generic (PLEG): container finished" podID="4ce6c227-efce-459e-be29-9276329560ea" containerID="fadc39cb4613473d324e7ee89a7fa105b4a1c1e8c30a775530ef8b396c30ad78" exitCode=143 Oct 05 21:11:50 crc kubenswrapper[4754]: I1005 21:11:50.732799 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"4ce6c227-efce-459e-be29-9276329560ea","Type":"ContainerDied","Data":"faffd538d0e452dc90ce161248c8b227666300ba17506311f3e003f5bdeea6e4"} Oct 05 21:11:50 crc kubenswrapper[4754]: I1005 21:11:50.732818 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"4ce6c227-efce-459e-be29-9276329560ea","Type":"ContainerDied","Data":"fadc39cb4613473d324e7ee89a7fa105b4a1c1e8c30a775530ef8b396c30ad78"} Oct 05 21:11:50 crc kubenswrapper[4754]: I1005 21:11:50.766456 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qws7z\" (UniqueName: \"kubernetes.io/projected/be464a13-b373-46f0-bcfa-fcf9f491652e-kube-api-access-qws7z\") pod \"barbican-3d83-account-create-hq42j\" (UID: \"be464a13-b373-46f0-bcfa-fcf9f491652e\") " pod="openstack/barbican-3d83-account-create-hq42j" Oct 05 21:11:50 crc kubenswrapper[4754]: I1005 21:11:50.888308 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-xncdg"] Oct 05 21:11:50 crc kubenswrapper[4754]: I1005 21:11:50.889752 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-xncdg" Oct 05 21:11:50 crc kubenswrapper[4754]: I1005 21:11:50.897829 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-xncdg"] Oct 05 21:11:50 crc kubenswrapper[4754]: I1005 21:11:50.899759 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-s6bxr" Oct 05 21:11:50 crc kubenswrapper[4754]: I1005 21:11:50.899976 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Oct 05 21:11:50 crc kubenswrapper[4754]: I1005 21:11:50.900095 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Oct 05 21:11:50 crc kubenswrapper[4754]: I1005 21:11:50.924307 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-3d83-account-create-hq42j" Oct 05 21:11:50 crc kubenswrapper[4754]: I1005 21:11:50.930102 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tvn5l\" (UniqueName: \"kubernetes.io/projected/94d0ae23-e26d-4bca-b7f3-e4af07ac2722-kube-api-access-tvn5l\") pod \"cinder-db-sync-xncdg\" (UID: \"94d0ae23-e26d-4bca-b7f3-e4af07ac2722\") " pod="openstack/cinder-db-sync-xncdg" Oct 05 21:11:50 crc kubenswrapper[4754]: I1005 21:11:50.930423 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94d0ae23-e26d-4bca-b7f3-e4af07ac2722-combined-ca-bundle\") pod \"cinder-db-sync-xncdg\" (UID: \"94d0ae23-e26d-4bca-b7f3-e4af07ac2722\") " pod="openstack/cinder-db-sync-xncdg" Oct 05 21:11:50 crc kubenswrapper[4754]: I1005 21:11:50.930591 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/94d0ae23-e26d-4bca-b7f3-e4af07ac2722-db-sync-config-data\") pod \"cinder-db-sync-xncdg\" (UID: \"94d0ae23-e26d-4bca-b7f3-e4af07ac2722\") " pod="openstack/cinder-db-sync-xncdg" Oct 05 21:11:50 crc kubenswrapper[4754]: I1005 21:11:50.930644 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/94d0ae23-e26d-4bca-b7f3-e4af07ac2722-scripts\") pod \"cinder-db-sync-xncdg\" (UID: \"94d0ae23-e26d-4bca-b7f3-e4af07ac2722\") " pod="openstack/cinder-db-sync-xncdg" Oct 05 21:11:50 crc kubenswrapper[4754]: I1005 21:11:50.930696 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/94d0ae23-e26d-4bca-b7f3-e4af07ac2722-config-data\") pod \"cinder-db-sync-xncdg\" (UID: \"94d0ae23-e26d-4bca-b7f3-e4af07ac2722\") " pod="openstack/cinder-db-sync-xncdg" Oct 05 21:11:50 crc kubenswrapper[4754]: I1005 21:11:50.930876 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/94d0ae23-e26d-4bca-b7f3-e4af07ac2722-etc-machine-id\") pod \"cinder-db-sync-xncdg\" (UID: \"94d0ae23-e26d-4bca-b7f3-e4af07ac2722\") " pod="openstack/cinder-db-sync-xncdg" Oct 05 21:11:51 crc kubenswrapper[4754]: I1005 21:11:51.032856 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tvn5l\" (UniqueName: \"kubernetes.io/projected/94d0ae23-e26d-4bca-b7f3-e4af07ac2722-kube-api-access-tvn5l\") pod \"cinder-db-sync-xncdg\" (UID: \"94d0ae23-e26d-4bca-b7f3-e4af07ac2722\") " pod="openstack/cinder-db-sync-xncdg" Oct 05 21:11:51 crc kubenswrapper[4754]: I1005 21:11:51.032916 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94d0ae23-e26d-4bca-b7f3-e4af07ac2722-combined-ca-bundle\") pod \"cinder-db-sync-xncdg\" (UID: \"94d0ae23-e26d-4bca-b7f3-e4af07ac2722\") " pod="openstack/cinder-db-sync-xncdg" Oct 05 21:11:51 crc kubenswrapper[4754]: I1005 21:11:51.032935 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/94d0ae23-e26d-4bca-b7f3-e4af07ac2722-db-sync-config-data\") pod \"cinder-db-sync-xncdg\" (UID: \"94d0ae23-e26d-4bca-b7f3-e4af07ac2722\") " pod="openstack/cinder-db-sync-xncdg" Oct 05 21:11:51 crc kubenswrapper[4754]: I1005 21:11:51.032971 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/94d0ae23-e26d-4bca-b7f3-e4af07ac2722-scripts\") pod \"cinder-db-sync-xncdg\" (UID: \"94d0ae23-e26d-4bca-b7f3-e4af07ac2722\") " pod="openstack/cinder-db-sync-xncdg" Oct 05 21:11:51 crc kubenswrapper[4754]: I1005 21:11:51.033001 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/94d0ae23-e26d-4bca-b7f3-e4af07ac2722-config-data\") pod \"cinder-db-sync-xncdg\" (UID: \"94d0ae23-e26d-4bca-b7f3-e4af07ac2722\") " pod="openstack/cinder-db-sync-xncdg" Oct 05 21:11:51 crc kubenswrapper[4754]: I1005 21:11:51.033050 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/94d0ae23-e26d-4bca-b7f3-e4af07ac2722-etc-machine-id\") pod \"cinder-db-sync-xncdg\" (UID: \"94d0ae23-e26d-4bca-b7f3-e4af07ac2722\") " pod="openstack/cinder-db-sync-xncdg" Oct 05 21:11:51 crc kubenswrapper[4754]: I1005 21:11:51.033113 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/94d0ae23-e26d-4bca-b7f3-e4af07ac2722-etc-machine-id\") pod \"cinder-db-sync-xncdg\" (UID: \"94d0ae23-e26d-4bca-b7f3-e4af07ac2722\") " pod="openstack/cinder-db-sync-xncdg" Oct 05 21:11:51 crc kubenswrapper[4754]: I1005 21:11:51.037635 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/94d0ae23-e26d-4bca-b7f3-e4af07ac2722-db-sync-config-data\") pod \"cinder-db-sync-xncdg\" (UID: \"94d0ae23-e26d-4bca-b7f3-e4af07ac2722\") " pod="openstack/cinder-db-sync-xncdg" Oct 05 21:11:51 crc kubenswrapper[4754]: I1005 21:11:51.038128 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/94d0ae23-e26d-4bca-b7f3-e4af07ac2722-config-data\") pod \"cinder-db-sync-xncdg\" (UID: \"94d0ae23-e26d-4bca-b7f3-e4af07ac2722\") " pod="openstack/cinder-db-sync-xncdg" Oct 05 21:11:51 crc kubenswrapper[4754]: I1005 21:11:51.038285 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94d0ae23-e26d-4bca-b7f3-e4af07ac2722-combined-ca-bundle\") pod \"cinder-db-sync-xncdg\" (UID: \"94d0ae23-e26d-4bca-b7f3-e4af07ac2722\") " pod="openstack/cinder-db-sync-xncdg" Oct 05 21:11:51 crc kubenswrapper[4754]: I1005 21:11:51.053685 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tvn5l\" (UniqueName: \"kubernetes.io/projected/94d0ae23-e26d-4bca-b7f3-e4af07ac2722-kube-api-access-tvn5l\") pod \"cinder-db-sync-xncdg\" (UID: \"94d0ae23-e26d-4bca-b7f3-e4af07ac2722\") " pod="openstack/cinder-db-sync-xncdg" Oct 05 21:11:51 crc kubenswrapper[4754]: I1005 21:11:51.058908 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/94d0ae23-e26d-4bca-b7f3-e4af07ac2722-scripts\") pod \"cinder-db-sync-xncdg\" (UID: \"94d0ae23-e26d-4bca-b7f3-e4af07ac2722\") " pod="openstack/cinder-db-sync-xncdg" Oct 05 21:11:51 crc kubenswrapper[4754]: I1005 21:11:51.120941 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-drwtj"] Oct 05 21:11:51 crc kubenswrapper[4754]: I1005 21:11:51.122077 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-drwtj" Oct 05 21:11:51 crc kubenswrapper[4754]: I1005 21:11:51.124627 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Oct 05 21:11:51 crc kubenswrapper[4754]: I1005 21:11:51.125946 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-zvvtc" Oct 05 21:11:51 crc kubenswrapper[4754]: I1005 21:11:51.126041 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Oct 05 21:11:51 crc kubenswrapper[4754]: I1005 21:11:51.149563 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-drwtj"] Oct 05 21:11:51 crc kubenswrapper[4754]: I1005 21:11:51.209425 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-xncdg" Oct 05 21:11:51 crc kubenswrapper[4754]: I1005 21:11:51.245775 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2rknr\" (UniqueName: \"kubernetes.io/projected/5594b283-6565-4ab3-89bd-9769992f6a20-kube-api-access-2rknr\") pod \"neutron-db-sync-drwtj\" (UID: \"5594b283-6565-4ab3-89bd-9769992f6a20\") " pod="openstack/neutron-db-sync-drwtj" Oct 05 21:11:51 crc kubenswrapper[4754]: I1005 21:11:51.245864 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5594b283-6565-4ab3-89bd-9769992f6a20-combined-ca-bundle\") pod \"neutron-db-sync-drwtj\" (UID: \"5594b283-6565-4ab3-89bd-9769992f6a20\") " pod="openstack/neutron-db-sync-drwtj" Oct 05 21:11:51 crc kubenswrapper[4754]: I1005 21:11:51.245901 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/5594b283-6565-4ab3-89bd-9769992f6a20-config\") pod \"neutron-db-sync-drwtj\" (UID: \"5594b283-6565-4ab3-89bd-9769992f6a20\") " pod="openstack/neutron-db-sync-drwtj" Oct 05 21:11:51 crc kubenswrapper[4754]: I1005 21:11:51.347594 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5594b283-6565-4ab3-89bd-9769992f6a20-combined-ca-bundle\") pod \"neutron-db-sync-drwtj\" (UID: \"5594b283-6565-4ab3-89bd-9769992f6a20\") " pod="openstack/neutron-db-sync-drwtj" Oct 05 21:11:51 crc kubenswrapper[4754]: I1005 21:11:51.347662 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/5594b283-6565-4ab3-89bd-9769992f6a20-config\") pod \"neutron-db-sync-drwtj\" (UID: \"5594b283-6565-4ab3-89bd-9769992f6a20\") " pod="openstack/neutron-db-sync-drwtj" Oct 05 21:11:51 crc kubenswrapper[4754]: I1005 21:11:51.347739 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2rknr\" (UniqueName: \"kubernetes.io/projected/5594b283-6565-4ab3-89bd-9769992f6a20-kube-api-access-2rknr\") pod \"neutron-db-sync-drwtj\" (UID: \"5594b283-6565-4ab3-89bd-9769992f6a20\") " pod="openstack/neutron-db-sync-drwtj" Oct 05 21:11:51 crc kubenswrapper[4754]: I1005 21:11:51.351977 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/5594b283-6565-4ab3-89bd-9769992f6a20-config\") pod \"neutron-db-sync-drwtj\" (UID: \"5594b283-6565-4ab3-89bd-9769992f6a20\") " pod="openstack/neutron-db-sync-drwtj" Oct 05 21:11:51 crc kubenswrapper[4754]: I1005 21:11:51.353116 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5594b283-6565-4ab3-89bd-9769992f6a20-combined-ca-bundle\") pod \"neutron-db-sync-drwtj\" (UID: \"5594b283-6565-4ab3-89bd-9769992f6a20\") " pod="openstack/neutron-db-sync-drwtj" Oct 05 21:11:51 crc kubenswrapper[4754]: I1005 21:11:51.371057 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2rknr\" (UniqueName: \"kubernetes.io/projected/5594b283-6565-4ab3-89bd-9769992f6a20-kube-api-access-2rknr\") pod \"neutron-db-sync-drwtj\" (UID: \"5594b283-6565-4ab3-89bd-9769992f6a20\") " pod="openstack/neutron-db-sync-drwtj" Oct 05 21:11:51 crc kubenswrapper[4754]: I1005 21:11:51.466062 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-drwtj" Oct 05 21:11:52 crc kubenswrapper[4754]: I1005 21:11:52.561729 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-fvkbb" Oct 05 21:11:52 crc kubenswrapper[4754]: I1005 21:11:52.668750 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/cbf1bb4f-3997-4581-ba76-64b10ba3f686-fernet-keys\") pod \"cbf1bb4f-3997-4581-ba76-64b10ba3f686\" (UID: \"cbf1bb4f-3997-4581-ba76-64b10ba3f686\") " Oct 05 21:11:52 crc kubenswrapper[4754]: I1005 21:11:52.669027 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cbf1bb4f-3997-4581-ba76-64b10ba3f686-config-data\") pod \"cbf1bb4f-3997-4581-ba76-64b10ba3f686\" (UID: \"cbf1bb4f-3997-4581-ba76-64b10ba3f686\") " Oct 05 21:11:52 crc kubenswrapper[4754]: I1005 21:11:52.669116 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/cbf1bb4f-3997-4581-ba76-64b10ba3f686-credential-keys\") pod \"cbf1bb4f-3997-4581-ba76-64b10ba3f686\" (UID: \"cbf1bb4f-3997-4581-ba76-64b10ba3f686\") " Oct 05 21:11:52 crc kubenswrapper[4754]: I1005 21:11:52.669216 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbf1bb4f-3997-4581-ba76-64b10ba3f686-combined-ca-bundle\") pod \"cbf1bb4f-3997-4581-ba76-64b10ba3f686\" (UID: \"cbf1bb4f-3997-4581-ba76-64b10ba3f686\") " Oct 05 21:11:52 crc kubenswrapper[4754]: I1005 21:11:52.669300 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cbf1bb4f-3997-4581-ba76-64b10ba3f686-scripts\") pod \"cbf1bb4f-3997-4581-ba76-64b10ba3f686\" (UID: \"cbf1bb4f-3997-4581-ba76-64b10ba3f686\") " Oct 05 21:11:52 crc kubenswrapper[4754]: I1005 21:11:52.669436 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2zgjt\" (UniqueName: \"kubernetes.io/projected/cbf1bb4f-3997-4581-ba76-64b10ba3f686-kube-api-access-2zgjt\") pod \"cbf1bb4f-3997-4581-ba76-64b10ba3f686\" (UID: \"cbf1bb4f-3997-4581-ba76-64b10ba3f686\") " Oct 05 21:11:52 crc kubenswrapper[4754]: I1005 21:11:52.679406 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cbf1bb4f-3997-4581-ba76-64b10ba3f686-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "cbf1bb4f-3997-4581-ba76-64b10ba3f686" (UID: "cbf1bb4f-3997-4581-ba76-64b10ba3f686"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:11:52 crc kubenswrapper[4754]: I1005 21:11:52.688129 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cbf1bb4f-3997-4581-ba76-64b10ba3f686-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "cbf1bb4f-3997-4581-ba76-64b10ba3f686" (UID: "cbf1bb4f-3997-4581-ba76-64b10ba3f686"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:11:52 crc kubenswrapper[4754]: I1005 21:11:52.697245 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cbf1bb4f-3997-4581-ba76-64b10ba3f686-scripts" (OuterVolumeSpecName: "scripts") pod "cbf1bb4f-3997-4581-ba76-64b10ba3f686" (UID: "cbf1bb4f-3997-4581-ba76-64b10ba3f686"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:11:52 crc kubenswrapper[4754]: I1005 21:11:52.697271 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cbf1bb4f-3997-4581-ba76-64b10ba3f686-kube-api-access-2zgjt" (OuterVolumeSpecName: "kube-api-access-2zgjt") pod "cbf1bb4f-3997-4581-ba76-64b10ba3f686" (UID: "cbf1bb4f-3997-4581-ba76-64b10ba3f686"). InnerVolumeSpecName "kube-api-access-2zgjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:11:52 crc kubenswrapper[4754]: I1005 21:11:52.709280 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cbf1bb4f-3997-4581-ba76-64b10ba3f686-config-data" (OuterVolumeSpecName: "config-data") pod "cbf1bb4f-3997-4581-ba76-64b10ba3f686" (UID: "cbf1bb4f-3997-4581-ba76-64b10ba3f686"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:11:52 crc kubenswrapper[4754]: I1005 21:11:52.713432 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cbf1bb4f-3997-4581-ba76-64b10ba3f686-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cbf1bb4f-3997-4581-ba76-64b10ba3f686" (UID: "cbf1bb4f-3997-4581-ba76-64b10ba3f686"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:11:52 crc kubenswrapper[4754]: I1005 21:11:52.762692 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-fvkbb" event={"ID":"cbf1bb4f-3997-4581-ba76-64b10ba3f686","Type":"ContainerDied","Data":"f681d7e7b627b29f4e56fe9964d7e02052e49864eb30180b1eeab9cdd31695dd"} Oct 05 21:11:52 crc kubenswrapper[4754]: I1005 21:11:52.762742 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f681d7e7b627b29f4e56fe9964d7e02052e49864eb30180b1eeab9cdd31695dd" Oct 05 21:11:52 crc kubenswrapper[4754]: I1005 21:11:52.762721 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-fvkbb" Oct 05 21:11:52 crc kubenswrapper[4754]: I1005 21:11:52.772803 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2zgjt\" (UniqueName: \"kubernetes.io/projected/cbf1bb4f-3997-4581-ba76-64b10ba3f686-kube-api-access-2zgjt\") on node \"crc\" DevicePath \"\"" Oct 05 21:11:52 crc kubenswrapper[4754]: I1005 21:11:52.772841 4754 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/cbf1bb4f-3997-4581-ba76-64b10ba3f686-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 05 21:11:52 crc kubenswrapper[4754]: I1005 21:11:52.772857 4754 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cbf1bb4f-3997-4581-ba76-64b10ba3f686-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 21:11:52 crc kubenswrapper[4754]: I1005 21:11:52.772868 4754 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/cbf1bb4f-3997-4581-ba76-64b10ba3f686-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 05 21:11:52 crc kubenswrapper[4754]: I1005 21:11:52.772878 4754 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbf1bb4f-3997-4581-ba76-64b10ba3f686-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 21:11:52 crc kubenswrapper[4754]: I1005 21:11:52.772889 4754 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cbf1bb4f-3997-4581-ba76-64b10ba3f686-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 21:11:52 crc kubenswrapper[4754]: I1005 21:11:52.902354 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-fvkbb"] Oct 05 21:11:52 crc kubenswrapper[4754]: I1005 21:11:52.910540 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-fvkbb"] Oct 05 21:11:52 crc kubenswrapper[4754]: I1005 21:11:52.994326 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-s4lpl"] Oct 05 21:11:52 crc kubenswrapper[4754]: E1005 21:11:52.995008 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cbf1bb4f-3997-4581-ba76-64b10ba3f686" containerName="keystone-bootstrap" Oct 05 21:11:52 crc kubenswrapper[4754]: I1005 21:11:52.995026 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="cbf1bb4f-3997-4581-ba76-64b10ba3f686" containerName="keystone-bootstrap" Oct 05 21:11:52 crc kubenswrapper[4754]: I1005 21:11:52.995272 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="cbf1bb4f-3997-4581-ba76-64b10ba3f686" containerName="keystone-bootstrap" Oct 05 21:11:52 crc kubenswrapper[4754]: I1005 21:11:52.996071 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-s4lpl" Oct 05 21:11:52 crc kubenswrapper[4754]: I1005 21:11:52.998309 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-5846s" Oct 05 21:11:52 crc kubenswrapper[4754]: I1005 21:11:52.998569 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 05 21:11:52 crc kubenswrapper[4754]: I1005 21:11:52.998637 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 05 21:11:52 crc kubenswrapper[4754]: I1005 21:11:52.998870 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 05 21:11:53 crc kubenswrapper[4754]: I1005 21:11:53.027283 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-s4lpl"] Oct 05 21:11:53 crc kubenswrapper[4754]: I1005 21:11:53.081657 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8705e88f-0fb4-4e7c-be21-d7bed4a8bede-config-data\") pod \"keystone-bootstrap-s4lpl\" (UID: \"8705e88f-0fb4-4e7c-be21-d7bed4a8bede\") " pod="openstack/keystone-bootstrap-s4lpl" Oct 05 21:11:53 crc kubenswrapper[4754]: I1005 21:11:53.081719 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8705e88f-0fb4-4e7c-be21-d7bed4a8bede-scripts\") pod \"keystone-bootstrap-s4lpl\" (UID: \"8705e88f-0fb4-4e7c-be21-d7bed4a8bede\") " pod="openstack/keystone-bootstrap-s4lpl" Oct 05 21:11:53 crc kubenswrapper[4754]: I1005 21:11:53.081760 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/8705e88f-0fb4-4e7c-be21-d7bed4a8bede-credential-keys\") pod \"keystone-bootstrap-s4lpl\" (UID: \"8705e88f-0fb4-4e7c-be21-d7bed4a8bede\") " pod="openstack/keystone-bootstrap-s4lpl" Oct 05 21:11:53 crc kubenswrapper[4754]: I1005 21:11:53.081853 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8705e88f-0fb4-4e7c-be21-d7bed4a8bede-fernet-keys\") pod \"keystone-bootstrap-s4lpl\" (UID: \"8705e88f-0fb4-4e7c-be21-d7bed4a8bede\") " pod="openstack/keystone-bootstrap-s4lpl" Oct 05 21:11:53 crc kubenswrapper[4754]: I1005 21:11:53.081878 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cqfh2\" (UniqueName: \"kubernetes.io/projected/8705e88f-0fb4-4e7c-be21-d7bed4a8bede-kube-api-access-cqfh2\") pod \"keystone-bootstrap-s4lpl\" (UID: \"8705e88f-0fb4-4e7c-be21-d7bed4a8bede\") " pod="openstack/keystone-bootstrap-s4lpl" Oct 05 21:11:53 crc kubenswrapper[4754]: I1005 21:11:53.081898 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8705e88f-0fb4-4e7c-be21-d7bed4a8bede-combined-ca-bundle\") pod \"keystone-bootstrap-s4lpl\" (UID: \"8705e88f-0fb4-4e7c-be21-d7bed4a8bede\") " pod="openstack/keystone-bootstrap-s4lpl" Oct 05 21:11:53 crc kubenswrapper[4754]: I1005 21:11:53.183337 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8705e88f-0fb4-4e7c-be21-d7bed4a8bede-config-data\") pod \"keystone-bootstrap-s4lpl\" (UID: \"8705e88f-0fb4-4e7c-be21-d7bed4a8bede\") " pod="openstack/keystone-bootstrap-s4lpl" Oct 05 21:11:53 crc kubenswrapper[4754]: I1005 21:11:53.183415 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8705e88f-0fb4-4e7c-be21-d7bed4a8bede-scripts\") pod \"keystone-bootstrap-s4lpl\" (UID: \"8705e88f-0fb4-4e7c-be21-d7bed4a8bede\") " pod="openstack/keystone-bootstrap-s4lpl" Oct 05 21:11:53 crc kubenswrapper[4754]: I1005 21:11:53.183463 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/8705e88f-0fb4-4e7c-be21-d7bed4a8bede-credential-keys\") pod \"keystone-bootstrap-s4lpl\" (UID: \"8705e88f-0fb4-4e7c-be21-d7bed4a8bede\") " pod="openstack/keystone-bootstrap-s4lpl" Oct 05 21:11:53 crc kubenswrapper[4754]: I1005 21:11:53.183681 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8705e88f-0fb4-4e7c-be21-d7bed4a8bede-fernet-keys\") pod \"keystone-bootstrap-s4lpl\" (UID: \"8705e88f-0fb4-4e7c-be21-d7bed4a8bede\") " pod="openstack/keystone-bootstrap-s4lpl" Oct 05 21:11:53 crc kubenswrapper[4754]: I1005 21:11:53.183728 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqfh2\" (UniqueName: \"kubernetes.io/projected/8705e88f-0fb4-4e7c-be21-d7bed4a8bede-kube-api-access-cqfh2\") pod \"keystone-bootstrap-s4lpl\" (UID: \"8705e88f-0fb4-4e7c-be21-d7bed4a8bede\") " pod="openstack/keystone-bootstrap-s4lpl" Oct 05 21:11:53 crc kubenswrapper[4754]: I1005 21:11:53.183756 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8705e88f-0fb4-4e7c-be21-d7bed4a8bede-combined-ca-bundle\") pod \"keystone-bootstrap-s4lpl\" (UID: \"8705e88f-0fb4-4e7c-be21-d7bed4a8bede\") " pod="openstack/keystone-bootstrap-s4lpl" Oct 05 21:11:53 crc kubenswrapper[4754]: I1005 21:11:53.191843 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/8705e88f-0fb4-4e7c-be21-d7bed4a8bede-credential-keys\") pod \"keystone-bootstrap-s4lpl\" (UID: \"8705e88f-0fb4-4e7c-be21-d7bed4a8bede\") " pod="openstack/keystone-bootstrap-s4lpl" Oct 05 21:11:53 crc kubenswrapper[4754]: I1005 21:11:53.193157 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8705e88f-0fb4-4e7c-be21-d7bed4a8bede-fernet-keys\") pod \"keystone-bootstrap-s4lpl\" (UID: \"8705e88f-0fb4-4e7c-be21-d7bed4a8bede\") " pod="openstack/keystone-bootstrap-s4lpl" Oct 05 21:11:53 crc kubenswrapper[4754]: I1005 21:11:53.195107 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8705e88f-0fb4-4e7c-be21-d7bed4a8bede-config-data\") pod \"keystone-bootstrap-s4lpl\" (UID: \"8705e88f-0fb4-4e7c-be21-d7bed4a8bede\") " pod="openstack/keystone-bootstrap-s4lpl" Oct 05 21:11:53 crc kubenswrapper[4754]: I1005 21:11:53.197189 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8705e88f-0fb4-4e7c-be21-d7bed4a8bede-combined-ca-bundle\") pod \"keystone-bootstrap-s4lpl\" (UID: \"8705e88f-0fb4-4e7c-be21-d7bed4a8bede\") " pod="openstack/keystone-bootstrap-s4lpl" Oct 05 21:11:53 crc kubenswrapper[4754]: I1005 21:11:53.197818 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8705e88f-0fb4-4e7c-be21-d7bed4a8bede-scripts\") pod \"keystone-bootstrap-s4lpl\" (UID: \"8705e88f-0fb4-4e7c-be21-d7bed4a8bede\") " pod="openstack/keystone-bootstrap-s4lpl" Oct 05 21:11:53 crc kubenswrapper[4754]: I1005 21:11:53.200659 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqfh2\" (UniqueName: \"kubernetes.io/projected/8705e88f-0fb4-4e7c-be21-d7bed4a8bede-kube-api-access-cqfh2\") pod \"keystone-bootstrap-s4lpl\" (UID: \"8705e88f-0fb4-4e7c-be21-d7bed4a8bede\") " pod="openstack/keystone-bootstrap-s4lpl" Oct 05 21:11:53 crc kubenswrapper[4754]: I1005 21:11:53.316062 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-s4lpl" Oct 05 21:11:53 crc kubenswrapper[4754]: I1005 21:11:53.530634 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-8b5c85b87-lsj2f" Oct 05 21:11:53 crc kubenswrapper[4754]: I1005 21:11:53.580154 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-7lxjx"] Oct 05 21:11:53 crc kubenswrapper[4754]: I1005 21:11:53.581256 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-77585f5f8c-7lxjx" podUID="7f0ec46e-d8ac-4749-b7d5-736dba5964f9" containerName="dnsmasq-dns" containerID="cri-o://3b296eec4386637feee1cf621c596d6555e98cfd40bfd7e08031ea4ee1ac7e03" gracePeriod=10 Oct 05 21:11:53 crc kubenswrapper[4754]: I1005 21:11:53.800462 4754 generic.go:334] "Generic (PLEG): container finished" podID="7f0ec46e-d8ac-4749-b7d5-736dba5964f9" containerID="3b296eec4386637feee1cf621c596d6555e98cfd40bfd7e08031ea4ee1ac7e03" exitCode=0 Oct 05 21:11:53 crc kubenswrapper[4754]: I1005 21:11:53.800530 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77585f5f8c-7lxjx" event={"ID":"7f0ec46e-d8ac-4749-b7d5-736dba5964f9","Type":"ContainerDied","Data":"3b296eec4386637feee1cf621c596d6555e98cfd40bfd7e08031ea4ee1ac7e03"} Oct 05 21:11:54 crc kubenswrapper[4754]: I1005 21:11:54.848686 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cbf1bb4f-3997-4581-ba76-64b10ba3f686" path="/var/lib/kubelet/pods/cbf1bb4f-3997-4581-ba76-64b10ba3f686/volumes" Oct 05 21:11:55 crc kubenswrapper[4754]: I1005 21:11:55.220522 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-6d595fb7c-nnfgp"] Oct 05 21:11:55 crc kubenswrapper[4754]: I1005 21:11:55.261461 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-68c6b67864-9msm5"] Oct 05 21:11:55 crc kubenswrapper[4754]: I1005 21:11:55.266274 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-68c6b67864-9msm5" Oct 05 21:11:55 crc kubenswrapper[4754]: I1005 21:11:55.271138 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-horizon-svc" Oct 05 21:11:55 crc kubenswrapper[4754]: I1005 21:11:55.302196 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-68c6b67864-9msm5"] Oct 05 21:11:55 crc kubenswrapper[4754]: I1005 21:11:55.328706 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/27cd6be5-f5eb-4158-bdd1-37f4df3d530d-logs\") pod \"horizon-68c6b67864-9msm5\" (UID: \"27cd6be5-f5eb-4158-bdd1-37f4df3d530d\") " pod="openstack/horizon-68c6b67864-9msm5" Oct 05 21:11:55 crc kubenswrapper[4754]: I1005 21:11:55.328754 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tmrsq\" (UniqueName: \"kubernetes.io/projected/27cd6be5-f5eb-4158-bdd1-37f4df3d530d-kube-api-access-tmrsq\") pod \"horizon-68c6b67864-9msm5\" (UID: \"27cd6be5-f5eb-4158-bdd1-37f4df3d530d\") " pod="openstack/horizon-68c6b67864-9msm5" Oct 05 21:11:55 crc kubenswrapper[4754]: I1005 21:11:55.328781 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/27cd6be5-f5eb-4158-bdd1-37f4df3d530d-horizon-secret-key\") pod \"horizon-68c6b67864-9msm5\" (UID: \"27cd6be5-f5eb-4158-bdd1-37f4df3d530d\") " pod="openstack/horizon-68c6b67864-9msm5" Oct 05 21:11:55 crc kubenswrapper[4754]: I1005 21:11:55.328836 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27cd6be5-f5eb-4158-bdd1-37f4df3d530d-combined-ca-bundle\") pod \"horizon-68c6b67864-9msm5\" (UID: \"27cd6be5-f5eb-4158-bdd1-37f4df3d530d\") " pod="openstack/horizon-68c6b67864-9msm5" Oct 05 21:11:55 crc kubenswrapper[4754]: I1005 21:11:55.328865 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/27cd6be5-f5eb-4158-bdd1-37f4df3d530d-scripts\") pod \"horizon-68c6b67864-9msm5\" (UID: \"27cd6be5-f5eb-4158-bdd1-37f4df3d530d\") " pod="openstack/horizon-68c6b67864-9msm5" Oct 05 21:11:55 crc kubenswrapper[4754]: I1005 21:11:55.328893 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/27cd6be5-f5eb-4158-bdd1-37f4df3d530d-config-data\") pod \"horizon-68c6b67864-9msm5\" (UID: \"27cd6be5-f5eb-4158-bdd1-37f4df3d530d\") " pod="openstack/horizon-68c6b67864-9msm5" Oct 05 21:11:55 crc kubenswrapper[4754]: I1005 21:11:55.328936 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/27cd6be5-f5eb-4158-bdd1-37f4df3d530d-horizon-tls-certs\") pod \"horizon-68c6b67864-9msm5\" (UID: \"27cd6be5-f5eb-4158-bdd1-37f4df3d530d\") " pod="openstack/horizon-68c6b67864-9msm5" Oct 05 21:11:55 crc kubenswrapper[4754]: I1005 21:11:55.369072 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-5f98f885c7-9ckfb"] Oct 05 21:11:55 crc kubenswrapper[4754]: I1005 21:11:55.387955 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-574454d6cb-vbnk6"] Oct 05 21:11:55 crc kubenswrapper[4754]: I1005 21:11:55.394571 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-574454d6cb-vbnk6" Oct 05 21:11:55 crc kubenswrapper[4754]: I1005 21:11:55.426137 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-574454d6cb-vbnk6"] Oct 05 21:11:55 crc kubenswrapper[4754]: I1005 21:11:55.436185 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/27cd6be5-f5eb-4158-bdd1-37f4df3d530d-horizon-secret-key\") pod \"horizon-68c6b67864-9msm5\" (UID: \"27cd6be5-f5eb-4158-bdd1-37f4df3d530d\") " pod="openstack/horizon-68c6b67864-9msm5" Oct 05 21:11:55 crc kubenswrapper[4754]: I1005 21:11:55.436286 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27cd6be5-f5eb-4158-bdd1-37f4df3d530d-combined-ca-bundle\") pod \"horizon-68c6b67864-9msm5\" (UID: \"27cd6be5-f5eb-4158-bdd1-37f4df3d530d\") " pod="openstack/horizon-68c6b67864-9msm5" Oct 05 21:11:55 crc kubenswrapper[4754]: I1005 21:11:55.436321 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/27cd6be5-f5eb-4158-bdd1-37f4df3d530d-scripts\") pod \"horizon-68c6b67864-9msm5\" (UID: \"27cd6be5-f5eb-4158-bdd1-37f4df3d530d\") " pod="openstack/horizon-68c6b67864-9msm5" Oct 05 21:11:55 crc kubenswrapper[4754]: I1005 21:11:55.436368 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/27cd6be5-f5eb-4158-bdd1-37f4df3d530d-config-data\") pod \"horizon-68c6b67864-9msm5\" (UID: \"27cd6be5-f5eb-4158-bdd1-37f4df3d530d\") " pod="openstack/horizon-68c6b67864-9msm5" Oct 05 21:11:55 crc kubenswrapper[4754]: I1005 21:11:55.436421 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/27cd6be5-f5eb-4158-bdd1-37f4df3d530d-horizon-tls-certs\") pod \"horizon-68c6b67864-9msm5\" (UID: \"27cd6be5-f5eb-4158-bdd1-37f4df3d530d\") " pod="openstack/horizon-68c6b67864-9msm5" Oct 05 21:11:55 crc kubenswrapper[4754]: I1005 21:11:55.436455 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/27cd6be5-f5eb-4158-bdd1-37f4df3d530d-logs\") pod \"horizon-68c6b67864-9msm5\" (UID: \"27cd6be5-f5eb-4158-bdd1-37f4df3d530d\") " pod="openstack/horizon-68c6b67864-9msm5" Oct 05 21:11:55 crc kubenswrapper[4754]: I1005 21:11:55.436481 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tmrsq\" (UniqueName: \"kubernetes.io/projected/27cd6be5-f5eb-4158-bdd1-37f4df3d530d-kube-api-access-tmrsq\") pod \"horizon-68c6b67864-9msm5\" (UID: \"27cd6be5-f5eb-4158-bdd1-37f4df3d530d\") " pod="openstack/horizon-68c6b67864-9msm5" Oct 05 21:11:55 crc kubenswrapper[4754]: I1005 21:11:55.437825 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/27cd6be5-f5eb-4158-bdd1-37f4df3d530d-config-data\") pod \"horizon-68c6b67864-9msm5\" (UID: \"27cd6be5-f5eb-4158-bdd1-37f4df3d530d\") " pod="openstack/horizon-68c6b67864-9msm5" Oct 05 21:11:55 crc kubenswrapper[4754]: I1005 21:11:55.437870 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/27cd6be5-f5eb-4158-bdd1-37f4df3d530d-scripts\") pod \"horizon-68c6b67864-9msm5\" (UID: \"27cd6be5-f5eb-4158-bdd1-37f4df3d530d\") " pod="openstack/horizon-68c6b67864-9msm5" Oct 05 21:11:55 crc kubenswrapper[4754]: I1005 21:11:55.438067 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/27cd6be5-f5eb-4158-bdd1-37f4df3d530d-logs\") pod \"horizon-68c6b67864-9msm5\" (UID: \"27cd6be5-f5eb-4158-bdd1-37f4df3d530d\") " pod="openstack/horizon-68c6b67864-9msm5" Oct 05 21:11:55 crc kubenswrapper[4754]: I1005 21:11:55.446702 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27cd6be5-f5eb-4158-bdd1-37f4df3d530d-combined-ca-bundle\") pod \"horizon-68c6b67864-9msm5\" (UID: \"27cd6be5-f5eb-4158-bdd1-37f4df3d530d\") " pod="openstack/horizon-68c6b67864-9msm5" Oct 05 21:11:55 crc kubenswrapper[4754]: I1005 21:11:55.450050 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/27cd6be5-f5eb-4158-bdd1-37f4df3d530d-horizon-tls-certs\") pod \"horizon-68c6b67864-9msm5\" (UID: \"27cd6be5-f5eb-4158-bdd1-37f4df3d530d\") " pod="openstack/horizon-68c6b67864-9msm5" Oct 05 21:11:55 crc kubenswrapper[4754]: I1005 21:11:55.462345 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/27cd6be5-f5eb-4158-bdd1-37f4df3d530d-horizon-secret-key\") pod \"horizon-68c6b67864-9msm5\" (UID: \"27cd6be5-f5eb-4158-bdd1-37f4df3d530d\") " pod="openstack/horizon-68c6b67864-9msm5" Oct 05 21:11:55 crc kubenswrapper[4754]: I1005 21:11:55.496506 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tmrsq\" (UniqueName: \"kubernetes.io/projected/27cd6be5-f5eb-4158-bdd1-37f4df3d530d-kube-api-access-tmrsq\") pod \"horizon-68c6b67864-9msm5\" (UID: \"27cd6be5-f5eb-4158-bdd1-37f4df3d530d\") " pod="openstack/horizon-68c6b67864-9msm5" Oct 05 21:11:55 crc kubenswrapper[4754]: I1005 21:11:55.537578 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/39d1aac3-14a1-45bf-9a9d-ba4cdb0afcb1-scripts\") pod \"horizon-574454d6cb-vbnk6\" (UID: \"39d1aac3-14a1-45bf-9a9d-ba4cdb0afcb1\") " pod="openstack/horizon-574454d6cb-vbnk6" Oct 05 21:11:55 crc kubenswrapper[4754]: I1005 21:11:55.537633 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bb5z7\" (UniqueName: \"kubernetes.io/projected/39d1aac3-14a1-45bf-9a9d-ba4cdb0afcb1-kube-api-access-bb5z7\") pod \"horizon-574454d6cb-vbnk6\" (UID: \"39d1aac3-14a1-45bf-9a9d-ba4cdb0afcb1\") " pod="openstack/horizon-574454d6cb-vbnk6" Oct 05 21:11:55 crc kubenswrapper[4754]: I1005 21:11:55.537668 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/39d1aac3-14a1-45bf-9a9d-ba4cdb0afcb1-horizon-tls-certs\") pod \"horizon-574454d6cb-vbnk6\" (UID: \"39d1aac3-14a1-45bf-9a9d-ba4cdb0afcb1\") " pod="openstack/horizon-574454d6cb-vbnk6" Oct 05 21:11:55 crc kubenswrapper[4754]: I1005 21:11:55.537691 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39d1aac3-14a1-45bf-9a9d-ba4cdb0afcb1-combined-ca-bundle\") pod \"horizon-574454d6cb-vbnk6\" (UID: \"39d1aac3-14a1-45bf-9a9d-ba4cdb0afcb1\") " pod="openstack/horizon-574454d6cb-vbnk6" Oct 05 21:11:55 crc kubenswrapper[4754]: I1005 21:11:55.537739 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/39d1aac3-14a1-45bf-9a9d-ba4cdb0afcb1-horizon-secret-key\") pod \"horizon-574454d6cb-vbnk6\" (UID: \"39d1aac3-14a1-45bf-9a9d-ba4cdb0afcb1\") " pod="openstack/horizon-574454d6cb-vbnk6" Oct 05 21:11:55 crc kubenswrapper[4754]: I1005 21:11:55.537764 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/39d1aac3-14a1-45bf-9a9d-ba4cdb0afcb1-config-data\") pod \"horizon-574454d6cb-vbnk6\" (UID: \"39d1aac3-14a1-45bf-9a9d-ba4cdb0afcb1\") " pod="openstack/horizon-574454d6cb-vbnk6" Oct 05 21:11:55 crc kubenswrapper[4754]: I1005 21:11:55.537842 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/39d1aac3-14a1-45bf-9a9d-ba4cdb0afcb1-logs\") pod \"horizon-574454d6cb-vbnk6\" (UID: \"39d1aac3-14a1-45bf-9a9d-ba4cdb0afcb1\") " pod="openstack/horizon-574454d6cb-vbnk6" Oct 05 21:11:55 crc kubenswrapper[4754]: I1005 21:11:55.603068 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-68c6b67864-9msm5" Oct 05 21:11:55 crc kubenswrapper[4754]: I1005 21:11:55.639585 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/39d1aac3-14a1-45bf-9a9d-ba4cdb0afcb1-logs\") pod \"horizon-574454d6cb-vbnk6\" (UID: \"39d1aac3-14a1-45bf-9a9d-ba4cdb0afcb1\") " pod="openstack/horizon-574454d6cb-vbnk6" Oct 05 21:11:55 crc kubenswrapper[4754]: I1005 21:11:55.639635 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/39d1aac3-14a1-45bf-9a9d-ba4cdb0afcb1-scripts\") pod \"horizon-574454d6cb-vbnk6\" (UID: \"39d1aac3-14a1-45bf-9a9d-ba4cdb0afcb1\") " pod="openstack/horizon-574454d6cb-vbnk6" Oct 05 21:11:55 crc kubenswrapper[4754]: I1005 21:11:55.639658 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bb5z7\" (UniqueName: \"kubernetes.io/projected/39d1aac3-14a1-45bf-9a9d-ba4cdb0afcb1-kube-api-access-bb5z7\") pod \"horizon-574454d6cb-vbnk6\" (UID: \"39d1aac3-14a1-45bf-9a9d-ba4cdb0afcb1\") " pod="openstack/horizon-574454d6cb-vbnk6" Oct 05 21:11:55 crc kubenswrapper[4754]: I1005 21:11:55.640024 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/39d1aac3-14a1-45bf-9a9d-ba4cdb0afcb1-horizon-tls-certs\") pod \"horizon-574454d6cb-vbnk6\" (UID: \"39d1aac3-14a1-45bf-9a9d-ba4cdb0afcb1\") " pod="openstack/horizon-574454d6cb-vbnk6" Oct 05 21:11:55 crc kubenswrapper[4754]: I1005 21:11:55.640059 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39d1aac3-14a1-45bf-9a9d-ba4cdb0afcb1-combined-ca-bundle\") pod \"horizon-574454d6cb-vbnk6\" (UID: \"39d1aac3-14a1-45bf-9a9d-ba4cdb0afcb1\") " pod="openstack/horizon-574454d6cb-vbnk6" Oct 05 21:11:55 crc kubenswrapper[4754]: I1005 21:11:55.640111 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/39d1aac3-14a1-45bf-9a9d-ba4cdb0afcb1-horizon-secret-key\") pod \"horizon-574454d6cb-vbnk6\" (UID: \"39d1aac3-14a1-45bf-9a9d-ba4cdb0afcb1\") " pod="openstack/horizon-574454d6cb-vbnk6" Oct 05 21:11:55 crc kubenswrapper[4754]: I1005 21:11:55.640144 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/39d1aac3-14a1-45bf-9a9d-ba4cdb0afcb1-config-data\") pod \"horizon-574454d6cb-vbnk6\" (UID: \"39d1aac3-14a1-45bf-9a9d-ba4cdb0afcb1\") " pod="openstack/horizon-574454d6cb-vbnk6" Oct 05 21:11:55 crc kubenswrapper[4754]: I1005 21:11:55.645634 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/39d1aac3-14a1-45bf-9a9d-ba4cdb0afcb1-logs\") pod \"horizon-574454d6cb-vbnk6\" (UID: \"39d1aac3-14a1-45bf-9a9d-ba4cdb0afcb1\") " pod="openstack/horizon-574454d6cb-vbnk6" Oct 05 21:11:55 crc kubenswrapper[4754]: I1005 21:11:55.646229 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/39d1aac3-14a1-45bf-9a9d-ba4cdb0afcb1-scripts\") pod \"horizon-574454d6cb-vbnk6\" (UID: \"39d1aac3-14a1-45bf-9a9d-ba4cdb0afcb1\") " pod="openstack/horizon-574454d6cb-vbnk6" Oct 05 21:11:55 crc kubenswrapper[4754]: I1005 21:11:55.658037 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/39d1aac3-14a1-45bf-9a9d-ba4cdb0afcb1-horizon-secret-key\") pod \"horizon-574454d6cb-vbnk6\" (UID: \"39d1aac3-14a1-45bf-9a9d-ba4cdb0afcb1\") " pod="openstack/horizon-574454d6cb-vbnk6" Oct 05 21:11:55 crc kubenswrapper[4754]: I1005 21:11:55.659117 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/39d1aac3-14a1-45bf-9a9d-ba4cdb0afcb1-config-data\") pod \"horizon-574454d6cb-vbnk6\" (UID: \"39d1aac3-14a1-45bf-9a9d-ba4cdb0afcb1\") " pod="openstack/horizon-574454d6cb-vbnk6" Oct 05 21:11:55 crc kubenswrapper[4754]: I1005 21:11:55.664618 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39d1aac3-14a1-45bf-9a9d-ba4cdb0afcb1-combined-ca-bundle\") pod \"horizon-574454d6cb-vbnk6\" (UID: \"39d1aac3-14a1-45bf-9a9d-ba4cdb0afcb1\") " pod="openstack/horizon-574454d6cb-vbnk6" Oct 05 21:11:55 crc kubenswrapper[4754]: I1005 21:11:55.670868 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bb5z7\" (UniqueName: \"kubernetes.io/projected/39d1aac3-14a1-45bf-9a9d-ba4cdb0afcb1-kube-api-access-bb5z7\") pod \"horizon-574454d6cb-vbnk6\" (UID: \"39d1aac3-14a1-45bf-9a9d-ba4cdb0afcb1\") " pod="openstack/horizon-574454d6cb-vbnk6" Oct 05 21:11:55 crc kubenswrapper[4754]: I1005 21:11:55.690970 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/39d1aac3-14a1-45bf-9a9d-ba4cdb0afcb1-horizon-tls-certs\") pod \"horizon-574454d6cb-vbnk6\" (UID: \"39d1aac3-14a1-45bf-9a9d-ba4cdb0afcb1\") " pod="openstack/horizon-574454d6cb-vbnk6" Oct 05 21:11:55 crc kubenswrapper[4754]: I1005 21:11:55.720263 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-574454d6cb-vbnk6" Oct 05 21:11:57 crc kubenswrapper[4754]: I1005 21:11:57.428939 4754 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-77585f5f8c-7lxjx" podUID="7f0ec46e-d8ac-4749-b7d5-736dba5964f9" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.125:5353: connect: connection refused" Oct 05 21:11:59 crc kubenswrapper[4754]: I1005 21:11:59.196179 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 05 21:11:59 crc kubenswrapper[4754]: I1005 21:11:59.321442 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/81cb813f-8f33-44b6-8e43-62ee912608ce-logs\") pod \"81cb813f-8f33-44b6-8e43-62ee912608ce\" (UID: \"81cb813f-8f33-44b6-8e43-62ee912608ce\") " Oct 05 21:11:59 crc kubenswrapper[4754]: I1005 21:11:59.321502 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h2jht\" (UniqueName: \"kubernetes.io/projected/81cb813f-8f33-44b6-8e43-62ee912608ce-kube-api-access-h2jht\") pod \"81cb813f-8f33-44b6-8e43-62ee912608ce\" (UID: \"81cb813f-8f33-44b6-8e43-62ee912608ce\") " Oct 05 21:11:59 crc kubenswrapper[4754]: I1005 21:11:59.321526 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81cb813f-8f33-44b6-8e43-62ee912608ce-combined-ca-bundle\") pod \"81cb813f-8f33-44b6-8e43-62ee912608ce\" (UID: \"81cb813f-8f33-44b6-8e43-62ee912608ce\") " Oct 05 21:11:59 crc kubenswrapper[4754]: I1005 21:11:59.321555 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/81cb813f-8f33-44b6-8e43-62ee912608ce-scripts\") pod \"81cb813f-8f33-44b6-8e43-62ee912608ce\" (UID: \"81cb813f-8f33-44b6-8e43-62ee912608ce\") " Oct 05 21:11:59 crc kubenswrapper[4754]: I1005 21:11:59.321629 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/81cb813f-8f33-44b6-8e43-62ee912608ce-httpd-run\") pod \"81cb813f-8f33-44b6-8e43-62ee912608ce\" (UID: \"81cb813f-8f33-44b6-8e43-62ee912608ce\") " Oct 05 21:11:59 crc kubenswrapper[4754]: I1005 21:11:59.321672 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81cb813f-8f33-44b6-8e43-62ee912608ce-config-data\") pod \"81cb813f-8f33-44b6-8e43-62ee912608ce\" (UID: \"81cb813f-8f33-44b6-8e43-62ee912608ce\") " Oct 05 21:11:59 crc kubenswrapper[4754]: I1005 21:11:59.321701 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"81cb813f-8f33-44b6-8e43-62ee912608ce\" (UID: \"81cb813f-8f33-44b6-8e43-62ee912608ce\") " Oct 05 21:11:59 crc kubenswrapper[4754]: I1005 21:11:59.322430 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/81cb813f-8f33-44b6-8e43-62ee912608ce-logs" (OuterVolumeSpecName: "logs") pod "81cb813f-8f33-44b6-8e43-62ee912608ce" (UID: "81cb813f-8f33-44b6-8e43-62ee912608ce"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 21:11:59 crc kubenswrapper[4754]: I1005 21:11:59.322640 4754 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/81cb813f-8f33-44b6-8e43-62ee912608ce-logs\") on node \"crc\" DevicePath \"\"" Oct 05 21:11:59 crc kubenswrapper[4754]: I1005 21:11:59.322974 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/81cb813f-8f33-44b6-8e43-62ee912608ce-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "81cb813f-8f33-44b6-8e43-62ee912608ce" (UID: "81cb813f-8f33-44b6-8e43-62ee912608ce"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 21:11:59 crc kubenswrapper[4754]: I1005 21:11:59.328637 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/81cb813f-8f33-44b6-8e43-62ee912608ce-scripts" (OuterVolumeSpecName: "scripts") pod "81cb813f-8f33-44b6-8e43-62ee912608ce" (UID: "81cb813f-8f33-44b6-8e43-62ee912608ce"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:11:59 crc kubenswrapper[4754]: I1005 21:11:59.333888 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/81cb813f-8f33-44b6-8e43-62ee912608ce-kube-api-access-h2jht" (OuterVolumeSpecName: "kube-api-access-h2jht") pod "81cb813f-8f33-44b6-8e43-62ee912608ce" (UID: "81cb813f-8f33-44b6-8e43-62ee912608ce"). InnerVolumeSpecName "kube-api-access-h2jht". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:11:59 crc kubenswrapper[4754]: I1005 21:11:59.338961 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "glance") pod "81cb813f-8f33-44b6-8e43-62ee912608ce" (UID: "81cb813f-8f33-44b6-8e43-62ee912608ce"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 05 21:11:59 crc kubenswrapper[4754]: I1005 21:11:59.390793 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/81cb813f-8f33-44b6-8e43-62ee912608ce-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "81cb813f-8f33-44b6-8e43-62ee912608ce" (UID: "81cb813f-8f33-44b6-8e43-62ee912608ce"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:11:59 crc kubenswrapper[4754]: I1005 21:11:59.403011 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/81cb813f-8f33-44b6-8e43-62ee912608ce-config-data" (OuterVolumeSpecName: "config-data") pod "81cb813f-8f33-44b6-8e43-62ee912608ce" (UID: "81cb813f-8f33-44b6-8e43-62ee912608ce"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:11:59 crc kubenswrapper[4754]: I1005 21:11:59.425164 4754 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/81cb813f-8f33-44b6-8e43-62ee912608ce-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 05 21:11:59 crc kubenswrapper[4754]: I1005 21:11:59.425199 4754 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81cb813f-8f33-44b6-8e43-62ee912608ce-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 21:11:59 crc kubenswrapper[4754]: I1005 21:11:59.425226 4754 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Oct 05 21:11:59 crc kubenswrapper[4754]: I1005 21:11:59.425238 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h2jht\" (UniqueName: \"kubernetes.io/projected/81cb813f-8f33-44b6-8e43-62ee912608ce-kube-api-access-h2jht\") on node \"crc\" DevicePath \"\"" Oct 05 21:11:59 crc kubenswrapper[4754]: I1005 21:11:59.425248 4754 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81cb813f-8f33-44b6-8e43-62ee912608ce-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 21:11:59 crc kubenswrapper[4754]: I1005 21:11:59.425257 4754 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/81cb813f-8f33-44b6-8e43-62ee912608ce-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 21:11:59 crc kubenswrapper[4754]: I1005 21:11:59.447992 4754 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Oct 05 21:11:59 crc kubenswrapper[4754]: I1005 21:11:59.527530 4754 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Oct 05 21:11:59 crc kubenswrapper[4754]: I1005 21:11:59.876964 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"81cb813f-8f33-44b6-8e43-62ee912608ce","Type":"ContainerDied","Data":"f9477fc2456936555beee0d94a22d56142582fc2dcb9e5ebd2fef64825c59b21"} Oct 05 21:11:59 crc kubenswrapper[4754]: I1005 21:11:59.877021 4754 scope.go:117] "RemoveContainer" containerID="9d1b636f33f2a4ee25b0ee5a822723b84ba632b207a15b251d5d962b9f42bf61" Oct 05 21:11:59 crc kubenswrapper[4754]: I1005 21:11:59.877677 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 05 21:11:59 crc kubenswrapper[4754]: I1005 21:11:59.958620 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 05 21:11:59 crc kubenswrapper[4754]: I1005 21:11:59.983582 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 05 21:11:59 crc kubenswrapper[4754]: I1005 21:11:59.994192 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 05 21:11:59 crc kubenswrapper[4754]: E1005 21:11:59.994552 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81cb813f-8f33-44b6-8e43-62ee912608ce" containerName="glance-log" Oct 05 21:11:59 crc kubenswrapper[4754]: I1005 21:11:59.994568 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="81cb813f-8f33-44b6-8e43-62ee912608ce" containerName="glance-log" Oct 05 21:11:59 crc kubenswrapper[4754]: E1005 21:11:59.994593 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81cb813f-8f33-44b6-8e43-62ee912608ce" containerName="glance-httpd" Oct 05 21:11:59 crc kubenswrapper[4754]: I1005 21:11:59.994599 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="81cb813f-8f33-44b6-8e43-62ee912608ce" containerName="glance-httpd" Oct 05 21:11:59 crc kubenswrapper[4754]: I1005 21:11:59.994784 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="81cb813f-8f33-44b6-8e43-62ee912608ce" containerName="glance-httpd" Oct 05 21:11:59 crc kubenswrapper[4754]: I1005 21:11:59.994797 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="81cb813f-8f33-44b6-8e43-62ee912608ce" containerName="glance-log" Oct 05 21:11:59 crc kubenswrapper[4754]: I1005 21:11:59.995685 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 05 21:11:59 crc kubenswrapper[4754]: I1005 21:11:59.999044 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 05 21:12:00 crc kubenswrapper[4754]: I1005 21:12:00.000086 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Oct 05 21:12:00 crc kubenswrapper[4754]: I1005 21:12:00.002163 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 05 21:12:00 crc kubenswrapper[4754]: I1005 21:12:00.140151 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fb20a498-c584-4d04-b3bd-fd3517d48d4f-scripts\") pod \"glance-default-internal-api-0\" (UID: \"fb20a498-c584-4d04-b3bd-fd3517d48d4f\") " pod="openstack/glance-default-internal-api-0" Oct 05 21:12:00 crc kubenswrapper[4754]: I1005 21:12:00.140214 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fb20a498-c584-4d04-b3bd-fd3517d48d4f-logs\") pod \"glance-default-internal-api-0\" (UID: \"fb20a498-c584-4d04-b3bd-fd3517d48d4f\") " pod="openstack/glance-default-internal-api-0" Oct 05 21:12:00 crc kubenswrapper[4754]: I1005 21:12:00.140291 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s2dpj\" (UniqueName: \"kubernetes.io/projected/fb20a498-c584-4d04-b3bd-fd3517d48d4f-kube-api-access-s2dpj\") pod \"glance-default-internal-api-0\" (UID: \"fb20a498-c584-4d04-b3bd-fd3517d48d4f\") " pod="openstack/glance-default-internal-api-0" Oct 05 21:12:00 crc kubenswrapper[4754]: I1005 21:12:00.140330 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb20a498-c584-4d04-b3bd-fd3517d48d4f-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"fb20a498-c584-4d04-b3bd-fd3517d48d4f\") " pod="openstack/glance-default-internal-api-0" Oct 05 21:12:00 crc kubenswrapper[4754]: I1005 21:12:00.140355 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb20a498-c584-4d04-b3bd-fd3517d48d4f-config-data\") pod \"glance-default-internal-api-0\" (UID: \"fb20a498-c584-4d04-b3bd-fd3517d48d4f\") " pod="openstack/glance-default-internal-api-0" Oct 05 21:12:00 crc kubenswrapper[4754]: I1005 21:12:00.140380 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-internal-api-0\" (UID: \"fb20a498-c584-4d04-b3bd-fd3517d48d4f\") " pod="openstack/glance-default-internal-api-0" Oct 05 21:12:00 crc kubenswrapper[4754]: I1005 21:12:00.140448 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb20a498-c584-4d04-b3bd-fd3517d48d4f-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"fb20a498-c584-4d04-b3bd-fd3517d48d4f\") " pod="openstack/glance-default-internal-api-0" Oct 05 21:12:00 crc kubenswrapper[4754]: I1005 21:12:00.140525 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/fb20a498-c584-4d04-b3bd-fd3517d48d4f-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"fb20a498-c584-4d04-b3bd-fd3517d48d4f\") " pod="openstack/glance-default-internal-api-0" Oct 05 21:12:00 crc kubenswrapper[4754]: I1005 21:12:00.241993 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fb20a498-c584-4d04-b3bd-fd3517d48d4f-scripts\") pod \"glance-default-internal-api-0\" (UID: \"fb20a498-c584-4d04-b3bd-fd3517d48d4f\") " pod="openstack/glance-default-internal-api-0" Oct 05 21:12:00 crc kubenswrapper[4754]: I1005 21:12:00.242031 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fb20a498-c584-4d04-b3bd-fd3517d48d4f-logs\") pod \"glance-default-internal-api-0\" (UID: \"fb20a498-c584-4d04-b3bd-fd3517d48d4f\") " pod="openstack/glance-default-internal-api-0" Oct 05 21:12:00 crc kubenswrapper[4754]: I1005 21:12:00.242066 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dpj\" (UniqueName: \"kubernetes.io/projected/fb20a498-c584-4d04-b3bd-fd3517d48d4f-kube-api-access-s2dpj\") pod \"glance-default-internal-api-0\" (UID: \"fb20a498-c584-4d04-b3bd-fd3517d48d4f\") " pod="openstack/glance-default-internal-api-0" Oct 05 21:12:00 crc kubenswrapper[4754]: I1005 21:12:00.242097 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb20a498-c584-4d04-b3bd-fd3517d48d4f-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"fb20a498-c584-4d04-b3bd-fd3517d48d4f\") " pod="openstack/glance-default-internal-api-0" Oct 05 21:12:00 crc kubenswrapper[4754]: I1005 21:12:00.242116 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb20a498-c584-4d04-b3bd-fd3517d48d4f-config-data\") pod \"glance-default-internal-api-0\" (UID: \"fb20a498-c584-4d04-b3bd-fd3517d48d4f\") " pod="openstack/glance-default-internal-api-0" Oct 05 21:12:00 crc kubenswrapper[4754]: I1005 21:12:00.242133 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-internal-api-0\" (UID: \"fb20a498-c584-4d04-b3bd-fd3517d48d4f\") " pod="openstack/glance-default-internal-api-0" Oct 05 21:12:00 crc kubenswrapper[4754]: I1005 21:12:00.242178 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb20a498-c584-4d04-b3bd-fd3517d48d4f-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"fb20a498-c584-4d04-b3bd-fd3517d48d4f\") " pod="openstack/glance-default-internal-api-0" Oct 05 21:12:00 crc kubenswrapper[4754]: I1005 21:12:00.242218 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/fb20a498-c584-4d04-b3bd-fd3517d48d4f-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"fb20a498-c584-4d04-b3bd-fd3517d48d4f\") " pod="openstack/glance-default-internal-api-0" Oct 05 21:12:00 crc kubenswrapper[4754]: I1005 21:12:00.242667 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/fb20a498-c584-4d04-b3bd-fd3517d48d4f-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"fb20a498-c584-4d04-b3bd-fd3517d48d4f\") " pod="openstack/glance-default-internal-api-0" Oct 05 21:12:00 crc kubenswrapper[4754]: I1005 21:12:00.243481 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fb20a498-c584-4d04-b3bd-fd3517d48d4f-logs\") pod \"glance-default-internal-api-0\" (UID: \"fb20a498-c584-4d04-b3bd-fd3517d48d4f\") " pod="openstack/glance-default-internal-api-0" Oct 05 21:12:00 crc kubenswrapper[4754]: I1005 21:12:00.243963 4754 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-internal-api-0\" (UID: \"fb20a498-c584-4d04-b3bd-fd3517d48d4f\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/glance-default-internal-api-0" Oct 05 21:12:00 crc kubenswrapper[4754]: I1005 21:12:00.247237 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb20a498-c584-4d04-b3bd-fd3517d48d4f-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"fb20a498-c584-4d04-b3bd-fd3517d48d4f\") " pod="openstack/glance-default-internal-api-0" Oct 05 21:12:00 crc kubenswrapper[4754]: I1005 21:12:00.247624 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fb20a498-c584-4d04-b3bd-fd3517d48d4f-scripts\") pod \"glance-default-internal-api-0\" (UID: \"fb20a498-c584-4d04-b3bd-fd3517d48d4f\") " pod="openstack/glance-default-internal-api-0" Oct 05 21:12:00 crc kubenswrapper[4754]: I1005 21:12:00.250630 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb20a498-c584-4d04-b3bd-fd3517d48d4f-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"fb20a498-c584-4d04-b3bd-fd3517d48d4f\") " pod="openstack/glance-default-internal-api-0" Oct 05 21:12:00 crc kubenswrapper[4754]: I1005 21:12:00.266528 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dpj\" (UniqueName: \"kubernetes.io/projected/fb20a498-c584-4d04-b3bd-fd3517d48d4f-kube-api-access-s2dpj\") pod \"glance-default-internal-api-0\" (UID: \"fb20a498-c584-4d04-b3bd-fd3517d48d4f\") " pod="openstack/glance-default-internal-api-0" Oct 05 21:12:00 crc kubenswrapper[4754]: I1005 21:12:00.271150 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb20a498-c584-4d04-b3bd-fd3517d48d4f-config-data\") pod \"glance-default-internal-api-0\" (UID: \"fb20a498-c584-4d04-b3bd-fd3517d48d4f\") " pod="openstack/glance-default-internal-api-0" Oct 05 21:12:00 crc kubenswrapper[4754]: I1005 21:12:00.277135 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-internal-api-0\" (UID: \"fb20a498-c584-4d04-b3bd-fd3517d48d4f\") " pod="openstack/glance-default-internal-api-0" Oct 05 21:12:00 crc kubenswrapper[4754]: I1005 21:12:00.318016 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 05 21:12:00 crc kubenswrapper[4754]: I1005 21:12:00.848579 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="81cb813f-8f33-44b6-8e43-62ee912608ce" path="/var/lib/kubelet/pods/81cb813f-8f33-44b6-8e43-62ee912608ce/volumes" Oct 05 21:12:02 crc kubenswrapper[4754]: E1005 21:12:02.113048 4754 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-placement-api:current-podified" Oct 05 21:12:02 crc kubenswrapper[4754]: E1005 21:12:02.113829 4754 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:placement-db-sync,Image:quay.io/podified-antelope-centos9/openstack-placement-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/placement,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:placement-dbsync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-v775j,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42482,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-db-sync-xlw68_openstack(94c73d9b-e224-4788-995d-be11dcee7b3c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 05 21:12:02 crc kubenswrapper[4754]: E1005 21:12:02.115449 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/placement-db-sync-xlw68" podUID="94c73d9b-e224-4788-995d-be11dcee7b3c" Oct 05 21:12:02 crc kubenswrapper[4754]: E1005 21:12:02.126070 4754 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Oct 05 21:12:02 crc kubenswrapper[4754]: E1005 21:12:02.126270 4754 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n8ch5dch54dh56ch56fh59bh657h7dh55ch556h68fh5f7h5d7h56fh594h5d4hddh55dh65fh59fh69h577h5cdhdbh84h559h5f4h5bh5d9h56dh58dh6dq,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-4gzct,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-5f98f885c7-9ckfb_openstack(146ce51d-4eb0-4aae-98ad-08bc8a8f9728): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 05 21:12:02 crc kubenswrapper[4754]: E1005 21:12:02.129241 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-5f98f885c7-9ckfb" podUID="146ce51d-4eb0-4aae-98ad-08bc8a8f9728" Oct 05 21:12:02 crc kubenswrapper[4754]: I1005 21:12:02.292026 4754 scope.go:117] "RemoveContainer" containerID="05a2fef72ece381d89a223af19f4818d6d2b4308ccededf2975de60136b2855c" Oct 05 21:12:02 crc kubenswrapper[4754]: I1005 21:12:02.444099 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 05 21:12:02 crc kubenswrapper[4754]: I1005 21:12:02.494828 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4ce6c227-efce-459e-be29-9276329560ea-logs\") pod \"4ce6c227-efce-459e-be29-9276329560ea\" (UID: \"4ce6c227-efce-459e-be29-9276329560ea\") " Oct 05 21:12:02 crc kubenswrapper[4754]: I1005 21:12:02.494958 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4ce6c227-efce-459e-be29-9276329560ea-scripts\") pod \"4ce6c227-efce-459e-be29-9276329560ea\" (UID: \"4ce6c227-efce-459e-be29-9276329560ea\") " Oct 05 21:12:02 crc kubenswrapper[4754]: I1005 21:12:02.494988 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ce6c227-efce-459e-be29-9276329560ea-combined-ca-bundle\") pod \"4ce6c227-efce-459e-be29-9276329560ea\" (UID: \"4ce6c227-efce-459e-be29-9276329560ea\") " Oct 05 21:12:02 crc kubenswrapper[4754]: I1005 21:12:02.495125 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ce6c227-efce-459e-be29-9276329560ea-config-data\") pod \"4ce6c227-efce-459e-be29-9276329560ea\" (UID: \"4ce6c227-efce-459e-be29-9276329560ea\") " Oct 05 21:12:02 crc kubenswrapper[4754]: I1005 21:12:02.495143 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2mfbh\" (UniqueName: \"kubernetes.io/projected/4ce6c227-efce-459e-be29-9276329560ea-kube-api-access-2mfbh\") pod \"4ce6c227-efce-459e-be29-9276329560ea\" (UID: \"4ce6c227-efce-459e-be29-9276329560ea\") " Oct 05 21:12:02 crc kubenswrapper[4754]: I1005 21:12:02.495185 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"4ce6c227-efce-459e-be29-9276329560ea\" (UID: \"4ce6c227-efce-459e-be29-9276329560ea\") " Oct 05 21:12:02 crc kubenswrapper[4754]: I1005 21:12:02.495216 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4ce6c227-efce-459e-be29-9276329560ea-httpd-run\") pod \"4ce6c227-efce-459e-be29-9276329560ea\" (UID: \"4ce6c227-efce-459e-be29-9276329560ea\") " Oct 05 21:12:02 crc kubenswrapper[4754]: I1005 21:12:02.497819 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4ce6c227-efce-459e-be29-9276329560ea-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "4ce6c227-efce-459e-be29-9276329560ea" (UID: "4ce6c227-efce-459e-be29-9276329560ea"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 21:12:02 crc kubenswrapper[4754]: I1005 21:12:02.502651 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4ce6c227-efce-459e-be29-9276329560ea-logs" (OuterVolumeSpecName: "logs") pod "4ce6c227-efce-459e-be29-9276329560ea" (UID: "4ce6c227-efce-459e-be29-9276329560ea"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 21:12:02 crc kubenswrapper[4754]: I1005 21:12:02.526182 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ce6c227-efce-459e-be29-9276329560ea-scripts" (OuterVolumeSpecName: "scripts") pod "4ce6c227-efce-459e-be29-9276329560ea" (UID: "4ce6c227-efce-459e-be29-9276329560ea"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:12:02 crc kubenswrapper[4754]: I1005 21:12:02.530171 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4ce6c227-efce-459e-be29-9276329560ea-kube-api-access-2mfbh" (OuterVolumeSpecName: "kube-api-access-2mfbh") pod "4ce6c227-efce-459e-be29-9276329560ea" (UID: "4ce6c227-efce-459e-be29-9276329560ea"). InnerVolumeSpecName "kube-api-access-2mfbh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:12:02 crc kubenswrapper[4754]: I1005 21:12:02.547406 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "glance") pod "4ce6c227-efce-459e-be29-9276329560ea" (UID: "4ce6c227-efce-459e-be29-9276329560ea"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 05 21:12:02 crc kubenswrapper[4754]: I1005 21:12:02.589312 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ce6c227-efce-459e-be29-9276329560ea-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4ce6c227-efce-459e-be29-9276329560ea" (UID: "4ce6c227-efce-459e-be29-9276329560ea"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:12:02 crc kubenswrapper[4754]: I1005 21:12:02.601693 4754 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4ce6c227-efce-459e-be29-9276329560ea-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 21:12:02 crc kubenswrapper[4754]: I1005 21:12:02.601914 4754 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ce6c227-efce-459e-be29-9276329560ea-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 21:12:02 crc kubenswrapper[4754]: I1005 21:12:02.601931 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2mfbh\" (UniqueName: \"kubernetes.io/projected/4ce6c227-efce-459e-be29-9276329560ea-kube-api-access-2mfbh\") on node \"crc\" DevicePath \"\"" Oct 05 21:12:02 crc kubenswrapper[4754]: I1005 21:12:02.601989 4754 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Oct 05 21:12:02 crc kubenswrapper[4754]: I1005 21:12:02.602000 4754 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4ce6c227-efce-459e-be29-9276329560ea-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 05 21:12:02 crc kubenswrapper[4754]: I1005 21:12:02.602011 4754 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4ce6c227-efce-459e-be29-9276329560ea-logs\") on node \"crc\" DevicePath \"\"" Oct 05 21:12:02 crc kubenswrapper[4754]: I1005 21:12:02.646257 4754 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Oct 05 21:12:02 crc kubenswrapper[4754]: I1005 21:12:02.678248 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ce6c227-efce-459e-be29-9276329560ea-config-data" (OuterVolumeSpecName: "config-data") pod "4ce6c227-efce-459e-be29-9276329560ea" (UID: "4ce6c227-efce-459e-be29-9276329560ea"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:12:02 crc kubenswrapper[4754]: I1005 21:12:02.703469 4754 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ce6c227-efce-459e-be29-9276329560ea-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 21:12:02 crc kubenswrapper[4754]: I1005 21:12:02.703610 4754 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Oct 05 21:12:02 crc kubenswrapper[4754]: I1005 21:12:02.734570 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77585f5f8c-7lxjx" Oct 05 21:12:02 crc kubenswrapper[4754]: I1005 21:12:02.806085 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7f0ec46e-d8ac-4749-b7d5-736dba5964f9-dns-swift-storage-0\") pod \"7f0ec46e-d8ac-4749-b7d5-736dba5964f9\" (UID: \"7f0ec46e-d8ac-4749-b7d5-736dba5964f9\") " Oct 05 21:12:02 crc kubenswrapper[4754]: I1005 21:12:02.806719 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7f0ec46e-d8ac-4749-b7d5-736dba5964f9-ovsdbserver-nb\") pod \"7f0ec46e-d8ac-4749-b7d5-736dba5964f9\" (UID: \"7f0ec46e-d8ac-4749-b7d5-736dba5964f9\") " Oct 05 21:12:02 crc kubenswrapper[4754]: I1005 21:12:02.807086 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7f0ec46e-d8ac-4749-b7d5-736dba5964f9-ovsdbserver-sb\") pod \"7f0ec46e-d8ac-4749-b7d5-736dba5964f9\" (UID: \"7f0ec46e-d8ac-4749-b7d5-736dba5964f9\") " Oct 05 21:12:02 crc kubenswrapper[4754]: I1005 21:12:02.807193 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wd6zp\" (UniqueName: \"kubernetes.io/projected/7f0ec46e-d8ac-4749-b7d5-736dba5964f9-kube-api-access-wd6zp\") pod \"7f0ec46e-d8ac-4749-b7d5-736dba5964f9\" (UID: \"7f0ec46e-d8ac-4749-b7d5-736dba5964f9\") " Oct 05 21:12:02 crc kubenswrapper[4754]: I1005 21:12:02.807256 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7f0ec46e-d8ac-4749-b7d5-736dba5964f9-config\") pod \"7f0ec46e-d8ac-4749-b7d5-736dba5964f9\" (UID: \"7f0ec46e-d8ac-4749-b7d5-736dba5964f9\") " Oct 05 21:12:02 crc kubenswrapper[4754]: I1005 21:12:02.807337 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7f0ec46e-d8ac-4749-b7d5-736dba5964f9-dns-svc\") pod \"7f0ec46e-d8ac-4749-b7d5-736dba5964f9\" (UID: \"7f0ec46e-d8ac-4749-b7d5-736dba5964f9\") " Oct 05 21:12:02 crc kubenswrapper[4754]: I1005 21:12:02.834262 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7f0ec46e-d8ac-4749-b7d5-736dba5964f9-kube-api-access-wd6zp" (OuterVolumeSpecName: "kube-api-access-wd6zp") pod "7f0ec46e-d8ac-4749-b7d5-736dba5964f9" (UID: "7f0ec46e-d8ac-4749-b7d5-736dba5964f9"). InnerVolumeSpecName "kube-api-access-wd6zp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:12:02 crc kubenswrapper[4754]: I1005 21:12:02.883210 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-drwtj"] Oct 05 21:12:02 crc kubenswrapper[4754]: I1005 21:12:02.911714 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wd6zp\" (UniqueName: \"kubernetes.io/projected/7f0ec46e-d8ac-4749-b7d5-736dba5964f9-kube-api-access-wd6zp\") on node \"crc\" DevicePath \"\"" Oct 05 21:12:02 crc kubenswrapper[4754]: I1005 21:12:02.918096 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"4ce6c227-efce-459e-be29-9276329560ea","Type":"ContainerDied","Data":"ae13d386ea4170e09acc7de99487262694163c237998f9629ec91b16b520f3db"} Oct 05 21:12:02 crc kubenswrapper[4754]: I1005 21:12:02.918155 4754 scope.go:117] "RemoveContainer" containerID="faffd538d0e452dc90ce161248c8b227666300ba17506311f3e003f5bdeea6e4" Oct 05 21:12:02 crc kubenswrapper[4754]: I1005 21:12:02.918233 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 05 21:12:02 crc kubenswrapper[4754]: W1005 21:12:02.918335 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5594b283_6565_4ab3_89bd_9769992f6a20.slice/crio-a31e7bf693a7dec198c99da63f8bd527f44e35c07981f8741533f043e5bae8c5 WatchSource:0}: Error finding container a31e7bf693a7dec198c99da63f8bd527f44e35c07981f8741533f043e5bae8c5: Status 404 returned error can't find the container with id a31e7bf693a7dec198c99da63f8bd527f44e35c07981f8741533f043e5bae8c5 Oct 05 21:12:02 crc kubenswrapper[4754]: I1005 21:12:02.920919 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77585f5f8c-7lxjx" event={"ID":"7f0ec46e-d8ac-4749-b7d5-736dba5964f9","Type":"ContainerDied","Data":"899f15b73841d1c02c0a24103eeeb8df271d5d86eb9182cde2e7f54080c6af16"} Oct 05 21:12:02 crc kubenswrapper[4754]: I1005 21:12:02.920999 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77585f5f8c-7lxjx" Oct 05 21:12:02 crc kubenswrapper[4754]: I1005 21:12:02.949399 4754 scope.go:117] "RemoveContainer" containerID="fadc39cb4613473d324e7ee89a7fa105b4a1c1e8c30a775530ef8b396c30ad78" Oct 05 21:12:02 crc kubenswrapper[4754]: E1005 21:12:02.949451 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-placement-api:current-podified\\\"\"" pod="openstack/placement-db-sync-xlw68" podUID="94c73d9b-e224-4788-995d-be11dcee7b3c" Oct 05 21:12:03 crc kubenswrapper[4754]: I1005 21:12:03.026441 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 05 21:12:03 crc kubenswrapper[4754]: I1005 21:12:03.066422 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 05 21:12:03 crc kubenswrapper[4754]: I1005 21:12:03.080637 4754 scope.go:117] "RemoveContainer" containerID="3b296eec4386637feee1cf621c596d6555e98cfd40bfd7e08031ea4ee1ac7e03" Oct 05 21:12:03 crc kubenswrapper[4754]: I1005 21:12:03.119603 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 05 21:12:03 crc kubenswrapper[4754]: E1005 21:12:03.120201 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f0ec46e-d8ac-4749-b7d5-736dba5964f9" containerName="init" Oct 05 21:12:03 crc kubenswrapper[4754]: I1005 21:12:03.120217 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f0ec46e-d8ac-4749-b7d5-736dba5964f9" containerName="init" Oct 05 21:12:03 crc kubenswrapper[4754]: E1005 21:12:03.120234 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ce6c227-efce-459e-be29-9276329560ea" containerName="glance-httpd" Oct 05 21:12:03 crc kubenswrapper[4754]: I1005 21:12:03.120240 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ce6c227-efce-459e-be29-9276329560ea" containerName="glance-httpd" Oct 05 21:12:03 crc kubenswrapper[4754]: E1005 21:12:03.120250 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ce6c227-efce-459e-be29-9276329560ea" containerName="glance-log" Oct 05 21:12:03 crc kubenswrapper[4754]: I1005 21:12:03.120263 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ce6c227-efce-459e-be29-9276329560ea" containerName="glance-log" Oct 05 21:12:03 crc kubenswrapper[4754]: E1005 21:12:03.120278 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f0ec46e-d8ac-4749-b7d5-736dba5964f9" containerName="dnsmasq-dns" Oct 05 21:12:03 crc kubenswrapper[4754]: I1005 21:12:03.120285 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f0ec46e-d8ac-4749-b7d5-736dba5964f9" containerName="dnsmasq-dns" Oct 05 21:12:03 crc kubenswrapper[4754]: I1005 21:12:03.120457 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="7f0ec46e-d8ac-4749-b7d5-736dba5964f9" containerName="dnsmasq-dns" Oct 05 21:12:03 crc kubenswrapper[4754]: I1005 21:12:03.120468 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ce6c227-efce-459e-be29-9276329560ea" containerName="glance-httpd" Oct 05 21:12:03 crc kubenswrapper[4754]: I1005 21:12:03.120484 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ce6c227-efce-459e-be29-9276329560ea" containerName="glance-log" Oct 05 21:12:03 crc kubenswrapper[4754]: I1005 21:12:03.128313 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 05 21:12:03 crc kubenswrapper[4754]: I1005 21:12:03.137768 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Oct 05 21:12:03 crc kubenswrapper[4754]: I1005 21:12:03.139131 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 05 21:12:03 crc kubenswrapper[4754]: I1005 21:12:03.163114 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 05 21:12:03 crc kubenswrapper[4754]: I1005 21:12:03.170834 4754 scope.go:117] "RemoveContainer" containerID="1666bb36d8d6aaba6b4ed3941d3eb3397fd3d42047e5eb00b6bec97ac6d0d12c" Oct 05 21:12:03 crc kubenswrapper[4754]: I1005 21:12:03.209054 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7f0ec46e-d8ac-4749-b7d5-736dba5964f9-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "7f0ec46e-d8ac-4749-b7d5-736dba5964f9" (UID: "7f0ec46e-d8ac-4749-b7d5-736dba5964f9"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 21:12:03 crc kubenswrapper[4754]: I1005 21:12:03.217977 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-574454d6cb-vbnk6"] Oct 05 21:12:03 crc kubenswrapper[4754]: I1005 21:12:03.221361 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6p5nx\" (UniqueName: \"kubernetes.io/projected/85cf3d49-cc86-43ad-ad97-0e4476de6ee6-kube-api-access-6p5nx\") pod \"glance-default-external-api-0\" (UID: \"85cf3d49-cc86-43ad-ad97-0e4476de6ee6\") " pod="openstack/glance-default-external-api-0" Oct 05 21:12:03 crc kubenswrapper[4754]: I1005 21:12:03.221414 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/85cf3d49-cc86-43ad-ad97-0e4476de6ee6-logs\") pod \"glance-default-external-api-0\" (UID: \"85cf3d49-cc86-43ad-ad97-0e4476de6ee6\") " pod="openstack/glance-default-external-api-0" Oct 05 21:12:03 crc kubenswrapper[4754]: I1005 21:12:03.221455 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"85cf3d49-cc86-43ad-ad97-0e4476de6ee6\") " pod="openstack/glance-default-external-api-0" Oct 05 21:12:03 crc kubenswrapper[4754]: I1005 21:12:03.221520 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/85cf3d49-cc86-43ad-ad97-0e4476de6ee6-config-data\") pod \"glance-default-external-api-0\" (UID: \"85cf3d49-cc86-43ad-ad97-0e4476de6ee6\") " pod="openstack/glance-default-external-api-0" Oct 05 21:12:03 crc kubenswrapper[4754]: I1005 21:12:03.221551 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/85cf3d49-cc86-43ad-ad97-0e4476de6ee6-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"85cf3d49-cc86-43ad-ad97-0e4476de6ee6\") " pod="openstack/glance-default-external-api-0" Oct 05 21:12:03 crc kubenswrapper[4754]: I1005 21:12:03.221575 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85cf3d49-cc86-43ad-ad97-0e4476de6ee6-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"85cf3d49-cc86-43ad-ad97-0e4476de6ee6\") " pod="openstack/glance-default-external-api-0" Oct 05 21:12:03 crc kubenswrapper[4754]: I1005 21:12:03.221606 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/85cf3d49-cc86-43ad-ad97-0e4476de6ee6-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"85cf3d49-cc86-43ad-ad97-0e4476de6ee6\") " pod="openstack/glance-default-external-api-0" Oct 05 21:12:03 crc kubenswrapper[4754]: I1005 21:12:03.221629 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/85cf3d49-cc86-43ad-ad97-0e4476de6ee6-scripts\") pod \"glance-default-external-api-0\" (UID: \"85cf3d49-cc86-43ad-ad97-0e4476de6ee6\") " pod="openstack/glance-default-external-api-0" Oct 05 21:12:03 crc kubenswrapper[4754]: I1005 21:12:03.221665 4754 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7f0ec46e-d8ac-4749-b7d5-736dba5964f9-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 05 21:12:03 crc kubenswrapper[4754]: I1005 21:12:03.233347 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-xncdg"] Oct 05 21:12:03 crc kubenswrapper[4754]: I1005 21:12:03.265676 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7f0ec46e-d8ac-4749-b7d5-736dba5964f9-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "7f0ec46e-d8ac-4749-b7d5-736dba5964f9" (UID: "7f0ec46e-d8ac-4749-b7d5-736dba5964f9"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 21:12:03 crc kubenswrapper[4754]: I1005 21:12:03.270515 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7f0ec46e-d8ac-4749-b7d5-736dba5964f9-config" (OuterVolumeSpecName: "config") pod "7f0ec46e-d8ac-4749-b7d5-736dba5964f9" (UID: "7f0ec46e-d8ac-4749-b7d5-736dba5964f9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 21:12:03 crc kubenswrapper[4754]: I1005 21:12:03.321690 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7f0ec46e-d8ac-4749-b7d5-736dba5964f9-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "7f0ec46e-d8ac-4749-b7d5-736dba5964f9" (UID: "7f0ec46e-d8ac-4749-b7d5-736dba5964f9"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 21:12:03 crc kubenswrapper[4754]: I1005 21:12:03.323773 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/85cf3d49-cc86-43ad-ad97-0e4476de6ee6-config-data\") pod \"glance-default-external-api-0\" (UID: \"85cf3d49-cc86-43ad-ad97-0e4476de6ee6\") " pod="openstack/glance-default-external-api-0" Oct 05 21:12:03 crc kubenswrapper[4754]: I1005 21:12:03.323826 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/85cf3d49-cc86-43ad-ad97-0e4476de6ee6-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"85cf3d49-cc86-43ad-ad97-0e4476de6ee6\") " pod="openstack/glance-default-external-api-0" Oct 05 21:12:03 crc kubenswrapper[4754]: I1005 21:12:03.323867 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85cf3d49-cc86-43ad-ad97-0e4476de6ee6-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"85cf3d49-cc86-43ad-ad97-0e4476de6ee6\") " pod="openstack/glance-default-external-api-0" Oct 05 21:12:03 crc kubenswrapper[4754]: I1005 21:12:03.323899 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/85cf3d49-cc86-43ad-ad97-0e4476de6ee6-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"85cf3d49-cc86-43ad-ad97-0e4476de6ee6\") " pod="openstack/glance-default-external-api-0" Oct 05 21:12:03 crc kubenswrapper[4754]: I1005 21:12:03.323930 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/85cf3d49-cc86-43ad-ad97-0e4476de6ee6-scripts\") pod \"glance-default-external-api-0\" (UID: \"85cf3d49-cc86-43ad-ad97-0e4476de6ee6\") " pod="openstack/glance-default-external-api-0" Oct 05 21:12:03 crc kubenswrapper[4754]: I1005 21:12:03.323955 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6p5nx\" (UniqueName: \"kubernetes.io/projected/85cf3d49-cc86-43ad-ad97-0e4476de6ee6-kube-api-access-6p5nx\") pod \"glance-default-external-api-0\" (UID: \"85cf3d49-cc86-43ad-ad97-0e4476de6ee6\") " pod="openstack/glance-default-external-api-0" Oct 05 21:12:03 crc kubenswrapper[4754]: I1005 21:12:03.323983 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/85cf3d49-cc86-43ad-ad97-0e4476de6ee6-logs\") pod \"glance-default-external-api-0\" (UID: \"85cf3d49-cc86-43ad-ad97-0e4476de6ee6\") " pod="openstack/glance-default-external-api-0" Oct 05 21:12:03 crc kubenswrapper[4754]: I1005 21:12:03.324019 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"85cf3d49-cc86-43ad-ad97-0e4476de6ee6\") " pod="openstack/glance-default-external-api-0" Oct 05 21:12:03 crc kubenswrapper[4754]: I1005 21:12:03.324085 4754 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7f0ec46e-d8ac-4749-b7d5-736dba5964f9-config\") on node \"crc\" DevicePath \"\"" Oct 05 21:12:03 crc kubenswrapper[4754]: I1005 21:12:03.324095 4754 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7f0ec46e-d8ac-4749-b7d5-736dba5964f9-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 05 21:12:03 crc kubenswrapper[4754]: I1005 21:12:03.324107 4754 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7f0ec46e-d8ac-4749-b7d5-736dba5964f9-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 05 21:12:03 crc kubenswrapper[4754]: I1005 21:12:03.329849 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/85cf3d49-cc86-43ad-ad97-0e4476de6ee6-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"85cf3d49-cc86-43ad-ad97-0e4476de6ee6\") " pod="openstack/glance-default-external-api-0" Oct 05 21:12:03 crc kubenswrapper[4754]: I1005 21:12:03.332055 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/85cf3d49-cc86-43ad-ad97-0e4476de6ee6-logs\") pod \"glance-default-external-api-0\" (UID: \"85cf3d49-cc86-43ad-ad97-0e4476de6ee6\") " pod="openstack/glance-default-external-api-0" Oct 05 21:12:03 crc kubenswrapper[4754]: I1005 21:12:03.332701 4754 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"85cf3d49-cc86-43ad-ad97-0e4476de6ee6\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/glance-default-external-api-0" Oct 05 21:12:03 crc kubenswrapper[4754]: I1005 21:12:03.335080 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/85cf3d49-cc86-43ad-ad97-0e4476de6ee6-scripts\") pod \"glance-default-external-api-0\" (UID: \"85cf3d49-cc86-43ad-ad97-0e4476de6ee6\") " pod="openstack/glance-default-external-api-0" Oct 05 21:12:03 crc kubenswrapper[4754]: I1005 21:12:03.337830 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/85cf3d49-cc86-43ad-ad97-0e4476de6ee6-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"85cf3d49-cc86-43ad-ad97-0e4476de6ee6\") " pod="openstack/glance-default-external-api-0" Oct 05 21:12:03 crc kubenswrapper[4754]: I1005 21:12:03.340165 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/85cf3d49-cc86-43ad-ad97-0e4476de6ee6-config-data\") pod \"glance-default-external-api-0\" (UID: \"85cf3d49-cc86-43ad-ad97-0e4476de6ee6\") " pod="openstack/glance-default-external-api-0" Oct 05 21:12:03 crc kubenswrapper[4754]: I1005 21:12:03.346358 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-3d83-account-create-hq42j"] Oct 05 21:12:03 crc kubenswrapper[4754]: I1005 21:12:03.348911 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7f0ec46e-d8ac-4749-b7d5-736dba5964f9-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "7f0ec46e-d8ac-4749-b7d5-736dba5964f9" (UID: "7f0ec46e-d8ac-4749-b7d5-736dba5964f9"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 21:12:03 crc kubenswrapper[4754]: I1005 21:12:03.390335 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6p5nx\" (UniqueName: \"kubernetes.io/projected/85cf3d49-cc86-43ad-ad97-0e4476de6ee6-kube-api-access-6p5nx\") pod \"glance-default-external-api-0\" (UID: \"85cf3d49-cc86-43ad-ad97-0e4476de6ee6\") " pod="openstack/glance-default-external-api-0" Oct 05 21:12:03 crc kubenswrapper[4754]: I1005 21:12:03.392016 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85cf3d49-cc86-43ad-ad97-0e4476de6ee6-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"85cf3d49-cc86-43ad-ad97-0e4476de6ee6\") " pod="openstack/glance-default-external-api-0" Oct 05 21:12:03 crc kubenswrapper[4754]: I1005 21:12:03.400886 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"85cf3d49-cc86-43ad-ad97-0e4476de6ee6\") " pod="openstack/glance-default-external-api-0" Oct 05 21:12:03 crc kubenswrapper[4754]: I1005 21:12:03.428384 4754 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7f0ec46e-d8ac-4749-b7d5-736dba5964f9-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 05 21:12:03 crc kubenswrapper[4754]: I1005 21:12:03.479181 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 05 21:12:03 crc kubenswrapper[4754]: I1005 21:12:03.599082 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-7lxjx"] Oct 05 21:12:03 crc kubenswrapper[4754]: I1005 21:12:03.613617 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-7lxjx"] Oct 05 21:12:03 crc kubenswrapper[4754]: I1005 21:12:03.726481 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-s4lpl"] Oct 05 21:12:03 crc kubenswrapper[4754]: I1005 21:12:03.770522 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-68c6b67864-9msm5"] Oct 05 21:12:03 crc kubenswrapper[4754]: I1005 21:12:03.781030 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5f98f885c7-9ckfb" Oct 05 21:12:03 crc kubenswrapper[4754]: W1005 21:12:03.813608 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod27cd6be5_f5eb_4158_bdd1_37f4df3d530d.slice/crio-de518e9a66f985fdfb6488044bfa833556110cf2d73cb9ff476ce82862781eae WatchSource:0}: Error finding container de518e9a66f985fdfb6488044bfa833556110cf2d73cb9ff476ce82862781eae: Status 404 returned error can't find the container with id de518e9a66f985fdfb6488044bfa833556110cf2d73cb9ff476ce82862781eae Oct 05 21:12:03 crc kubenswrapper[4754]: I1005 21:12:03.839508 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/146ce51d-4eb0-4aae-98ad-08bc8a8f9728-logs\") pod \"146ce51d-4eb0-4aae-98ad-08bc8a8f9728\" (UID: \"146ce51d-4eb0-4aae-98ad-08bc8a8f9728\") " Oct 05 21:12:03 crc kubenswrapper[4754]: I1005 21:12:03.839551 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/146ce51d-4eb0-4aae-98ad-08bc8a8f9728-config-data\") pod \"146ce51d-4eb0-4aae-98ad-08bc8a8f9728\" (UID: \"146ce51d-4eb0-4aae-98ad-08bc8a8f9728\") " Oct 05 21:12:03 crc kubenswrapper[4754]: I1005 21:12:03.839644 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/146ce51d-4eb0-4aae-98ad-08bc8a8f9728-scripts\") pod \"146ce51d-4eb0-4aae-98ad-08bc8a8f9728\" (UID: \"146ce51d-4eb0-4aae-98ad-08bc8a8f9728\") " Oct 05 21:12:03 crc kubenswrapper[4754]: I1005 21:12:03.839729 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/146ce51d-4eb0-4aae-98ad-08bc8a8f9728-horizon-secret-key\") pod \"146ce51d-4eb0-4aae-98ad-08bc8a8f9728\" (UID: \"146ce51d-4eb0-4aae-98ad-08bc8a8f9728\") " Oct 05 21:12:03 crc kubenswrapper[4754]: I1005 21:12:03.839847 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4gzct\" (UniqueName: \"kubernetes.io/projected/146ce51d-4eb0-4aae-98ad-08bc8a8f9728-kube-api-access-4gzct\") pod \"146ce51d-4eb0-4aae-98ad-08bc8a8f9728\" (UID: \"146ce51d-4eb0-4aae-98ad-08bc8a8f9728\") " Oct 05 21:12:03 crc kubenswrapper[4754]: I1005 21:12:03.843743 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/146ce51d-4eb0-4aae-98ad-08bc8a8f9728-logs" (OuterVolumeSpecName: "logs") pod "146ce51d-4eb0-4aae-98ad-08bc8a8f9728" (UID: "146ce51d-4eb0-4aae-98ad-08bc8a8f9728"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 21:12:03 crc kubenswrapper[4754]: I1005 21:12:03.844582 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/146ce51d-4eb0-4aae-98ad-08bc8a8f9728-scripts" (OuterVolumeSpecName: "scripts") pod "146ce51d-4eb0-4aae-98ad-08bc8a8f9728" (UID: "146ce51d-4eb0-4aae-98ad-08bc8a8f9728"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 21:12:03 crc kubenswrapper[4754]: I1005 21:12:03.846133 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/146ce51d-4eb0-4aae-98ad-08bc8a8f9728-config-data" (OuterVolumeSpecName: "config-data") pod "146ce51d-4eb0-4aae-98ad-08bc8a8f9728" (UID: "146ce51d-4eb0-4aae-98ad-08bc8a8f9728"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 21:12:03 crc kubenswrapper[4754]: I1005 21:12:03.852455 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/146ce51d-4eb0-4aae-98ad-08bc8a8f9728-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "146ce51d-4eb0-4aae-98ad-08bc8a8f9728" (UID: "146ce51d-4eb0-4aae-98ad-08bc8a8f9728"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:12:03 crc kubenswrapper[4754]: I1005 21:12:03.853008 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/146ce51d-4eb0-4aae-98ad-08bc8a8f9728-kube-api-access-4gzct" (OuterVolumeSpecName: "kube-api-access-4gzct") pod "146ce51d-4eb0-4aae-98ad-08bc8a8f9728" (UID: "146ce51d-4eb0-4aae-98ad-08bc8a8f9728"). InnerVolumeSpecName "kube-api-access-4gzct". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:12:03 crc kubenswrapper[4754]: I1005 21:12:03.876661 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 05 21:12:03 crc kubenswrapper[4754]: I1005 21:12:03.942031 4754 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/146ce51d-4eb0-4aae-98ad-08bc8a8f9728-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 05 21:12:03 crc kubenswrapper[4754]: I1005 21:12:03.942337 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4gzct\" (UniqueName: \"kubernetes.io/projected/146ce51d-4eb0-4aae-98ad-08bc8a8f9728-kube-api-access-4gzct\") on node \"crc\" DevicePath \"\"" Oct 05 21:12:03 crc kubenswrapper[4754]: I1005 21:12:03.942347 4754 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/146ce51d-4eb0-4aae-98ad-08bc8a8f9728-logs\") on node \"crc\" DevicePath \"\"" Oct 05 21:12:03 crc kubenswrapper[4754]: I1005 21:12:03.942360 4754 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/146ce51d-4eb0-4aae-98ad-08bc8a8f9728-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 21:12:03 crc kubenswrapper[4754]: I1005 21:12:03.942368 4754 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/146ce51d-4eb0-4aae-98ad-08bc8a8f9728-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 21:12:03 crc kubenswrapper[4754]: I1005 21:12:03.986272 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-drwtj" event={"ID":"5594b283-6565-4ab3-89bd-9769992f6a20","Type":"ContainerStarted","Data":"35e7ec9f9fdd5504e8cead852c32e929a754fb8f8f40cf4c89ee0c93643b374b"} Oct 05 21:12:03 crc kubenswrapper[4754]: I1005 21:12:03.986334 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-drwtj" event={"ID":"5594b283-6565-4ab3-89bd-9769992f6a20","Type":"ContainerStarted","Data":"a31e7bf693a7dec198c99da63f8bd527f44e35c07981f8741533f043e5bae8c5"} Oct 05 21:12:04 crc kubenswrapper[4754]: I1005 21:12:04.015004 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-drwtj" podStartSLOduration=13.014985166 podStartE2EDuration="13.014985166s" podCreationTimestamp="2025-10-05 21:11:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 21:12:04.01362949 +0000 UTC m=+1047.917748200" watchObservedRunningTime="2025-10-05 21:12:04.014985166 +0000 UTC m=+1047.919103866" Oct 05 21:12:04 crc kubenswrapper[4754]: I1005 21:12:04.020911 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0a10eb04-c1a7-447a-9400-87b39b1befdf","Type":"ContainerStarted","Data":"ec5e8223346ee4ff9e8367f4a328b561205c831c46f42c18f05917162be8e6a3"} Oct 05 21:12:04 crc kubenswrapper[4754]: I1005 21:12:04.023326 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-574454d6cb-vbnk6" event={"ID":"39d1aac3-14a1-45bf-9a9d-ba4cdb0afcb1","Type":"ContainerStarted","Data":"8f287041ecc74efdef130437536715ba47925b3906c22d822a16a9520f31e235"} Oct 05 21:12:04 crc kubenswrapper[4754]: I1005 21:12:04.023367 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-574454d6cb-vbnk6" event={"ID":"39d1aac3-14a1-45bf-9a9d-ba4cdb0afcb1","Type":"ContainerStarted","Data":"509b112f421a86be318fdf96489bccdc3a7da4f167c9b145821c1598940e3851"} Oct 05 21:12:04 crc kubenswrapper[4754]: I1005 21:12:04.031669 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-689b67764f-n4l6h" event={"ID":"eea94483-ef0a-4518-8258-587ee1112608","Type":"ContainerStarted","Data":"7a3be6d1860224c38bae686b11dc667c26380ccacdec9958d497ad2fde6b53c3"} Oct 05 21:12:04 crc kubenswrapper[4754]: I1005 21:12:04.031742 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-689b67764f-n4l6h" event={"ID":"eea94483-ef0a-4518-8258-587ee1112608","Type":"ContainerStarted","Data":"8e78a86fa9ef8e80c4b54e2989ab5675be879fe0716c3b91dc3600b2f0845d0c"} Oct 05 21:12:04 crc kubenswrapper[4754]: I1005 21:12:04.031953 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-689b67764f-n4l6h" podUID="eea94483-ef0a-4518-8258-587ee1112608" containerName="horizon-log" containerID="cri-o://8e78a86fa9ef8e80c4b54e2989ab5675be879fe0716c3b91dc3600b2f0845d0c" gracePeriod=30 Oct 05 21:12:04 crc kubenswrapper[4754]: I1005 21:12:04.032863 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-689b67764f-n4l6h" podUID="eea94483-ef0a-4518-8258-587ee1112608" containerName="horizon" containerID="cri-o://7a3be6d1860224c38bae686b11dc667c26380ccacdec9958d497ad2fde6b53c3" gracePeriod=30 Oct 05 21:12:04 crc kubenswrapper[4754]: I1005 21:12:04.049471 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6d595fb7c-nnfgp" event={"ID":"6f47b134-d55b-454d-8084-7676eb153105","Type":"ContainerStarted","Data":"90a1ac4ad2b42fce36308ae6b7c3ddb93e5d7a13682a2554e3b20086da7c736f"} Oct 05 21:12:04 crc kubenswrapper[4754]: I1005 21:12:04.049543 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6d595fb7c-nnfgp" event={"ID":"6f47b134-d55b-454d-8084-7676eb153105","Type":"ContainerStarted","Data":"3145d30e064762755f608d429f86b8dbe13676109d0cc8b22b9dc0af78ebbe01"} Oct 05 21:12:04 crc kubenswrapper[4754]: I1005 21:12:04.049685 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-6d595fb7c-nnfgp" podUID="6f47b134-d55b-454d-8084-7676eb153105" containerName="horizon-log" containerID="cri-o://3145d30e064762755f608d429f86b8dbe13676109d0cc8b22b9dc0af78ebbe01" gracePeriod=30 Oct 05 21:12:04 crc kubenswrapper[4754]: I1005 21:12:04.050130 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-6d595fb7c-nnfgp" podUID="6f47b134-d55b-454d-8084-7676eb153105" containerName="horizon" containerID="cri-o://90a1ac4ad2b42fce36308ae6b7c3ddb93e5d7a13682a2554e3b20086da7c736f" gracePeriod=30 Oct 05 21:12:04 crc kubenswrapper[4754]: I1005 21:12:04.050980 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-689b67764f-n4l6h" podStartSLOduration=4.942787568 podStartE2EDuration="23.050968962s" podCreationTimestamp="2025-10-05 21:11:41 +0000 UTC" firstStartedPulling="2025-10-05 21:11:44.147982641 +0000 UTC m=+1028.052101351" lastFinishedPulling="2025-10-05 21:12:02.256164035 +0000 UTC m=+1046.160282745" observedRunningTime="2025-10-05 21:12:04.049290228 +0000 UTC m=+1047.953408938" watchObservedRunningTime="2025-10-05 21:12:04.050968962 +0000 UTC m=+1047.955087672" Oct 05 21:12:04 crc kubenswrapper[4754]: I1005 21:12:04.065826 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-xncdg" event={"ID":"94d0ae23-e26d-4bca-b7f3-e4af07ac2722","Type":"ContainerStarted","Data":"77b307198d532f13d65299e2dad4c683500be12a880a83e15d30c5537d8c355f"} Oct 05 21:12:04 crc kubenswrapper[4754]: I1005 21:12:04.071452 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5f98f885c7-9ckfb" event={"ID":"146ce51d-4eb0-4aae-98ad-08bc8a8f9728","Type":"ContainerDied","Data":"cab648bee655ca280f477f5718ea20e53b388bab4b68ad37032e5beed7eed19f"} Oct 05 21:12:04 crc kubenswrapper[4754]: I1005 21:12:04.071810 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5f98f885c7-9ckfb" Oct 05 21:12:04 crc kubenswrapper[4754]: I1005 21:12:04.076686 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-68c6b67864-9msm5" event={"ID":"27cd6be5-f5eb-4158-bdd1-37f4df3d530d","Type":"ContainerStarted","Data":"de518e9a66f985fdfb6488044bfa833556110cf2d73cb9ff476ce82862781eae"} Oct 05 21:12:04 crc kubenswrapper[4754]: I1005 21:12:04.091169 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-6d595fb7c-nnfgp" podStartSLOduration=3.892459401 podStartE2EDuration="22.09114925s" podCreationTimestamp="2025-10-05 21:11:42 +0000 UTC" firstStartedPulling="2025-10-05 21:11:44.234935356 +0000 UTC m=+1028.139054056" lastFinishedPulling="2025-10-05 21:12:02.433625195 +0000 UTC m=+1046.337743905" observedRunningTime="2025-10-05 21:12:04.085327997 +0000 UTC m=+1047.989446717" watchObservedRunningTime="2025-10-05 21:12:04.09114925 +0000 UTC m=+1047.995267960" Oct 05 21:12:04 crc kubenswrapper[4754]: I1005 21:12:04.099456 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-s4lpl" event={"ID":"8705e88f-0fb4-4e7c-be21-d7bed4a8bede","Type":"ContainerStarted","Data":"974e8ad43cf78d6bc39959a4be53df734f231c3d68a2bb2d826fd58c08821ac9"} Oct 05 21:12:04 crc kubenswrapper[4754]: I1005 21:12:04.109314 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-3d83-account-create-hq42j" event={"ID":"be464a13-b373-46f0-bcfa-fcf9f491652e","Type":"ContainerStarted","Data":"6c2f653b095b4d003d9c9987856ca69ed98dbc3f4baf9d8b18f7966747cb44a7"} Oct 05 21:12:04 crc kubenswrapper[4754]: I1005 21:12:04.109812 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-3d83-account-create-hq42j" event={"ID":"be464a13-b373-46f0-bcfa-fcf9f491652e","Type":"ContainerStarted","Data":"36705fb521f8f0a306126789a49d4aa68b0bf4f37fbedc7bf8860ec37b402dc0"} Oct 05 21:12:04 crc kubenswrapper[4754]: I1005 21:12:04.112993 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"fb20a498-c584-4d04-b3bd-fd3517d48d4f","Type":"ContainerStarted","Data":"443c8126464536f5f38add9cc7d2e1e4ac37674aee47c1bcd6bfba501063bdd5"} Oct 05 21:12:04 crc kubenswrapper[4754]: I1005 21:12:04.301242 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 05 21:12:04 crc kubenswrapper[4754]: I1005 21:12:04.481083 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-5f98f885c7-9ckfb"] Oct 05 21:12:04 crc kubenswrapper[4754]: I1005 21:12:04.544019 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-5f98f885c7-9ckfb"] Oct 05 21:12:04 crc kubenswrapper[4754]: I1005 21:12:04.854483 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="146ce51d-4eb0-4aae-98ad-08bc8a8f9728" path="/var/lib/kubelet/pods/146ce51d-4eb0-4aae-98ad-08bc8a8f9728/volumes" Oct 05 21:12:04 crc kubenswrapper[4754]: I1005 21:12:04.856586 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4ce6c227-efce-459e-be29-9276329560ea" path="/var/lib/kubelet/pods/4ce6c227-efce-459e-be29-9276329560ea/volumes" Oct 05 21:12:04 crc kubenswrapper[4754]: I1005 21:12:04.857811 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7f0ec46e-d8ac-4749-b7d5-736dba5964f9" path="/var/lib/kubelet/pods/7f0ec46e-d8ac-4749-b7d5-736dba5964f9/volumes" Oct 05 21:12:05 crc kubenswrapper[4754]: I1005 21:12:05.133437 4754 generic.go:334] "Generic (PLEG): container finished" podID="be464a13-b373-46f0-bcfa-fcf9f491652e" containerID="6c2f653b095b4d003d9c9987856ca69ed98dbc3f4baf9d8b18f7966747cb44a7" exitCode=0 Oct 05 21:12:05 crc kubenswrapper[4754]: I1005 21:12:05.133544 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-3d83-account-create-hq42j" event={"ID":"be464a13-b373-46f0-bcfa-fcf9f491652e","Type":"ContainerDied","Data":"6c2f653b095b4d003d9c9987856ca69ed98dbc3f4baf9d8b18f7966747cb44a7"} Oct 05 21:12:05 crc kubenswrapper[4754]: I1005 21:12:05.145533 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-68c6b67864-9msm5" event={"ID":"27cd6be5-f5eb-4158-bdd1-37f4df3d530d","Type":"ContainerStarted","Data":"550fc17784cbed0030f262eb4233588b6160edd99bfa3b03eaec56012f272639"} Oct 05 21:12:05 crc kubenswrapper[4754]: I1005 21:12:05.145576 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-68c6b67864-9msm5" event={"ID":"27cd6be5-f5eb-4158-bdd1-37f4df3d530d","Type":"ContainerStarted","Data":"48ce6ae318ddb68e70258170c00936a21c53d9535eb9576fac3f4ad34120b991"} Oct 05 21:12:05 crc kubenswrapper[4754]: I1005 21:12:05.157837 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"fb20a498-c584-4d04-b3bd-fd3517d48d4f","Type":"ContainerStarted","Data":"133bffa9d8418e6d9fd4f21872636494b0cc27e364eb272cbd0c43c78f09f737"} Oct 05 21:12:05 crc kubenswrapper[4754]: I1005 21:12:05.158915 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"85cf3d49-cc86-43ad-ad97-0e4476de6ee6","Type":"ContainerStarted","Data":"dfab69117ac0d0afa11a7554d470ac8b1555ae8cceee3021b3d6bce7fa4433e2"} Oct 05 21:12:05 crc kubenswrapper[4754]: I1005 21:12:05.180338 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-s4lpl" event={"ID":"8705e88f-0fb4-4e7c-be21-d7bed4a8bede","Type":"ContainerStarted","Data":"4f500128b3304f698cca44b4340c265d04313ef491ceec4ef0596e6aa8e57918"} Oct 05 21:12:05 crc kubenswrapper[4754]: I1005 21:12:05.192259 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-574454d6cb-vbnk6" event={"ID":"39d1aac3-14a1-45bf-9a9d-ba4cdb0afcb1","Type":"ContainerStarted","Data":"4c8ff8949b109b6ab42fd3b43db16a006408ec8d790928e7785deedd5ef7d473"} Oct 05 21:12:05 crc kubenswrapper[4754]: I1005 21:12:05.215303 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-s4lpl" podStartSLOduration=13.21528747 podStartE2EDuration="13.21528747s" podCreationTimestamp="2025-10-05 21:11:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 21:12:05.209942549 +0000 UTC m=+1049.114061259" watchObservedRunningTime="2025-10-05 21:12:05.21528747 +0000 UTC m=+1049.119406180" Oct 05 21:12:05 crc kubenswrapper[4754]: I1005 21:12:05.215393 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-68c6b67864-9msm5" podStartSLOduration=10.215388463 podStartE2EDuration="10.215388463s" podCreationTimestamp="2025-10-05 21:11:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 21:12:05.17993897 +0000 UTC m=+1049.084057680" watchObservedRunningTime="2025-10-05 21:12:05.215388463 +0000 UTC m=+1049.119507163" Oct 05 21:12:05 crc kubenswrapper[4754]: I1005 21:12:05.233483 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-574454d6cb-vbnk6" podStartSLOduration=10.233464018 podStartE2EDuration="10.233464018s" podCreationTimestamp="2025-10-05 21:11:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 21:12:05.231335792 +0000 UTC m=+1049.135454492" watchObservedRunningTime="2025-10-05 21:12:05.233464018 +0000 UTC m=+1049.137582728" Oct 05 21:12:05 crc kubenswrapper[4754]: I1005 21:12:05.604626 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-68c6b67864-9msm5" Oct 05 21:12:05 crc kubenswrapper[4754]: I1005 21:12:05.606297 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-68c6b67864-9msm5" Oct 05 21:12:05 crc kubenswrapper[4754]: I1005 21:12:05.720623 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-574454d6cb-vbnk6" Oct 05 21:12:05 crc kubenswrapper[4754]: I1005 21:12:05.720678 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-574454d6cb-vbnk6" Oct 05 21:12:06 crc kubenswrapper[4754]: I1005 21:12:06.222319 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"85cf3d49-cc86-43ad-ad97-0e4476de6ee6","Type":"ContainerStarted","Data":"6d5bd90dbdb3906057a2a30e1be4ae1d3f097b8e2659882a66083011d71ad64d"} Oct 05 21:12:06 crc kubenswrapper[4754]: I1005 21:12:06.687933 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-3d83-account-create-hq42j" Oct 05 21:12:06 crc kubenswrapper[4754]: I1005 21:12:06.713155 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qws7z\" (UniqueName: \"kubernetes.io/projected/be464a13-b373-46f0-bcfa-fcf9f491652e-kube-api-access-qws7z\") pod \"be464a13-b373-46f0-bcfa-fcf9f491652e\" (UID: \"be464a13-b373-46f0-bcfa-fcf9f491652e\") " Oct 05 21:12:06 crc kubenswrapper[4754]: I1005 21:12:06.721844 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/be464a13-b373-46f0-bcfa-fcf9f491652e-kube-api-access-qws7z" (OuterVolumeSpecName: "kube-api-access-qws7z") pod "be464a13-b373-46f0-bcfa-fcf9f491652e" (UID: "be464a13-b373-46f0-bcfa-fcf9f491652e"). InnerVolumeSpecName "kube-api-access-qws7z". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:12:06 crc kubenswrapper[4754]: I1005 21:12:06.817191 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qws7z\" (UniqueName: \"kubernetes.io/projected/be464a13-b373-46f0-bcfa-fcf9f491652e-kube-api-access-qws7z\") on node \"crc\" DevicePath \"\"" Oct 05 21:12:07 crc kubenswrapper[4754]: I1005 21:12:07.239931 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-3d83-account-create-hq42j" Oct 05 21:12:07 crc kubenswrapper[4754]: I1005 21:12:07.240663 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-3d83-account-create-hq42j" event={"ID":"be464a13-b373-46f0-bcfa-fcf9f491652e","Type":"ContainerDied","Data":"36705fb521f8f0a306126789a49d4aa68b0bf4f37fbedc7bf8860ec37b402dc0"} Oct 05 21:12:07 crc kubenswrapper[4754]: I1005 21:12:07.240692 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="36705fb521f8f0a306126789a49d4aa68b0bf4f37fbedc7bf8860ec37b402dc0" Oct 05 21:12:07 crc kubenswrapper[4754]: I1005 21:12:07.430013 4754 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-77585f5f8c-7lxjx" podUID="7f0ec46e-d8ac-4749-b7d5-736dba5964f9" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.125:5353: i/o timeout" Oct 05 21:12:08 crc kubenswrapper[4754]: I1005 21:12:08.252679 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"fb20a498-c584-4d04-b3bd-fd3517d48d4f","Type":"ContainerStarted","Data":"58ebd0ada1f6bb5f01842c688aacdea75887cf282cfdb70b776799a10eb20507"} Oct 05 21:12:08 crc kubenswrapper[4754]: I1005 21:12:08.262316 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"85cf3d49-cc86-43ad-ad97-0e4476de6ee6","Type":"ContainerStarted","Data":"61c6c3b71ff15d6f9ac21858b8162e6919168883eb60955b3b78ab5843bf6b52"} Oct 05 21:12:08 crc kubenswrapper[4754]: I1005 21:12:08.284830 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=9.28481039 podStartE2EDuration="9.28481039s" podCreationTimestamp="2025-10-05 21:11:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 21:12:08.279179012 +0000 UTC m=+1052.183297722" watchObservedRunningTime="2025-10-05 21:12:08.28481039 +0000 UTC m=+1052.188929100" Oct 05 21:12:09 crc kubenswrapper[4754]: I1005 21:12:09.283640 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0a10eb04-c1a7-447a-9400-87b39b1befdf","Type":"ContainerStarted","Data":"83a4f5b028fd5d8370357e2ce7359bcdbe5419e295f9c87528c4827687035543"} Oct 05 21:12:10 crc kubenswrapper[4754]: I1005 21:12:10.319411 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 05 21:12:10 crc kubenswrapper[4754]: I1005 21:12:10.319470 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 05 21:12:10 crc kubenswrapper[4754]: I1005 21:12:10.362759 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 05 21:12:10 crc kubenswrapper[4754]: I1005 21:12:10.396289 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=8.396269779 podStartE2EDuration="8.396269779s" podCreationTimestamp="2025-10-05 21:12:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 21:12:08.324917085 +0000 UTC m=+1052.229035785" watchObservedRunningTime="2025-10-05 21:12:10.396269779 +0000 UTC m=+1054.300388489" Oct 05 21:12:10 crc kubenswrapper[4754]: I1005 21:12:10.414547 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 05 21:12:11 crc kubenswrapper[4754]: I1005 21:12:11.041543 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-6wlrw"] Oct 05 21:12:11 crc kubenswrapper[4754]: E1005 21:12:11.042119 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be464a13-b373-46f0-bcfa-fcf9f491652e" containerName="mariadb-account-create" Oct 05 21:12:11 crc kubenswrapper[4754]: I1005 21:12:11.042143 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="be464a13-b373-46f0-bcfa-fcf9f491652e" containerName="mariadb-account-create" Oct 05 21:12:11 crc kubenswrapper[4754]: I1005 21:12:11.042324 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="be464a13-b373-46f0-bcfa-fcf9f491652e" containerName="mariadb-account-create" Oct 05 21:12:11 crc kubenswrapper[4754]: I1005 21:12:11.043097 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-6wlrw" Oct 05 21:12:11 crc kubenswrapper[4754]: I1005 21:12:11.048680 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Oct 05 21:12:11 crc kubenswrapper[4754]: I1005 21:12:11.048719 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-nv4lv" Oct 05 21:12:11 crc kubenswrapper[4754]: I1005 21:12:11.069669 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-6wlrw"] Oct 05 21:12:11 crc kubenswrapper[4754]: I1005 21:12:11.229622 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-78rcw\" (UniqueName: \"kubernetes.io/projected/99975fd1-2c5d-4a2d-9790-0ef13f9fc3f9-kube-api-access-78rcw\") pod \"barbican-db-sync-6wlrw\" (UID: \"99975fd1-2c5d-4a2d-9790-0ef13f9fc3f9\") " pod="openstack/barbican-db-sync-6wlrw" Oct 05 21:12:11 crc kubenswrapper[4754]: I1005 21:12:11.229694 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99975fd1-2c5d-4a2d-9790-0ef13f9fc3f9-combined-ca-bundle\") pod \"barbican-db-sync-6wlrw\" (UID: \"99975fd1-2c5d-4a2d-9790-0ef13f9fc3f9\") " pod="openstack/barbican-db-sync-6wlrw" Oct 05 21:12:11 crc kubenswrapper[4754]: I1005 21:12:11.230178 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/99975fd1-2c5d-4a2d-9790-0ef13f9fc3f9-db-sync-config-data\") pod \"barbican-db-sync-6wlrw\" (UID: \"99975fd1-2c5d-4a2d-9790-0ef13f9fc3f9\") " pod="openstack/barbican-db-sync-6wlrw" Oct 05 21:12:11 crc kubenswrapper[4754]: I1005 21:12:11.313630 4754 generic.go:334] "Generic (PLEG): container finished" podID="8705e88f-0fb4-4e7c-be21-d7bed4a8bede" containerID="4f500128b3304f698cca44b4340c265d04313ef491ceec4ef0596e6aa8e57918" exitCode=0 Oct 05 21:12:11 crc kubenswrapper[4754]: I1005 21:12:11.315051 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-s4lpl" event={"ID":"8705e88f-0fb4-4e7c-be21-d7bed4a8bede","Type":"ContainerDied","Data":"4f500128b3304f698cca44b4340c265d04313ef491ceec4ef0596e6aa8e57918"} Oct 05 21:12:11 crc kubenswrapper[4754]: I1005 21:12:11.315105 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 05 21:12:11 crc kubenswrapper[4754]: I1005 21:12:11.315118 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 05 21:12:11 crc kubenswrapper[4754]: I1005 21:12:11.337921 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-78rcw\" (UniqueName: \"kubernetes.io/projected/99975fd1-2c5d-4a2d-9790-0ef13f9fc3f9-kube-api-access-78rcw\") pod \"barbican-db-sync-6wlrw\" (UID: \"99975fd1-2c5d-4a2d-9790-0ef13f9fc3f9\") " pod="openstack/barbican-db-sync-6wlrw" Oct 05 21:12:11 crc kubenswrapper[4754]: I1005 21:12:11.337991 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99975fd1-2c5d-4a2d-9790-0ef13f9fc3f9-combined-ca-bundle\") pod \"barbican-db-sync-6wlrw\" (UID: \"99975fd1-2c5d-4a2d-9790-0ef13f9fc3f9\") " pod="openstack/barbican-db-sync-6wlrw" Oct 05 21:12:11 crc kubenswrapper[4754]: I1005 21:12:11.338071 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/99975fd1-2c5d-4a2d-9790-0ef13f9fc3f9-db-sync-config-data\") pod \"barbican-db-sync-6wlrw\" (UID: \"99975fd1-2c5d-4a2d-9790-0ef13f9fc3f9\") " pod="openstack/barbican-db-sync-6wlrw" Oct 05 21:12:11 crc kubenswrapper[4754]: I1005 21:12:11.365531 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99975fd1-2c5d-4a2d-9790-0ef13f9fc3f9-combined-ca-bundle\") pod \"barbican-db-sync-6wlrw\" (UID: \"99975fd1-2c5d-4a2d-9790-0ef13f9fc3f9\") " pod="openstack/barbican-db-sync-6wlrw" Oct 05 21:12:11 crc kubenswrapper[4754]: I1005 21:12:11.367205 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-78rcw\" (UniqueName: \"kubernetes.io/projected/99975fd1-2c5d-4a2d-9790-0ef13f9fc3f9-kube-api-access-78rcw\") pod \"barbican-db-sync-6wlrw\" (UID: \"99975fd1-2c5d-4a2d-9790-0ef13f9fc3f9\") " pod="openstack/barbican-db-sync-6wlrw" Oct 05 21:12:11 crc kubenswrapper[4754]: I1005 21:12:11.369640 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/99975fd1-2c5d-4a2d-9790-0ef13f9fc3f9-db-sync-config-data\") pod \"barbican-db-sync-6wlrw\" (UID: \"99975fd1-2c5d-4a2d-9790-0ef13f9fc3f9\") " pod="openstack/barbican-db-sync-6wlrw" Oct 05 21:12:11 crc kubenswrapper[4754]: I1005 21:12:11.668970 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-6wlrw" Oct 05 21:12:12 crc kubenswrapper[4754]: I1005 21:12:12.287160 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-6wlrw"] Oct 05 21:12:12 crc kubenswrapper[4754]: I1005 21:12:12.331521 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-689b67764f-n4l6h" Oct 05 21:12:12 crc kubenswrapper[4754]: I1005 21:12:12.371859 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-6wlrw" event={"ID":"99975fd1-2c5d-4a2d-9790-0ef13f9fc3f9","Type":"ContainerStarted","Data":"54a24f4b6483ff93adcdb0fd84d666cd20b275782362b87c4dc257b6a10e4c56"} Oct 05 21:12:12 crc kubenswrapper[4754]: I1005 21:12:12.611106 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-6d595fb7c-nnfgp" Oct 05 21:12:12 crc kubenswrapper[4754]: I1005 21:12:12.911978 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-s4lpl" Oct 05 21:12:13 crc kubenswrapper[4754]: I1005 21:12:13.092448 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8705e88f-0fb4-4e7c-be21-d7bed4a8bede-config-data\") pod \"8705e88f-0fb4-4e7c-be21-d7bed4a8bede\" (UID: \"8705e88f-0fb4-4e7c-be21-d7bed4a8bede\") " Oct 05 21:12:13 crc kubenswrapper[4754]: I1005 21:12:13.092608 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/8705e88f-0fb4-4e7c-be21-d7bed4a8bede-credential-keys\") pod \"8705e88f-0fb4-4e7c-be21-d7bed4a8bede\" (UID: \"8705e88f-0fb4-4e7c-be21-d7bed4a8bede\") " Oct 05 21:12:13 crc kubenswrapper[4754]: I1005 21:12:13.092684 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cqfh2\" (UniqueName: \"kubernetes.io/projected/8705e88f-0fb4-4e7c-be21-d7bed4a8bede-kube-api-access-cqfh2\") pod \"8705e88f-0fb4-4e7c-be21-d7bed4a8bede\" (UID: \"8705e88f-0fb4-4e7c-be21-d7bed4a8bede\") " Oct 05 21:12:13 crc kubenswrapper[4754]: I1005 21:12:13.092768 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8705e88f-0fb4-4e7c-be21-d7bed4a8bede-scripts\") pod \"8705e88f-0fb4-4e7c-be21-d7bed4a8bede\" (UID: \"8705e88f-0fb4-4e7c-be21-d7bed4a8bede\") " Oct 05 21:12:13 crc kubenswrapper[4754]: I1005 21:12:13.093189 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8705e88f-0fb4-4e7c-be21-d7bed4a8bede-combined-ca-bundle\") pod \"8705e88f-0fb4-4e7c-be21-d7bed4a8bede\" (UID: \"8705e88f-0fb4-4e7c-be21-d7bed4a8bede\") " Oct 05 21:12:13 crc kubenswrapper[4754]: I1005 21:12:13.093258 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8705e88f-0fb4-4e7c-be21-d7bed4a8bede-fernet-keys\") pod \"8705e88f-0fb4-4e7c-be21-d7bed4a8bede\" (UID: \"8705e88f-0fb4-4e7c-be21-d7bed4a8bede\") " Oct 05 21:12:13 crc kubenswrapper[4754]: I1005 21:12:13.100942 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8705e88f-0fb4-4e7c-be21-d7bed4a8bede-kube-api-access-cqfh2" (OuterVolumeSpecName: "kube-api-access-cqfh2") pod "8705e88f-0fb4-4e7c-be21-d7bed4a8bede" (UID: "8705e88f-0fb4-4e7c-be21-d7bed4a8bede"). InnerVolumeSpecName "kube-api-access-cqfh2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:12:13 crc kubenswrapper[4754]: I1005 21:12:13.121625 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8705e88f-0fb4-4e7c-be21-d7bed4a8bede-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "8705e88f-0fb4-4e7c-be21-d7bed4a8bede" (UID: "8705e88f-0fb4-4e7c-be21-d7bed4a8bede"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:12:13 crc kubenswrapper[4754]: I1005 21:12:13.132106 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8705e88f-0fb4-4e7c-be21-d7bed4a8bede-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "8705e88f-0fb4-4e7c-be21-d7bed4a8bede" (UID: "8705e88f-0fb4-4e7c-be21-d7bed4a8bede"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:12:13 crc kubenswrapper[4754]: I1005 21:12:13.139677 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8705e88f-0fb4-4e7c-be21-d7bed4a8bede-scripts" (OuterVolumeSpecName: "scripts") pod "8705e88f-0fb4-4e7c-be21-d7bed4a8bede" (UID: "8705e88f-0fb4-4e7c-be21-d7bed4a8bede"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:12:13 crc kubenswrapper[4754]: I1005 21:12:13.146035 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8705e88f-0fb4-4e7c-be21-d7bed4a8bede-config-data" (OuterVolumeSpecName: "config-data") pod "8705e88f-0fb4-4e7c-be21-d7bed4a8bede" (UID: "8705e88f-0fb4-4e7c-be21-d7bed4a8bede"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:12:13 crc kubenswrapper[4754]: I1005 21:12:13.190647 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8705e88f-0fb4-4e7c-be21-d7bed4a8bede-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8705e88f-0fb4-4e7c-be21-d7bed4a8bede" (UID: "8705e88f-0fb4-4e7c-be21-d7bed4a8bede"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:12:13 crc kubenswrapper[4754]: I1005 21:12:13.197674 4754 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8705e88f-0fb4-4e7c-be21-d7bed4a8bede-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 21:12:13 crc kubenswrapper[4754]: I1005 21:12:13.197741 4754 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8705e88f-0fb4-4e7c-be21-d7bed4a8bede-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 05 21:12:13 crc kubenswrapper[4754]: I1005 21:12:13.197757 4754 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8705e88f-0fb4-4e7c-be21-d7bed4a8bede-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 21:12:13 crc kubenswrapper[4754]: I1005 21:12:13.197768 4754 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/8705e88f-0fb4-4e7c-be21-d7bed4a8bede-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 05 21:12:13 crc kubenswrapper[4754]: I1005 21:12:13.197778 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cqfh2\" (UniqueName: \"kubernetes.io/projected/8705e88f-0fb4-4e7c-be21-d7bed4a8bede-kube-api-access-cqfh2\") on node \"crc\" DevicePath \"\"" Oct 05 21:12:13 crc kubenswrapper[4754]: I1005 21:12:13.197823 4754 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8705e88f-0fb4-4e7c-be21-d7bed4a8bede-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 21:12:13 crc kubenswrapper[4754]: I1005 21:12:13.430487 4754 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 05 21:12:13 crc kubenswrapper[4754]: I1005 21:12:13.431392 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-s4lpl" event={"ID":"8705e88f-0fb4-4e7c-be21-d7bed4a8bede","Type":"ContainerDied","Data":"974e8ad43cf78d6bc39959a4be53df734f231c3d68a2bb2d826fd58c08821ac9"} Oct 05 21:12:13 crc kubenswrapper[4754]: I1005 21:12:13.431419 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="974e8ad43cf78d6bc39959a4be53df734f231c3d68a2bb2d826fd58c08821ac9" Oct 05 21:12:13 crc kubenswrapper[4754]: I1005 21:12:13.431415 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-s4lpl" Oct 05 21:12:13 crc kubenswrapper[4754]: I1005 21:12:13.477060 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-55bcb5dcdd-jhgrw"] Oct 05 21:12:13 crc kubenswrapper[4754]: E1005 21:12:13.477551 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8705e88f-0fb4-4e7c-be21-d7bed4a8bede" containerName="keystone-bootstrap" Oct 05 21:12:13 crc kubenswrapper[4754]: I1005 21:12:13.477569 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="8705e88f-0fb4-4e7c-be21-d7bed4a8bede" containerName="keystone-bootstrap" Oct 05 21:12:13 crc kubenswrapper[4754]: I1005 21:12:13.479319 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="8705e88f-0fb4-4e7c-be21-d7bed4a8bede" containerName="keystone-bootstrap" Oct 05 21:12:13 crc kubenswrapper[4754]: I1005 21:12:13.480347 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 05 21:12:13 crc kubenswrapper[4754]: I1005 21:12:13.480369 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 05 21:12:13 crc kubenswrapper[4754]: I1005 21:12:13.480731 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-55bcb5dcdd-jhgrw" Oct 05 21:12:13 crc kubenswrapper[4754]: I1005 21:12:13.484915 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-5846s" Oct 05 21:12:13 crc kubenswrapper[4754]: I1005 21:12:13.485098 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 05 21:12:13 crc kubenswrapper[4754]: I1005 21:12:13.489449 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Oct 05 21:12:13 crc kubenswrapper[4754]: I1005 21:12:13.489792 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 05 21:12:13 crc kubenswrapper[4754]: I1005 21:12:13.489959 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Oct 05 21:12:13 crc kubenswrapper[4754]: I1005 21:12:13.492779 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 05 21:12:13 crc kubenswrapper[4754]: I1005 21:12:13.528853 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-55bcb5dcdd-jhgrw"] Oct 05 21:12:13 crc kubenswrapper[4754]: I1005 21:12:13.549930 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 05 21:12:13 crc kubenswrapper[4754]: I1005 21:12:13.599617 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 05 21:12:13 crc kubenswrapper[4754]: I1005 21:12:13.611572 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d4c2091b-e2d2-4861-8190-79d548039f18-internal-tls-certs\") pod \"keystone-55bcb5dcdd-jhgrw\" (UID: \"d4c2091b-e2d2-4861-8190-79d548039f18\") " pod="openstack/keystone-55bcb5dcdd-jhgrw" Oct 05 21:12:13 crc kubenswrapper[4754]: I1005 21:12:13.611627 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4c2091b-e2d2-4861-8190-79d548039f18-combined-ca-bundle\") pod \"keystone-55bcb5dcdd-jhgrw\" (UID: \"d4c2091b-e2d2-4861-8190-79d548039f18\") " pod="openstack/keystone-55bcb5dcdd-jhgrw" Oct 05 21:12:13 crc kubenswrapper[4754]: I1005 21:12:13.611787 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d4c2091b-e2d2-4861-8190-79d548039f18-fernet-keys\") pod \"keystone-55bcb5dcdd-jhgrw\" (UID: \"d4c2091b-e2d2-4861-8190-79d548039f18\") " pod="openstack/keystone-55bcb5dcdd-jhgrw" Oct 05 21:12:13 crc kubenswrapper[4754]: I1005 21:12:13.612094 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w2l7w\" (UniqueName: \"kubernetes.io/projected/d4c2091b-e2d2-4861-8190-79d548039f18-kube-api-access-w2l7w\") pod \"keystone-55bcb5dcdd-jhgrw\" (UID: \"d4c2091b-e2d2-4861-8190-79d548039f18\") " pod="openstack/keystone-55bcb5dcdd-jhgrw" Oct 05 21:12:13 crc kubenswrapper[4754]: I1005 21:12:13.612198 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d4c2091b-e2d2-4861-8190-79d548039f18-scripts\") pod \"keystone-55bcb5dcdd-jhgrw\" (UID: \"d4c2091b-e2d2-4861-8190-79d548039f18\") " pod="openstack/keystone-55bcb5dcdd-jhgrw" Oct 05 21:12:13 crc kubenswrapper[4754]: I1005 21:12:13.612698 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/d4c2091b-e2d2-4861-8190-79d548039f18-credential-keys\") pod \"keystone-55bcb5dcdd-jhgrw\" (UID: \"d4c2091b-e2d2-4861-8190-79d548039f18\") " pod="openstack/keystone-55bcb5dcdd-jhgrw" Oct 05 21:12:13 crc kubenswrapper[4754]: I1005 21:12:13.612757 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d4c2091b-e2d2-4861-8190-79d548039f18-public-tls-certs\") pod \"keystone-55bcb5dcdd-jhgrw\" (UID: \"d4c2091b-e2d2-4861-8190-79d548039f18\") " pod="openstack/keystone-55bcb5dcdd-jhgrw" Oct 05 21:12:13 crc kubenswrapper[4754]: I1005 21:12:13.612942 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4c2091b-e2d2-4861-8190-79d548039f18-config-data\") pod \"keystone-55bcb5dcdd-jhgrw\" (UID: \"d4c2091b-e2d2-4861-8190-79d548039f18\") " pod="openstack/keystone-55bcb5dcdd-jhgrw" Oct 05 21:12:13 crc kubenswrapper[4754]: I1005 21:12:13.715049 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w2l7w\" (UniqueName: \"kubernetes.io/projected/d4c2091b-e2d2-4861-8190-79d548039f18-kube-api-access-w2l7w\") pod \"keystone-55bcb5dcdd-jhgrw\" (UID: \"d4c2091b-e2d2-4861-8190-79d548039f18\") " pod="openstack/keystone-55bcb5dcdd-jhgrw" Oct 05 21:12:13 crc kubenswrapper[4754]: I1005 21:12:13.715109 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d4c2091b-e2d2-4861-8190-79d548039f18-scripts\") pod \"keystone-55bcb5dcdd-jhgrw\" (UID: \"d4c2091b-e2d2-4861-8190-79d548039f18\") " pod="openstack/keystone-55bcb5dcdd-jhgrw" Oct 05 21:12:13 crc kubenswrapper[4754]: I1005 21:12:13.715141 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/d4c2091b-e2d2-4861-8190-79d548039f18-credential-keys\") pod \"keystone-55bcb5dcdd-jhgrw\" (UID: \"d4c2091b-e2d2-4861-8190-79d548039f18\") " pod="openstack/keystone-55bcb5dcdd-jhgrw" Oct 05 21:12:13 crc kubenswrapper[4754]: I1005 21:12:13.715162 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d4c2091b-e2d2-4861-8190-79d548039f18-public-tls-certs\") pod \"keystone-55bcb5dcdd-jhgrw\" (UID: \"d4c2091b-e2d2-4861-8190-79d548039f18\") " pod="openstack/keystone-55bcb5dcdd-jhgrw" Oct 05 21:12:13 crc kubenswrapper[4754]: I1005 21:12:13.715217 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4c2091b-e2d2-4861-8190-79d548039f18-config-data\") pod \"keystone-55bcb5dcdd-jhgrw\" (UID: \"d4c2091b-e2d2-4861-8190-79d548039f18\") " pod="openstack/keystone-55bcb5dcdd-jhgrw" Oct 05 21:12:13 crc kubenswrapper[4754]: I1005 21:12:13.715247 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d4c2091b-e2d2-4861-8190-79d548039f18-internal-tls-certs\") pod \"keystone-55bcb5dcdd-jhgrw\" (UID: \"d4c2091b-e2d2-4861-8190-79d548039f18\") " pod="openstack/keystone-55bcb5dcdd-jhgrw" Oct 05 21:12:13 crc kubenswrapper[4754]: I1005 21:12:13.715264 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4c2091b-e2d2-4861-8190-79d548039f18-combined-ca-bundle\") pod \"keystone-55bcb5dcdd-jhgrw\" (UID: \"d4c2091b-e2d2-4861-8190-79d548039f18\") " pod="openstack/keystone-55bcb5dcdd-jhgrw" Oct 05 21:12:13 crc kubenswrapper[4754]: I1005 21:12:13.715310 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d4c2091b-e2d2-4861-8190-79d548039f18-fernet-keys\") pod \"keystone-55bcb5dcdd-jhgrw\" (UID: \"d4c2091b-e2d2-4861-8190-79d548039f18\") " pod="openstack/keystone-55bcb5dcdd-jhgrw" Oct 05 21:12:13 crc kubenswrapper[4754]: I1005 21:12:13.724027 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/d4c2091b-e2d2-4861-8190-79d548039f18-credential-keys\") pod \"keystone-55bcb5dcdd-jhgrw\" (UID: \"d4c2091b-e2d2-4861-8190-79d548039f18\") " pod="openstack/keystone-55bcb5dcdd-jhgrw" Oct 05 21:12:13 crc kubenswrapper[4754]: I1005 21:12:13.728408 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4c2091b-e2d2-4861-8190-79d548039f18-config-data\") pod \"keystone-55bcb5dcdd-jhgrw\" (UID: \"d4c2091b-e2d2-4861-8190-79d548039f18\") " pod="openstack/keystone-55bcb5dcdd-jhgrw" Oct 05 21:12:13 crc kubenswrapper[4754]: I1005 21:12:13.728635 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d4c2091b-e2d2-4861-8190-79d548039f18-fernet-keys\") pod \"keystone-55bcb5dcdd-jhgrw\" (UID: \"d4c2091b-e2d2-4861-8190-79d548039f18\") " pod="openstack/keystone-55bcb5dcdd-jhgrw" Oct 05 21:12:13 crc kubenswrapper[4754]: I1005 21:12:13.728764 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d4c2091b-e2d2-4861-8190-79d548039f18-internal-tls-certs\") pod \"keystone-55bcb5dcdd-jhgrw\" (UID: \"d4c2091b-e2d2-4861-8190-79d548039f18\") " pod="openstack/keystone-55bcb5dcdd-jhgrw" Oct 05 21:12:13 crc kubenswrapper[4754]: I1005 21:12:13.728857 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d4c2091b-e2d2-4861-8190-79d548039f18-scripts\") pod \"keystone-55bcb5dcdd-jhgrw\" (UID: \"d4c2091b-e2d2-4861-8190-79d548039f18\") " pod="openstack/keystone-55bcb5dcdd-jhgrw" Oct 05 21:12:13 crc kubenswrapper[4754]: I1005 21:12:13.733334 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4c2091b-e2d2-4861-8190-79d548039f18-combined-ca-bundle\") pod \"keystone-55bcb5dcdd-jhgrw\" (UID: \"d4c2091b-e2d2-4861-8190-79d548039f18\") " pod="openstack/keystone-55bcb5dcdd-jhgrw" Oct 05 21:12:13 crc kubenswrapper[4754]: I1005 21:12:13.737871 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d4c2091b-e2d2-4861-8190-79d548039f18-public-tls-certs\") pod \"keystone-55bcb5dcdd-jhgrw\" (UID: \"d4c2091b-e2d2-4861-8190-79d548039f18\") " pod="openstack/keystone-55bcb5dcdd-jhgrw" Oct 05 21:12:13 crc kubenswrapper[4754]: I1005 21:12:13.742540 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w2l7w\" (UniqueName: \"kubernetes.io/projected/d4c2091b-e2d2-4861-8190-79d548039f18-kube-api-access-w2l7w\") pod \"keystone-55bcb5dcdd-jhgrw\" (UID: \"d4c2091b-e2d2-4861-8190-79d548039f18\") " pod="openstack/keystone-55bcb5dcdd-jhgrw" Oct 05 21:12:13 crc kubenswrapper[4754]: I1005 21:12:13.842743 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-55bcb5dcdd-jhgrw" Oct 05 21:12:14 crc kubenswrapper[4754]: I1005 21:12:14.487067 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 05 21:12:14 crc kubenswrapper[4754]: I1005 21:12:14.488298 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 05 21:12:14 crc kubenswrapper[4754]: I1005 21:12:14.772304 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-55bcb5dcdd-jhgrw"] Oct 05 21:12:15 crc kubenswrapper[4754]: I1005 21:12:15.606197 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-68c6b67864-9msm5" podUID="27cd6be5-f5eb-4158-bdd1-37f4df3d530d" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.148:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.148:8443: connect: connection refused" Oct 05 21:12:15 crc kubenswrapper[4754]: I1005 21:12:15.723596 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-574454d6cb-vbnk6" podUID="39d1aac3-14a1-45bf-9a9d-ba4cdb0afcb1" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.149:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.149:8443: connect: connection refused" Oct 05 21:12:16 crc kubenswrapper[4754]: I1005 21:12:16.512232 4754 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 05 21:12:16 crc kubenswrapper[4754]: I1005 21:12:16.512736 4754 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 05 21:12:17 crc kubenswrapper[4754]: I1005 21:12:17.703449 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 05 21:12:17 crc kubenswrapper[4754]: I1005 21:12:17.703570 4754 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 05 21:12:17 crc kubenswrapper[4754]: I1005 21:12:17.707661 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 05 21:12:17 crc kubenswrapper[4754]: I1005 21:12:17.834351 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 05 21:12:17 crc kubenswrapper[4754]: I1005 21:12:17.834507 4754 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 05 21:12:17 crc kubenswrapper[4754]: I1005 21:12:17.870102 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 05 21:12:25 crc kubenswrapper[4754]: I1005 21:12:25.604321 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-68c6b67864-9msm5" podUID="27cd6be5-f5eb-4158-bdd1-37f4df3d530d" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.148:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.148:8443: connect: connection refused" Oct 05 21:12:25 crc kubenswrapper[4754]: I1005 21:12:25.721567 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-574454d6cb-vbnk6" podUID="39d1aac3-14a1-45bf-9a9d-ba4cdb0afcb1" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.149:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.149:8443: connect: connection refused" Oct 05 21:12:34 crc kubenswrapper[4754]: W1005 21:12:34.281878 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd4c2091b_e2d2_4861_8190_79d548039f18.slice/crio-58c1e3fefb788d4039d055aff53f85984fb74e5ce70a52911f42eaf5ca3e9637 WatchSource:0}: Error finding container 58c1e3fefb788d4039d055aff53f85984fb74e5ce70a52911f42eaf5ca3e9637: Status 404 returned error can't find the container with id 58c1e3fefb788d4039d055aff53f85984fb74e5ce70a52911f42eaf5ca3e9637 Oct 05 21:12:34 crc kubenswrapper[4754]: I1005 21:12:34.692427 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-55bcb5dcdd-jhgrw" event={"ID":"d4c2091b-e2d2-4861-8190-79d548039f18","Type":"ContainerStarted","Data":"58c1e3fefb788d4039d055aff53f85984fb74e5ce70a52911f42eaf5ca3e9637"} Oct 05 21:12:34 crc kubenswrapper[4754]: I1005 21:12:34.694898 4754 generic.go:334] "Generic (PLEG): container finished" podID="eea94483-ef0a-4518-8258-587ee1112608" containerID="7a3be6d1860224c38bae686b11dc667c26380ccacdec9958d497ad2fde6b53c3" exitCode=137 Oct 05 21:12:34 crc kubenswrapper[4754]: I1005 21:12:34.694920 4754 generic.go:334] "Generic (PLEG): container finished" podID="eea94483-ef0a-4518-8258-587ee1112608" containerID="8e78a86fa9ef8e80c4b54e2989ab5675be879fe0716c3b91dc3600b2f0845d0c" exitCode=137 Oct 05 21:12:34 crc kubenswrapper[4754]: I1005 21:12:34.694965 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-689b67764f-n4l6h" event={"ID":"eea94483-ef0a-4518-8258-587ee1112608","Type":"ContainerDied","Data":"7a3be6d1860224c38bae686b11dc667c26380ccacdec9958d497ad2fde6b53c3"} Oct 05 21:12:34 crc kubenswrapper[4754]: I1005 21:12:34.694985 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-689b67764f-n4l6h" event={"ID":"eea94483-ef0a-4518-8258-587ee1112608","Type":"ContainerDied","Data":"8e78a86fa9ef8e80c4b54e2989ab5675be879fe0716c3b91dc3600b2f0845d0c"} Oct 05 21:12:34 crc kubenswrapper[4754]: I1005 21:12:34.697280 4754 generic.go:334] "Generic (PLEG): container finished" podID="6f47b134-d55b-454d-8084-7676eb153105" containerID="90a1ac4ad2b42fce36308ae6b7c3ddb93e5d7a13682a2554e3b20086da7c736f" exitCode=137 Oct 05 21:12:34 crc kubenswrapper[4754]: I1005 21:12:34.697332 4754 generic.go:334] "Generic (PLEG): container finished" podID="6f47b134-d55b-454d-8084-7676eb153105" containerID="3145d30e064762755f608d429f86b8dbe13676109d0cc8b22b9dc0af78ebbe01" exitCode=137 Oct 05 21:12:34 crc kubenswrapper[4754]: I1005 21:12:34.697343 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6d595fb7c-nnfgp" event={"ID":"6f47b134-d55b-454d-8084-7676eb153105","Type":"ContainerDied","Data":"90a1ac4ad2b42fce36308ae6b7c3ddb93e5d7a13682a2554e3b20086da7c736f"} Oct 05 21:12:34 crc kubenswrapper[4754]: I1005 21:12:34.697399 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6d595fb7c-nnfgp" event={"ID":"6f47b134-d55b-454d-8084-7676eb153105","Type":"ContainerDied","Data":"3145d30e064762755f608d429f86b8dbe13676109d0cc8b22b9dc0af78ebbe01"} Oct 05 21:12:40 crc kubenswrapper[4754]: I1005 21:12:40.607726 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-68c6b67864-9msm5" podUID="27cd6be5-f5eb-4158-bdd1-37f4df3d530d" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.148:8443/dashboard/auth/login/?next=/dashboard/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 05 21:12:40 crc kubenswrapper[4754]: I1005 21:12:40.608368 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-68c6b67864-9msm5" Oct 05 21:12:40 crc kubenswrapper[4754]: I1005 21:12:40.609298 4754 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="horizon" containerStatusID={"Type":"cri-o","ID":"550fc17784cbed0030f262eb4233588b6160edd99bfa3b03eaec56012f272639"} pod="openstack/horizon-68c6b67864-9msm5" containerMessage="Container horizon failed startup probe, will be restarted" Oct 05 21:12:40 crc kubenswrapper[4754]: I1005 21:12:40.609484 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-68c6b67864-9msm5" podUID="27cd6be5-f5eb-4158-bdd1-37f4df3d530d" containerName="horizon" containerID="cri-o://550fc17784cbed0030f262eb4233588b6160edd99bfa3b03eaec56012f272639" gracePeriod=30 Oct 05 21:12:40 crc kubenswrapper[4754]: I1005 21:12:40.725669 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-574454d6cb-vbnk6" podUID="39d1aac3-14a1-45bf-9a9d-ba4cdb0afcb1" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.149:8443/dashboard/auth/login/?next=/dashboard/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 05 21:12:40 crc kubenswrapper[4754]: I1005 21:12:40.725775 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-574454d6cb-vbnk6" Oct 05 21:12:40 crc kubenswrapper[4754]: I1005 21:12:40.726642 4754 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="horizon" containerStatusID={"Type":"cri-o","ID":"4c8ff8949b109b6ab42fd3b43db16a006408ec8d790928e7785deedd5ef7d473"} pod="openstack/horizon-574454d6cb-vbnk6" containerMessage="Container horizon failed startup probe, will be restarted" Oct 05 21:12:40 crc kubenswrapper[4754]: I1005 21:12:40.726695 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-574454d6cb-vbnk6" podUID="39d1aac3-14a1-45bf-9a9d-ba4cdb0afcb1" containerName="horizon" containerID="cri-o://4c8ff8949b109b6ab42fd3b43db16a006408ec8d790928e7785deedd5ef7d473" gracePeriod=30 Oct 05 21:12:44 crc kubenswrapper[4754]: I1005 21:12:44.799856 4754 generic.go:334] "Generic (PLEG): container finished" podID="39d1aac3-14a1-45bf-9a9d-ba4cdb0afcb1" containerID="4c8ff8949b109b6ab42fd3b43db16a006408ec8d790928e7785deedd5ef7d473" exitCode=0 Oct 05 21:12:44 crc kubenswrapper[4754]: I1005 21:12:44.799956 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-574454d6cb-vbnk6" event={"ID":"39d1aac3-14a1-45bf-9a9d-ba4cdb0afcb1","Type":"ContainerDied","Data":"4c8ff8949b109b6ab42fd3b43db16a006408ec8d790928e7785deedd5ef7d473"} Oct 05 21:12:44 crc kubenswrapper[4754]: I1005 21:12:44.804278 4754 generic.go:334] "Generic (PLEG): container finished" podID="27cd6be5-f5eb-4158-bdd1-37f4df3d530d" containerID="550fc17784cbed0030f262eb4233588b6160edd99bfa3b03eaec56012f272639" exitCode=0 Oct 05 21:12:44 crc kubenswrapper[4754]: I1005 21:12:44.804365 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-68c6b67864-9msm5" event={"ID":"27cd6be5-f5eb-4158-bdd1-37f4df3d530d","Type":"ContainerDied","Data":"550fc17784cbed0030f262eb4233588b6160edd99bfa3b03eaec56012f272639"} Oct 05 21:12:48 crc kubenswrapper[4754]: E1005 21:12:48.502563 4754 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-placement-api:current-podified" Oct 05 21:12:48 crc kubenswrapper[4754]: E1005 21:12:48.503464 4754 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:placement-db-sync,Image:quay.io/podified-antelope-centos9/openstack-placement-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/placement,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:placement-dbsync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-v775j,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42482,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-db-sync-xlw68_openstack(94c73d9b-e224-4788-995d-be11dcee7b3c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 05 21:12:48 crc kubenswrapper[4754]: E1005 21:12:48.505793 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/placement-db-sync-xlw68" podUID="94c73d9b-e224-4788-995d-be11dcee7b3c" Oct 05 21:12:50 crc kubenswrapper[4754]: E1005 21:12:50.325642 4754 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Oct 05 21:12:50 crc kubenswrapper[4754]: E1005 21:12:50.326953 4754 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-tvn5l,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-xncdg_openstack(94d0ae23-e26d-4bca-b7f3-e4af07ac2722): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 05 21:12:50 crc kubenswrapper[4754]: E1005 21:12:50.328156 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-xncdg" podUID="94d0ae23-e26d-4bca-b7f3-e4af07ac2722" Oct 05 21:12:50 crc kubenswrapper[4754]: E1005 21:12:50.844164 4754 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified" Oct 05 21:12:50 crc kubenswrapper[4754]: E1005 21:12:50.844639 4754 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-78rcw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-6wlrw_openstack(99975fd1-2c5d-4a2d-9790-0ef13f9fc3f9): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 05 21:12:50 crc kubenswrapper[4754]: E1005 21:12:50.847064 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-6wlrw" podUID="99975fd1-2c5d-4a2d-9790-0ef13f9fc3f9" Oct 05 21:12:50 crc kubenswrapper[4754]: I1005 21:12:50.918938 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-689b67764f-n4l6h" event={"ID":"eea94483-ef0a-4518-8258-587ee1112608","Type":"ContainerDied","Data":"8949f2011a09fc96d23dd205200b33c72f26c7eec65ceea8683e2441f452b32c"} Oct 05 21:12:50 crc kubenswrapper[4754]: I1005 21:12:50.919447 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8949f2011a09fc96d23dd205200b33c72f26c7eec65ceea8683e2441f452b32c" Oct 05 21:12:50 crc kubenswrapper[4754]: I1005 21:12:50.929110 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6d595fb7c-nnfgp" event={"ID":"6f47b134-d55b-454d-8084-7676eb153105","Type":"ContainerDied","Data":"72ca63d3e5da90693a056fdb5f8a671f27ace95c9927768e78c24b40b7af13d3"} Oct 05 21:12:50 crc kubenswrapper[4754]: I1005 21:12:50.931186 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="72ca63d3e5da90693a056fdb5f8a671f27ace95c9927768e78c24b40b7af13d3" Oct 05 21:12:50 crc kubenswrapper[4754]: E1005 21:12:50.941129 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="openstack/cinder-db-sync-xncdg" podUID="94d0ae23-e26d-4bca-b7f3-e4af07ac2722" Oct 05 21:12:50 crc kubenswrapper[4754]: E1005 21:12:50.941963 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified\\\"\"" pod="openstack/barbican-db-sync-6wlrw" podUID="99975fd1-2c5d-4a2d-9790-0ef13f9fc3f9" Oct 05 21:12:50 crc kubenswrapper[4754]: I1005 21:12:50.992120 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-689b67764f-n4l6h" Oct 05 21:12:50 crc kubenswrapper[4754]: I1005 21:12:50.993202 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6d595fb7c-nnfgp" Oct 05 21:12:51 crc kubenswrapper[4754]: I1005 21:12:51.068244 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/6f47b134-d55b-454d-8084-7676eb153105-horizon-secret-key\") pod \"6f47b134-d55b-454d-8084-7676eb153105\" (UID: \"6f47b134-d55b-454d-8084-7676eb153105\") " Oct 05 21:12:51 crc kubenswrapper[4754]: I1005 21:12:51.068328 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6f47b134-d55b-454d-8084-7676eb153105-scripts\") pod \"6f47b134-d55b-454d-8084-7676eb153105\" (UID: \"6f47b134-d55b-454d-8084-7676eb153105\") " Oct 05 21:12:51 crc kubenswrapper[4754]: I1005 21:12:51.068389 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/eea94483-ef0a-4518-8258-587ee1112608-logs\") pod \"eea94483-ef0a-4518-8258-587ee1112608\" (UID: \"eea94483-ef0a-4518-8258-587ee1112608\") " Oct 05 21:12:51 crc kubenswrapper[4754]: I1005 21:12:51.068436 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hrgbh\" (UniqueName: \"kubernetes.io/projected/eea94483-ef0a-4518-8258-587ee1112608-kube-api-access-hrgbh\") pod \"eea94483-ef0a-4518-8258-587ee1112608\" (UID: \"eea94483-ef0a-4518-8258-587ee1112608\") " Oct 05 21:12:51 crc kubenswrapper[4754]: I1005 21:12:51.068487 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/eea94483-ef0a-4518-8258-587ee1112608-horizon-secret-key\") pod \"eea94483-ef0a-4518-8258-587ee1112608\" (UID: \"eea94483-ef0a-4518-8258-587ee1112608\") " Oct 05 21:12:51 crc kubenswrapper[4754]: I1005 21:12:51.068537 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/eea94483-ef0a-4518-8258-587ee1112608-scripts\") pod \"eea94483-ef0a-4518-8258-587ee1112608\" (UID: \"eea94483-ef0a-4518-8258-587ee1112608\") " Oct 05 21:12:51 crc kubenswrapper[4754]: I1005 21:12:51.068564 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6f47b134-d55b-454d-8084-7676eb153105-logs\") pod \"6f47b134-d55b-454d-8084-7676eb153105\" (UID: \"6f47b134-d55b-454d-8084-7676eb153105\") " Oct 05 21:12:51 crc kubenswrapper[4754]: I1005 21:12:51.068625 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/eea94483-ef0a-4518-8258-587ee1112608-config-data\") pod \"eea94483-ef0a-4518-8258-587ee1112608\" (UID: \"eea94483-ef0a-4518-8258-587ee1112608\") " Oct 05 21:12:51 crc kubenswrapper[4754]: I1005 21:12:51.068667 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6f47b134-d55b-454d-8084-7676eb153105-config-data\") pod \"6f47b134-d55b-454d-8084-7676eb153105\" (UID: \"6f47b134-d55b-454d-8084-7676eb153105\") " Oct 05 21:12:51 crc kubenswrapper[4754]: I1005 21:12:51.068704 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n2bwx\" (UniqueName: \"kubernetes.io/projected/6f47b134-d55b-454d-8084-7676eb153105-kube-api-access-n2bwx\") pod \"6f47b134-d55b-454d-8084-7676eb153105\" (UID: \"6f47b134-d55b-454d-8084-7676eb153105\") " Oct 05 21:12:51 crc kubenswrapper[4754]: I1005 21:12:51.072198 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6f47b134-d55b-454d-8084-7676eb153105-logs" (OuterVolumeSpecName: "logs") pod "6f47b134-d55b-454d-8084-7676eb153105" (UID: "6f47b134-d55b-454d-8084-7676eb153105"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 21:12:51 crc kubenswrapper[4754]: I1005 21:12:51.073640 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eea94483-ef0a-4518-8258-587ee1112608-logs" (OuterVolumeSpecName: "logs") pod "eea94483-ef0a-4518-8258-587ee1112608" (UID: "eea94483-ef0a-4518-8258-587ee1112608"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 21:12:51 crc kubenswrapper[4754]: I1005 21:12:51.079961 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eea94483-ef0a-4518-8258-587ee1112608-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "eea94483-ef0a-4518-8258-587ee1112608" (UID: "eea94483-ef0a-4518-8258-587ee1112608"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:12:51 crc kubenswrapper[4754]: I1005 21:12:51.086936 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6f47b134-d55b-454d-8084-7676eb153105-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "6f47b134-d55b-454d-8084-7676eb153105" (UID: "6f47b134-d55b-454d-8084-7676eb153105"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:12:51 crc kubenswrapper[4754]: I1005 21:12:51.087175 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eea94483-ef0a-4518-8258-587ee1112608-kube-api-access-hrgbh" (OuterVolumeSpecName: "kube-api-access-hrgbh") pod "eea94483-ef0a-4518-8258-587ee1112608" (UID: "eea94483-ef0a-4518-8258-587ee1112608"). InnerVolumeSpecName "kube-api-access-hrgbh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:12:51 crc kubenswrapper[4754]: I1005 21:12:51.088286 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6f47b134-d55b-454d-8084-7676eb153105-kube-api-access-n2bwx" (OuterVolumeSpecName: "kube-api-access-n2bwx") pod "6f47b134-d55b-454d-8084-7676eb153105" (UID: "6f47b134-d55b-454d-8084-7676eb153105"). InnerVolumeSpecName "kube-api-access-n2bwx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:12:51 crc kubenswrapper[4754]: I1005 21:12:51.153986 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6f47b134-d55b-454d-8084-7676eb153105-scripts" (OuterVolumeSpecName: "scripts") pod "6f47b134-d55b-454d-8084-7676eb153105" (UID: "6f47b134-d55b-454d-8084-7676eb153105"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 21:12:51 crc kubenswrapper[4754]: I1005 21:12:51.160340 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eea94483-ef0a-4518-8258-587ee1112608-config-data" (OuterVolumeSpecName: "config-data") pod "eea94483-ef0a-4518-8258-587ee1112608" (UID: "eea94483-ef0a-4518-8258-587ee1112608"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 21:12:51 crc kubenswrapper[4754]: I1005 21:12:51.161004 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6f47b134-d55b-454d-8084-7676eb153105-config-data" (OuterVolumeSpecName: "config-data") pod "6f47b134-d55b-454d-8084-7676eb153105" (UID: "6f47b134-d55b-454d-8084-7676eb153105"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 21:12:51 crc kubenswrapper[4754]: I1005 21:12:51.171274 4754 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/eea94483-ef0a-4518-8258-587ee1112608-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 21:12:51 crc kubenswrapper[4754]: I1005 21:12:51.171310 4754 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6f47b134-d55b-454d-8084-7676eb153105-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 21:12:51 crc kubenswrapper[4754]: I1005 21:12:51.171321 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n2bwx\" (UniqueName: \"kubernetes.io/projected/6f47b134-d55b-454d-8084-7676eb153105-kube-api-access-n2bwx\") on node \"crc\" DevicePath \"\"" Oct 05 21:12:51 crc kubenswrapper[4754]: I1005 21:12:51.171333 4754 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/6f47b134-d55b-454d-8084-7676eb153105-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 05 21:12:51 crc kubenswrapper[4754]: I1005 21:12:51.171342 4754 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6f47b134-d55b-454d-8084-7676eb153105-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 21:12:51 crc kubenswrapper[4754]: I1005 21:12:51.171350 4754 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/eea94483-ef0a-4518-8258-587ee1112608-logs\") on node \"crc\" DevicePath \"\"" Oct 05 21:12:51 crc kubenswrapper[4754]: I1005 21:12:51.171358 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hrgbh\" (UniqueName: \"kubernetes.io/projected/eea94483-ef0a-4518-8258-587ee1112608-kube-api-access-hrgbh\") on node \"crc\" DevicePath \"\"" Oct 05 21:12:51 crc kubenswrapper[4754]: I1005 21:12:51.171368 4754 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/eea94483-ef0a-4518-8258-587ee1112608-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 05 21:12:51 crc kubenswrapper[4754]: I1005 21:12:51.171376 4754 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6f47b134-d55b-454d-8084-7676eb153105-logs\") on node \"crc\" DevicePath \"\"" Oct 05 21:12:51 crc kubenswrapper[4754]: I1005 21:12:51.174296 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eea94483-ef0a-4518-8258-587ee1112608-scripts" (OuterVolumeSpecName: "scripts") pod "eea94483-ef0a-4518-8258-587ee1112608" (UID: "eea94483-ef0a-4518-8258-587ee1112608"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 21:12:51 crc kubenswrapper[4754]: I1005 21:12:51.274446 4754 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/eea94483-ef0a-4518-8258-587ee1112608-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 21:12:51 crc kubenswrapper[4754]: I1005 21:12:51.949126 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-574454d6cb-vbnk6" event={"ID":"39d1aac3-14a1-45bf-9a9d-ba4cdb0afcb1","Type":"ContainerStarted","Data":"7ebc5fb1dedef86d2a80ae2a6e400e3f44aaa47d3b94955c70a007eb48092a6a"} Oct 05 21:12:51 crc kubenswrapper[4754]: I1005 21:12:51.952306 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-55bcb5dcdd-jhgrw" event={"ID":"d4c2091b-e2d2-4861-8190-79d548039f18","Type":"ContainerStarted","Data":"10a1586a056f8707b3b49719bb5e8d84656a63c4c448d8ea167b806b48376c9b"} Oct 05 21:12:51 crc kubenswrapper[4754]: I1005 21:12:51.953583 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-55bcb5dcdd-jhgrw" Oct 05 21:12:51 crc kubenswrapper[4754]: I1005 21:12:51.957564 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-68c6b67864-9msm5" event={"ID":"27cd6be5-f5eb-4158-bdd1-37f4df3d530d","Type":"ContainerStarted","Data":"7d14664c74062c1ff554d9b5e811703f4db1749cb0131849ed4d6056fb07bc02"} Oct 05 21:12:51 crc kubenswrapper[4754]: I1005 21:12:51.961273 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-689b67764f-n4l6h" Oct 05 21:12:51 crc kubenswrapper[4754]: I1005 21:12:51.961292 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6d595fb7c-nnfgp" Oct 05 21:12:51 crc kubenswrapper[4754]: I1005 21:12:51.961281 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0a10eb04-c1a7-447a-9400-87b39b1befdf","Type":"ContainerStarted","Data":"bf230f1ca8b629d9031d88f12f9a25424a422a10bc1317c68aae6b99c5dee27a"} Oct 05 21:12:52 crc kubenswrapper[4754]: I1005 21:12:52.019115 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-55bcb5dcdd-jhgrw" podStartSLOduration=39.019096932 podStartE2EDuration="39.019096932s" podCreationTimestamp="2025-10-05 21:12:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 21:12:52.015370284 +0000 UTC m=+1095.919489014" watchObservedRunningTime="2025-10-05 21:12:52.019096932 +0000 UTC m=+1095.923215642" Oct 05 21:12:52 crc kubenswrapper[4754]: I1005 21:12:52.050825 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-689b67764f-n4l6h"] Oct 05 21:12:52 crc kubenswrapper[4754]: I1005 21:12:52.062992 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-689b67764f-n4l6h"] Oct 05 21:12:52 crc kubenswrapper[4754]: I1005 21:12:52.079764 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-6d595fb7c-nnfgp"] Oct 05 21:12:52 crc kubenswrapper[4754]: I1005 21:12:52.093676 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-6d595fb7c-nnfgp"] Oct 05 21:12:52 crc kubenswrapper[4754]: I1005 21:12:52.851066 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6f47b134-d55b-454d-8084-7676eb153105" path="/var/lib/kubelet/pods/6f47b134-d55b-454d-8084-7676eb153105/volumes" Oct 05 21:12:52 crc kubenswrapper[4754]: I1005 21:12:52.852144 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eea94483-ef0a-4518-8258-587ee1112608" path="/var/lib/kubelet/pods/eea94483-ef0a-4518-8258-587ee1112608/volumes" Oct 05 21:12:55 crc kubenswrapper[4754]: I1005 21:12:55.603747 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-68c6b67864-9msm5" Oct 05 21:12:55 crc kubenswrapper[4754]: I1005 21:12:55.605271 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-68c6b67864-9msm5" Oct 05 21:12:55 crc kubenswrapper[4754]: I1005 21:12:55.721306 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-574454d6cb-vbnk6" Oct 05 21:12:55 crc kubenswrapper[4754]: I1005 21:12:55.721362 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-574454d6cb-vbnk6" Oct 05 21:13:03 crc kubenswrapper[4754]: E1005 21:13:03.841771 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-placement-api:current-podified\\\"\"" pod="openstack/placement-db-sync-xlw68" podUID="94c73d9b-e224-4788-995d-be11dcee7b3c" Oct 05 21:13:05 crc kubenswrapper[4754]: I1005 21:13:05.244840 4754 patch_prober.go:28] interesting pod/machine-config-daemon-b2h9k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 21:13:05 crc kubenswrapper[4754]: I1005 21:13:05.245296 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 21:13:05 crc kubenswrapper[4754]: I1005 21:13:05.605541 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-68c6b67864-9msm5" podUID="27cd6be5-f5eb-4158-bdd1-37f4df3d530d" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.148:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.148:8443: connect: connection refused" Oct 05 21:13:05 crc kubenswrapper[4754]: I1005 21:13:05.724553 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-574454d6cb-vbnk6" podUID="39d1aac3-14a1-45bf-9a9d-ba4cdb0afcb1" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.149:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.149:8443: connect: connection refused" Oct 05 21:13:09 crc kubenswrapper[4754]: I1005 21:13:09.179870 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-6wlrw" event={"ID":"99975fd1-2c5d-4a2d-9790-0ef13f9fc3f9","Type":"ContainerStarted","Data":"4302b6bd8251e7a7c27e4b9908256f84d6cc923b3112ce45c885fc471fb96198"} Oct 05 21:13:09 crc kubenswrapper[4754]: I1005 21:13:09.199523 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0a10eb04-c1a7-447a-9400-87b39b1befdf","Type":"ContainerStarted","Data":"79984ab6a1f39eb83bb47a3828121ca61cdf9f93b2fe4295ce4a4275cdc9ff81"} Oct 05 21:13:09 crc kubenswrapper[4754]: I1005 21:13:09.199801 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0a10eb04-c1a7-447a-9400-87b39b1befdf" containerName="ceilometer-central-agent" containerID="cri-o://ec5e8223346ee4ff9e8367f4a328b561205c831c46f42c18f05917162be8e6a3" gracePeriod=30 Oct 05 21:13:09 crc kubenswrapper[4754]: I1005 21:13:09.200052 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 05 21:13:09 crc kubenswrapper[4754]: I1005 21:13:09.200186 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0a10eb04-c1a7-447a-9400-87b39b1befdf" containerName="proxy-httpd" containerID="cri-o://79984ab6a1f39eb83bb47a3828121ca61cdf9f93b2fe4295ce4a4275cdc9ff81" gracePeriod=30 Oct 05 21:13:09 crc kubenswrapper[4754]: I1005 21:13:09.200208 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0a10eb04-c1a7-447a-9400-87b39b1befdf" containerName="ceilometer-notification-agent" containerID="cri-o://83a4f5b028fd5d8370357e2ce7359bcdbe5419e295f9c87528c4827687035543" gracePeriod=30 Oct 05 21:13:09 crc kubenswrapper[4754]: I1005 21:13:09.200090 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0a10eb04-c1a7-447a-9400-87b39b1befdf" containerName="sg-core" containerID="cri-o://bf230f1ca8b629d9031d88f12f9a25424a422a10bc1317c68aae6b99c5dee27a" gracePeriod=30 Oct 05 21:13:09 crc kubenswrapper[4754]: I1005 21:13:09.214049 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-xncdg" event={"ID":"94d0ae23-e26d-4bca-b7f3-e4af07ac2722","Type":"ContainerStarted","Data":"907c4f77f15040be8d5f56e9c1085880e54be3a78590fa12671e33449f1ded81"} Oct 05 21:13:09 crc kubenswrapper[4754]: I1005 21:13:09.214211 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-6wlrw" podStartSLOduration=2.700309262 podStartE2EDuration="58.214162025s" podCreationTimestamp="2025-10-05 21:12:11 +0000 UTC" firstStartedPulling="2025-10-05 21:12:12.328575286 +0000 UTC m=+1056.232693996" lastFinishedPulling="2025-10-05 21:13:07.842428049 +0000 UTC m=+1111.746546759" observedRunningTime="2025-10-05 21:13:09.20297796 +0000 UTC m=+1113.107096670" watchObservedRunningTime="2025-10-05 21:13:09.214162025 +0000 UTC m=+1113.118280735" Oct 05 21:13:09 crc kubenswrapper[4754]: I1005 21:13:09.242636 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.585035133 podStartE2EDuration="1m27.242613723s" podCreationTimestamp="2025-10-05 21:11:42 +0000 UTC" firstStartedPulling="2025-10-05 21:11:44.206078318 +0000 UTC m=+1028.110197028" lastFinishedPulling="2025-10-05 21:13:07.863656908 +0000 UTC m=+1111.767775618" observedRunningTime="2025-10-05 21:13:09.236722508 +0000 UTC m=+1113.140841218" watchObservedRunningTime="2025-10-05 21:13:09.242613723 +0000 UTC m=+1113.146732433" Oct 05 21:13:09 crc kubenswrapper[4754]: I1005 21:13:09.257958 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-xncdg" podStartSLOduration=14.706517432 podStartE2EDuration="1m19.257943247s" podCreationTimestamp="2025-10-05 21:11:50 +0000 UTC" firstStartedPulling="2025-10-05 21:12:03.301896361 +0000 UTC m=+1047.206015061" lastFinishedPulling="2025-10-05 21:13:07.853322166 +0000 UTC m=+1111.757440876" observedRunningTime="2025-10-05 21:13:09.256631302 +0000 UTC m=+1113.160750002" watchObservedRunningTime="2025-10-05 21:13:09.257943247 +0000 UTC m=+1113.162061957" Oct 05 21:13:10 crc kubenswrapper[4754]: I1005 21:13:10.226513 4754 generic.go:334] "Generic (PLEG): container finished" podID="0a10eb04-c1a7-447a-9400-87b39b1befdf" containerID="79984ab6a1f39eb83bb47a3828121ca61cdf9f93b2fe4295ce4a4275cdc9ff81" exitCode=0 Oct 05 21:13:10 crc kubenswrapper[4754]: I1005 21:13:10.226559 4754 generic.go:334] "Generic (PLEG): container finished" podID="0a10eb04-c1a7-447a-9400-87b39b1befdf" containerID="bf230f1ca8b629d9031d88f12f9a25424a422a10bc1317c68aae6b99c5dee27a" exitCode=2 Oct 05 21:13:10 crc kubenswrapper[4754]: I1005 21:13:10.226567 4754 generic.go:334] "Generic (PLEG): container finished" podID="0a10eb04-c1a7-447a-9400-87b39b1befdf" containerID="ec5e8223346ee4ff9e8367f4a328b561205c831c46f42c18f05917162be8e6a3" exitCode=0 Oct 05 21:13:10 crc kubenswrapper[4754]: I1005 21:13:10.226609 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0a10eb04-c1a7-447a-9400-87b39b1befdf","Type":"ContainerDied","Data":"79984ab6a1f39eb83bb47a3828121ca61cdf9f93b2fe4295ce4a4275cdc9ff81"} Oct 05 21:13:10 crc kubenswrapper[4754]: I1005 21:13:10.226689 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0a10eb04-c1a7-447a-9400-87b39b1befdf","Type":"ContainerDied","Data":"bf230f1ca8b629d9031d88f12f9a25424a422a10bc1317c68aae6b99c5dee27a"} Oct 05 21:13:10 crc kubenswrapper[4754]: I1005 21:13:10.226705 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0a10eb04-c1a7-447a-9400-87b39b1befdf","Type":"ContainerDied","Data":"ec5e8223346ee4ff9e8367f4a328b561205c831c46f42c18f05917162be8e6a3"} Oct 05 21:13:11 crc kubenswrapper[4754]: I1005 21:13:11.249940 4754 generic.go:334] "Generic (PLEG): container finished" podID="0a10eb04-c1a7-447a-9400-87b39b1befdf" containerID="83a4f5b028fd5d8370357e2ce7359bcdbe5419e295f9c87528c4827687035543" exitCode=0 Oct 05 21:13:11 crc kubenswrapper[4754]: I1005 21:13:11.250206 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0a10eb04-c1a7-447a-9400-87b39b1befdf","Type":"ContainerDied","Data":"83a4f5b028fd5d8370357e2ce7359bcdbe5419e295f9c87528c4827687035543"} Oct 05 21:13:11 crc kubenswrapper[4754]: I1005 21:13:11.655212 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 05 21:13:11 crc kubenswrapper[4754]: I1005 21:13:11.732645 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a10eb04-c1a7-447a-9400-87b39b1befdf-combined-ca-bundle\") pod \"0a10eb04-c1a7-447a-9400-87b39b1befdf\" (UID: \"0a10eb04-c1a7-447a-9400-87b39b1befdf\") " Oct 05 21:13:11 crc kubenswrapper[4754]: I1005 21:13:11.732729 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0a10eb04-c1a7-447a-9400-87b39b1befdf-log-httpd\") pod \"0a10eb04-c1a7-447a-9400-87b39b1befdf\" (UID: \"0a10eb04-c1a7-447a-9400-87b39b1befdf\") " Oct 05 21:13:11 crc kubenswrapper[4754]: I1005 21:13:11.732774 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l7k9j\" (UniqueName: \"kubernetes.io/projected/0a10eb04-c1a7-447a-9400-87b39b1befdf-kube-api-access-l7k9j\") pod \"0a10eb04-c1a7-447a-9400-87b39b1befdf\" (UID: \"0a10eb04-c1a7-447a-9400-87b39b1befdf\") " Oct 05 21:13:11 crc kubenswrapper[4754]: I1005 21:13:11.732850 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a10eb04-c1a7-447a-9400-87b39b1befdf-config-data\") pod \"0a10eb04-c1a7-447a-9400-87b39b1befdf\" (UID: \"0a10eb04-c1a7-447a-9400-87b39b1befdf\") " Oct 05 21:13:11 crc kubenswrapper[4754]: I1005 21:13:11.732887 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0a10eb04-c1a7-447a-9400-87b39b1befdf-run-httpd\") pod \"0a10eb04-c1a7-447a-9400-87b39b1befdf\" (UID: \"0a10eb04-c1a7-447a-9400-87b39b1befdf\") " Oct 05 21:13:11 crc kubenswrapper[4754]: I1005 21:13:11.732912 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0a10eb04-c1a7-447a-9400-87b39b1befdf-scripts\") pod \"0a10eb04-c1a7-447a-9400-87b39b1befdf\" (UID: \"0a10eb04-c1a7-447a-9400-87b39b1befdf\") " Oct 05 21:13:11 crc kubenswrapper[4754]: I1005 21:13:11.733015 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0a10eb04-c1a7-447a-9400-87b39b1befdf-sg-core-conf-yaml\") pod \"0a10eb04-c1a7-447a-9400-87b39b1befdf\" (UID: \"0a10eb04-c1a7-447a-9400-87b39b1befdf\") " Oct 05 21:13:11 crc kubenswrapper[4754]: I1005 21:13:11.733814 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0a10eb04-c1a7-447a-9400-87b39b1befdf-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "0a10eb04-c1a7-447a-9400-87b39b1befdf" (UID: "0a10eb04-c1a7-447a-9400-87b39b1befdf"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 21:13:11 crc kubenswrapper[4754]: I1005 21:13:11.733972 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0a10eb04-c1a7-447a-9400-87b39b1befdf-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "0a10eb04-c1a7-447a-9400-87b39b1befdf" (UID: "0a10eb04-c1a7-447a-9400-87b39b1befdf"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 21:13:11 crc kubenswrapper[4754]: I1005 21:13:11.747191 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a10eb04-c1a7-447a-9400-87b39b1befdf-scripts" (OuterVolumeSpecName: "scripts") pod "0a10eb04-c1a7-447a-9400-87b39b1befdf" (UID: "0a10eb04-c1a7-447a-9400-87b39b1befdf"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:13:11 crc kubenswrapper[4754]: I1005 21:13:11.749812 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0a10eb04-c1a7-447a-9400-87b39b1befdf-kube-api-access-l7k9j" (OuterVolumeSpecName: "kube-api-access-l7k9j") pod "0a10eb04-c1a7-447a-9400-87b39b1befdf" (UID: "0a10eb04-c1a7-447a-9400-87b39b1befdf"). InnerVolumeSpecName "kube-api-access-l7k9j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:13:11 crc kubenswrapper[4754]: I1005 21:13:11.835274 4754 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0a10eb04-c1a7-447a-9400-87b39b1befdf-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 05 21:13:11 crc kubenswrapper[4754]: I1005 21:13:11.835322 4754 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0a10eb04-c1a7-447a-9400-87b39b1befdf-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 21:13:11 crc kubenswrapper[4754]: I1005 21:13:11.835331 4754 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0a10eb04-c1a7-447a-9400-87b39b1befdf-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 05 21:13:11 crc kubenswrapper[4754]: I1005 21:13:11.835341 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l7k9j\" (UniqueName: \"kubernetes.io/projected/0a10eb04-c1a7-447a-9400-87b39b1befdf-kube-api-access-l7k9j\") on node \"crc\" DevicePath \"\"" Oct 05 21:13:11 crc kubenswrapper[4754]: I1005 21:13:11.839282 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a10eb04-c1a7-447a-9400-87b39b1befdf-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0a10eb04-c1a7-447a-9400-87b39b1befdf" (UID: "0a10eb04-c1a7-447a-9400-87b39b1befdf"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:13:11 crc kubenswrapper[4754]: I1005 21:13:11.843648 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a10eb04-c1a7-447a-9400-87b39b1befdf-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "0a10eb04-c1a7-447a-9400-87b39b1befdf" (UID: "0a10eb04-c1a7-447a-9400-87b39b1befdf"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:13:11 crc kubenswrapper[4754]: I1005 21:13:11.869560 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a10eb04-c1a7-447a-9400-87b39b1befdf-config-data" (OuterVolumeSpecName: "config-data") pod "0a10eb04-c1a7-447a-9400-87b39b1befdf" (UID: "0a10eb04-c1a7-447a-9400-87b39b1befdf"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:13:11 crc kubenswrapper[4754]: I1005 21:13:11.936846 4754 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a10eb04-c1a7-447a-9400-87b39b1befdf-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 21:13:11 crc kubenswrapper[4754]: I1005 21:13:11.936885 4754 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0a10eb04-c1a7-447a-9400-87b39b1befdf-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 05 21:13:11 crc kubenswrapper[4754]: I1005 21:13:11.936899 4754 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a10eb04-c1a7-447a-9400-87b39b1befdf-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 21:13:12 crc kubenswrapper[4754]: I1005 21:13:12.285106 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0a10eb04-c1a7-447a-9400-87b39b1befdf","Type":"ContainerDied","Data":"656f92069277b501ccc8771c65b67ba371b0110cf87ab9be74b51aa3eacf78b6"} Oct 05 21:13:12 crc kubenswrapper[4754]: I1005 21:13:12.285172 4754 scope.go:117] "RemoveContainer" containerID="79984ab6a1f39eb83bb47a3828121ca61cdf9f93b2fe4295ce4a4275cdc9ff81" Oct 05 21:13:12 crc kubenswrapper[4754]: I1005 21:13:12.285210 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 05 21:13:12 crc kubenswrapper[4754]: I1005 21:13:12.345267 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 05 21:13:12 crc kubenswrapper[4754]: I1005 21:13:12.346542 4754 scope.go:117] "RemoveContainer" containerID="bf230f1ca8b629d9031d88f12f9a25424a422a10bc1317c68aae6b99c5dee27a" Oct 05 21:13:12 crc kubenswrapper[4754]: I1005 21:13:12.360398 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 05 21:13:12 crc kubenswrapper[4754]: I1005 21:13:12.397932 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 05 21:13:12 crc kubenswrapper[4754]: E1005 21:13:12.398971 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f47b134-d55b-454d-8084-7676eb153105" containerName="horizon" Oct 05 21:13:12 crc kubenswrapper[4754]: I1005 21:13:12.398994 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f47b134-d55b-454d-8084-7676eb153105" containerName="horizon" Oct 05 21:13:12 crc kubenswrapper[4754]: E1005 21:13:12.399007 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a10eb04-c1a7-447a-9400-87b39b1befdf" containerName="ceilometer-notification-agent" Oct 05 21:13:12 crc kubenswrapper[4754]: I1005 21:13:12.399016 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a10eb04-c1a7-447a-9400-87b39b1befdf" containerName="ceilometer-notification-agent" Oct 05 21:13:12 crc kubenswrapper[4754]: E1005 21:13:12.399033 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a10eb04-c1a7-447a-9400-87b39b1befdf" containerName="sg-core" Oct 05 21:13:12 crc kubenswrapper[4754]: I1005 21:13:12.399039 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a10eb04-c1a7-447a-9400-87b39b1befdf" containerName="sg-core" Oct 05 21:13:12 crc kubenswrapper[4754]: E1005 21:13:12.399054 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a10eb04-c1a7-447a-9400-87b39b1befdf" containerName="ceilometer-central-agent" Oct 05 21:13:12 crc kubenswrapper[4754]: I1005 21:13:12.399060 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a10eb04-c1a7-447a-9400-87b39b1befdf" containerName="ceilometer-central-agent" Oct 05 21:13:12 crc kubenswrapper[4754]: E1005 21:13:12.399073 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a10eb04-c1a7-447a-9400-87b39b1befdf" containerName="proxy-httpd" Oct 05 21:13:12 crc kubenswrapper[4754]: I1005 21:13:12.399080 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a10eb04-c1a7-447a-9400-87b39b1befdf" containerName="proxy-httpd" Oct 05 21:13:12 crc kubenswrapper[4754]: E1005 21:13:12.399097 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f47b134-d55b-454d-8084-7676eb153105" containerName="horizon-log" Oct 05 21:13:12 crc kubenswrapper[4754]: I1005 21:13:12.399103 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f47b134-d55b-454d-8084-7676eb153105" containerName="horizon-log" Oct 05 21:13:12 crc kubenswrapper[4754]: E1005 21:13:12.399115 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eea94483-ef0a-4518-8258-587ee1112608" containerName="horizon-log" Oct 05 21:13:12 crc kubenswrapper[4754]: I1005 21:13:12.399121 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="eea94483-ef0a-4518-8258-587ee1112608" containerName="horizon-log" Oct 05 21:13:12 crc kubenswrapper[4754]: E1005 21:13:12.399133 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eea94483-ef0a-4518-8258-587ee1112608" containerName="horizon" Oct 05 21:13:12 crc kubenswrapper[4754]: I1005 21:13:12.399142 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="eea94483-ef0a-4518-8258-587ee1112608" containerName="horizon" Oct 05 21:13:12 crc kubenswrapper[4754]: I1005 21:13:12.399322 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a10eb04-c1a7-447a-9400-87b39b1befdf" containerName="sg-core" Oct 05 21:13:12 crc kubenswrapper[4754]: I1005 21:13:12.399337 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a10eb04-c1a7-447a-9400-87b39b1befdf" containerName="ceilometer-central-agent" Oct 05 21:13:12 crc kubenswrapper[4754]: I1005 21:13:12.399347 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="eea94483-ef0a-4518-8258-587ee1112608" containerName="horizon" Oct 05 21:13:12 crc kubenswrapper[4754]: I1005 21:13:12.399362 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="6f47b134-d55b-454d-8084-7676eb153105" containerName="horizon" Oct 05 21:13:12 crc kubenswrapper[4754]: I1005 21:13:12.399371 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a10eb04-c1a7-447a-9400-87b39b1befdf" containerName="ceilometer-notification-agent" Oct 05 21:13:12 crc kubenswrapper[4754]: I1005 21:13:12.399385 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="eea94483-ef0a-4518-8258-587ee1112608" containerName="horizon-log" Oct 05 21:13:12 crc kubenswrapper[4754]: I1005 21:13:12.399395 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="6f47b134-d55b-454d-8084-7676eb153105" containerName="horizon-log" Oct 05 21:13:12 crc kubenswrapper[4754]: I1005 21:13:12.399405 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a10eb04-c1a7-447a-9400-87b39b1befdf" containerName="proxy-httpd" Oct 05 21:13:12 crc kubenswrapper[4754]: I1005 21:13:12.406488 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 05 21:13:12 crc kubenswrapper[4754]: I1005 21:13:12.421941 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 05 21:13:12 crc kubenswrapper[4754]: I1005 21:13:12.422413 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 05 21:13:12 crc kubenswrapper[4754]: I1005 21:13:12.431693 4754 scope.go:117] "RemoveContainer" containerID="83a4f5b028fd5d8370357e2ce7359bcdbe5419e295f9c87528c4827687035543" Oct 05 21:13:12 crc kubenswrapper[4754]: I1005 21:13:12.438900 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 05 21:13:12 crc kubenswrapper[4754]: I1005 21:13:12.447120 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9fbf0155-8666-4b0c-a321-8f6c788b529d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9fbf0155-8666-4b0c-a321-8f6c788b529d\") " pod="openstack/ceilometer-0" Oct 05 21:13:12 crc kubenswrapper[4754]: I1005 21:13:12.447191 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9fbf0155-8666-4b0c-a321-8f6c788b529d-config-data\") pod \"ceilometer-0\" (UID: \"9fbf0155-8666-4b0c-a321-8f6c788b529d\") " pod="openstack/ceilometer-0" Oct 05 21:13:12 crc kubenswrapper[4754]: I1005 21:13:12.447224 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9fbf0155-8666-4b0c-a321-8f6c788b529d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9fbf0155-8666-4b0c-a321-8f6c788b529d\") " pod="openstack/ceilometer-0" Oct 05 21:13:12 crc kubenswrapper[4754]: I1005 21:13:12.447259 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9fbf0155-8666-4b0c-a321-8f6c788b529d-log-httpd\") pod \"ceilometer-0\" (UID: \"9fbf0155-8666-4b0c-a321-8f6c788b529d\") " pod="openstack/ceilometer-0" Oct 05 21:13:12 crc kubenswrapper[4754]: I1005 21:13:12.447323 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9fbf0155-8666-4b0c-a321-8f6c788b529d-scripts\") pod \"ceilometer-0\" (UID: \"9fbf0155-8666-4b0c-a321-8f6c788b529d\") " pod="openstack/ceilometer-0" Oct 05 21:13:12 crc kubenswrapper[4754]: I1005 21:13:12.447345 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9fbf0155-8666-4b0c-a321-8f6c788b529d-run-httpd\") pod \"ceilometer-0\" (UID: \"9fbf0155-8666-4b0c-a321-8f6c788b529d\") " pod="openstack/ceilometer-0" Oct 05 21:13:12 crc kubenswrapper[4754]: I1005 21:13:12.447380 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wpr2t\" (UniqueName: \"kubernetes.io/projected/9fbf0155-8666-4b0c-a321-8f6c788b529d-kube-api-access-wpr2t\") pod \"ceilometer-0\" (UID: \"9fbf0155-8666-4b0c-a321-8f6c788b529d\") " pod="openstack/ceilometer-0" Oct 05 21:13:12 crc kubenswrapper[4754]: I1005 21:13:12.512043 4754 scope.go:117] "RemoveContainer" containerID="ec5e8223346ee4ff9e8367f4a328b561205c831c46f42c18f05917162be8e6a3" Oct 05 21:13:12 crc kubenswrapper[4754]: I1005 21:13:12.548435 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9fbf0155-8666-4b0c-a321-8f6c788b529d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9fbf0155-8666-4b0c-a321-8f6c788b529d\") " pod="openstack/ceilometer-0" Oct 05 21:13:12 crc kubenswrapper[4754]: I1005 21:13:12.548754 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9fbf0155-8666-4b0c-a321-8f6c788b529d-log-httpd\") pod \"ceilometer-0\" (UID: \"9fbf0155-8666-4b0c-a321-8f6c788b529d\") " pod="openstack/ceilometer-0" Oct 05 21:13:12 crc kubenswrapper[4754]: I1005 21:13:12.548898 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9fbf0155-8666-4b0c-a321-8f6c788b529d-scripts\") pod \"ceilometer-0\" (UID: \"9fbf0155-8666-4b0c-a321-8f6c788b529d\") " pod="openstack/ceilometer-0" Oct 05 21:13:12 crc kubenswrapper[4754]: I1005 21:13:12.549045 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9fbf0155-8666-4b0c-a321-8f6c788b529d-run-httpd\") pod \"ceilometer-0\" (UID: \"9fbf0155-8666-4b0c-a321-8f6c788b529d\") " pod="openstack/ceilometer-0" Oct 05 21:13:12 crc kubenswrapper[4754]: I1005 21:13:12.549134 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wpr2t\" (UniqueName: \"kubernetes.io/projected/9fbf0155-8666-4b0c-a321-8f6c788b529d-kube-api-access-wpr2t\") pod \"ceilometer-0\" (UID: \"9fbf0155-8666-4b0c-a321-8f6c788b529d\") " pod="openstack/ceilometer-0" Oct 05 21:13:12 crc kubenswrapper[4754]: I1005 21:13:12.549242 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9fbf0155-8666-4b0c-a321-8f6c788b529d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9fbf0155-8666-4b0c-a321-8f6c788b529d\") " pod="openstack/ceilometer-0" Oct 05 21:13:12 crc kubenswrapper[4754]: I1005 21:13:12.549310 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9fbf0155-8666-4b0c-a321-8f6c788b529d-config-data\") pod \"ceilometer-0\" (UID: \"9fbf0155-8666-4b0c-a321-8f6c788b529d\") " pod="openstack/ceilometer-0" Oct 05 21:13:12 crc kubenswrapper[4754]: I1005 21:13:12.550255 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9fbf0155-8666-4b0c-a321-8f6c788b529d-run-httpd\") pod \"ceilometer-0\" (UID: \"9fbf0155-8666-4b0c-a321-8f6c788b529d\") " pod="openstack/ceilometer-0" Oct 05 21:13:12 crc kubenswrapper[4754]: I1005 21:13:12.553482 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9fbf0155-8666-4b0c-a321-8f6c788b529d-log-httpd\") pod \"ceilometer-0\" (UID: \"9fbf0155-8666-4b0c-a321-8f6c788b529d\") " pod="openstack/ceilometer-0" Oct 05 21:13:12 crc kubenswrapper[4754]: I1005 21:13:12.569262 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9fbf0155-8666-4b0c-a321-8f6c788b529d-config-data\") pod \"ceilometer-0\" (UID: \"9fbf0155-8666-4b0c-a321-8f6c788b529d\") " pod="openstack/ceilometer-0" Oct 05 21:13:12 crc kubenswrapper[4754]: I1005 21:13:12.573319 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9fbf0155-8666-4b0c-a321-8f6c788b529d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9fbf0155-8666-4b0c-a321-8f6c788b529d\") " pod="openstack/ceilometer-0" Oct 05 21:13:12 crc kubenswrapper[4754]: I1005 21:13:12.576383 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9fbf0155-8666-4b0c-a321-8f6c788b529d-scripts\") pod \"ceilometer-0\" (UID: \"9fbf0155-8666-4b0c-a321-8f6c788b529d\") " pod="openstack/ceilometer-0" Oct 05 21:13:12 crc kubenswrapper[4754]: I1005 21:13:12.581289 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9fbf0155-8666-4b0c-a321-8f6c788b529d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9fbf0155-8666-4b0c-a321-8f6c788b529d\") " pod="openstack/ceilometer-0" Oct 05 21:13:12 crc kubenswrapper[4754]: I1005 21:13:12.586375 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wpr2t\" (UniqueName: \"kubernetes.io/projected/9fbf0155-8666-4b0c-a321-8f6c788b529d-kube-api-access-wpr2t\") pod \"ceilometer-0\" (UID: \"9fbf0155-8666-4b0c-a321-8f6c788b529d\") " pod="openstack/ceilometer-0" Oct 05 21:13:12 crc kubenswrapper[4754]: I1005 21:13:12.737905 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 05 21:13:12 crc kubenswrapper[4754]: I1005 21:13:12.852604 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0a10eb04-c1a7-447a-9400-87b39b1befdf" path="/var/lib/kubelet/pods/0a10eb04-c1a7-447a-9400-87b39b1befdf/volumes" Oct 05 21:13:13 crc kubenswrapper[4754]: I1005 21:13:13.297721 4754 generic.go:334] "Generic (PLEG): container finished" podID="99975fd1-2c5d-4a2d-9790-0ef13f9fc3f9" containerID="4302b6bd8251e7a7c27e4b9908256f84d6cc923b3112ce45c885fc471fb96198" exitCode=0 Oct 05 21:13:13 crc kubenswrapper[4754]: I1005 21:13:13.297826 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-6wlrw" event={"ID":"99975fd1-2c5d-4a2d-9790-0ef13f9fc3f9","Type":"ContainerDied","Data":"4302b6bd8251e7a7c27e4b9908256f84d6cc923b3112ce45c885fc471fb96198"} Oct 05 21:13:13 crc kubenswrapper[4754]: I1005 21:13:13.327483 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 05 21:13:14 crc kubenswrapper[4754]: I1005 21:13:14.328606 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9fbf0155-8666-4b0c-a321-8f6c788b529d","Type":"ContainerStarted","Data":"9cba681c08c510789981ad324899d5474bd4c967035d7a4f8ea21e4f81a5b44f"} Oct 05 21:13:14 crc kubenswrapper[4754]: I1005 21:13:14.329083 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9fbf0155-8666-4b0c-a321-8f6c788b529d","Type":"ContainerStarted","Data":"383df62c656a37781ded70118f6132b7393e16d8c2c33715b6ebe1528d5715dc"} Oct 05 21:13:14 crc kubenswrapper[4754]: I1005 21:13:14.791697 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-6wlrw" Oct 05 21:13:14 crc kubenswrapper[4754]: I1005 21:13:14.898031 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-78rcw\" (UniqueName: \"kubernetes.io/projected/99975fd1-2c5d-4a2d-9790-0ef13f9fc3f9-kube-api-access-78rcw\") pod \"99975fd1-2c5d-4a2d-9790-0ef13f9fc3f9\" (UID: \"99975fd1-2c5d-4a2d-9790-0ef13f9fc3f9\") " Oct 05 21:13:14 crc kubenswrapper[4754]: I1005 21:13:14.898206 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99975fd1-2c5d-4a2d-9790-0ef13f9fc3f9-combined-ca-bundle\") pod \"99975fd1-2c5d-4a2d-9790-0ef13f9fc3f9\" (UID: \"99975fd1-2c5d-4a2d-9790-0ef13f9fc3f9\") " Oct 05 21:13:14 crc kubenswrapper[4754]: I1005 21:13:14.898234 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/99975fd1-2c5d-4a2d-9790-0ef13f9fc3f9-db-sync-config-data\") pod \"99975fd1-2c5d-4a2d-9790-0ef13f9fc3f9\" (UID: \"99975fd1-2c5d-4a2d-9790-0ef13f9fc3f9\") " Oct 05 21:13:14 crc kubenswrapper[4754]: I1005 21:13:14.904604 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99975fd1-2c5d-4a2d-9790-0ef13f9fc3f9-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "99975fd1-2c5d-4a2d-9790-0ef13f9fc3f9" (UID: "99975fd1-2c5d-4a2d-9790-0ef13f9fc3f9"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:13:14 crc kubenswrapper[4754]: I1005 21:13:14.905488 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/99975fd1-2c5d-4a2d-9790-0ef13f9fc3f9-kube-api-access-78rcw" (OuterVolumeSpecName: "kube-api-access-78rcw") pod "99975fd1-2c5d-4a2d-9790-0ef13f9fc3f9" (UID: "99975fd1-2c5d-4a2d-9790-0ef13f9fc3f9"). InnerVolumeSpecName "kube-api-access-78rcw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:13:14 crc kubenswrapper[4754]: I1005 21:13:14.962547 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99975fd1-2c5d-4a2d-9790-0ef13f9fc3f9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "99975fd1-2c5d-4a2d-9790-0ef13f9fc3f9" (UID: "99975fd1-2c5d-4a2d-9790-0ef13f9fc3f9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:13:15 crc kubenswrapper[4754]: I1005 21:13:15.003132 4754 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99975fd1-2c5d-4a2d-9790-0ef13f9fc3f9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 21:13:15 crc kubenswrapper[4754]: I1005 21:13:15.003169 4754 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/99975fd1-2c5d-4a2d-9790-0ef13f9fc3f9-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 21:13:15 crc kubenswrapper[4754]: I1005 21:13:15.003181 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-78rcw\" (UniqueName: \"kubernetes.io/projected/99975fd1-2c5d-4a2d-9790-0ef13f9fc3f9-kube-api-access-78rcw\") on node \"crc\" DevicePath \"\"" Oct 05 21:13:15 crc kubenswrapper[4754]: I1005 21:13:15.341971 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-6wlrw" event={"ID":"99975fd1-2c5d-4a2d-9790-0ef13f9fc3f9","Type":"ContainerDied","Data":"54a24f4b6483ff93adcdb0fd84d666cd20b275782362b87c4dc257b6a10e4c56"} Oct 05 21:13:15 crc kubenswrapper[4754]: I1005 21:13:15.342015 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="54a24f4b6483ff93adcdb0fd84d666cd20b275782362b87c4dc257b6a10e4c56" Oct 05 21:13:15 crc kubenswrapper[4754]: I1005 21:13:15.342090 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-6wlrw" Oct 05 21:13:15 crc kubenswrapper[4754]: I1005 21:13:15.346250 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9fbf0155-8666-4b0c-a321-8f6c788b529d","Type":"ContainerStarted","Data":"5cf22c8730a91a7041f82212493156f36557b87917c64440ab525727adffaab2"} Oct 05 21:13:15 crc kubenswrapper[4754]: I1005 21:13:15.580736 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-54b7554787-vqrt4"] Oct 05 21:13:15 crc kubenswrapper[4754]: E1005 21:13:15.581160 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99975fd1-2c5d-4a2d-9790-0ef13f9fc3f9" containerName="barbican-db-sync" Oct 05 21:13:15 crc kubenswrapper[4754]: I1005 21:13:15.581177 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="99975fd1-2c5d-4a2d-9790-0ef13f9fc3f9" containerName="barbican-db-sync" Oct 05 21:13:15 crc kubenswrapper[4754]: I1005 21:13:15.581373 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="99975fd1-2c5d-4a2d-9790-0ef13f9fc3f9" containerName="barbican-db-sync" Oct 05 21:13:15 crc kubenswrapper[4754]: I1005 21:13:15.582322 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-54b7554787-vqrt4" Oct 05 21:13:15 crc kubenswrapper[4754]: I1005 21:13:15.589037 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-nv4lv" Oct 05 21:13:15 crc kubenswrapper[4754]: I1005 21:13:15.589260 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Oct 05 21:13:15 crc kubenswrapper[4754]: I1005 21:13:15.589367 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Oct 05 21:13:15 crc kubenswrapper[4754]: I1005 21:13:15.609869 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-54b7554787-vqrt4"] Oct 05 21:13:15 crc kubenswrapper[4754]: I1005 21:13:15.620848 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/346d8b39-98ea-442d-910e-b40a6e4b85c4-combined-ca-bundle\") pod \"barbican-worker-54b7554787-vqrt4\" (UID: \"346d8b39-98ea-442d-910e-b40a6e4b85c4\") " pod="openstack/barbican-worker-54b7554787-vqrt4" Oct 05 21:13:15 crc kubenswrapper[4754]: I1005 21:13:15.620915 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/346d8b39-98ea-442d-910e-b40a6e4b85c4-logs\") pod \"barbican-worker-54b7554787-vqrt4\" (UID: \"346d8b39-98ea-442d-910e-b40a6e4b85c4\") " pod="openstack/barbican-worker-54b7554787-vqrt4" Oct 05 21:13:15 crc kubenswrapper[4754]: I1005 21:13:15.620947 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/346d8b39-98ea-442d-910e-b40a6e4b85c4-config-data-custom\") pod \"barbican-worker-54b7554787-vqrt4\" (UID: \"346d8b39-98ea-442d-910e-b40a6e4b85c4\") " pod="openstack/barbican-worker-54b7554787-vqrt4" Oct 05 21:13:15 crc kubenswrapper[4754]: I1005 21:13:15.621000 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d9qlv\" (UniqueName: \"kubernetes.io/projected/346d8b39-98ea-442d-910e-b40a6e4b85c4-kube-api-access-d9qlv\") pod \"barbican-worker-54b7554787-vqrt4\" (UID: \"346d8b39-98ea-442d-910e-b40a6e4b85c4\") " pod="openstack/barbican-worker-54b7554787-vqrt4" Oct 05 21:13:15 crc kubenswrapper[4754]: I1005 21:13:15.621024 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/346d8b39-98ea-442d-910e-b40a6e4b85c4-config-data\") pod \"barbican-worker-54b7554787-vqrt4\" (UID: \"346d8b39-98ea-442d-910e-b40a6e4b85c4\") " pod="openstack/barbican-worker-54b7554787-vqrt4" Oct 05 21:13:15 crc kubenswrapper[4754]: I1005 21:13:15.704146 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-9dc4bfc9d-cfm98"] Oct 05 21:13:15 crc kubenswrapper[4754]: I1005 21:13:15.714413 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-59d5ff467f-8sfrm"] Oct 05 21:13:15 crc kubenswrapper[4754]: I1005 21:13:15.716039 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59d5ff467f-8sfrm" Oct 05 21:13:15 crc kubenswrapper[4754]: I1005 21:13:15.716733 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-9dc4bfc9d-cfm98" Oct 05 21:13:15 crc kubenswrapper[4754]: I1005 21:13:15.723123 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/346d8b39-98ea-442d-910e-b40a6e4b85c4-combined-ca-bundle\") pod \"barbican-worker-54b7554787-vqrt4\" (UID: \"346d8b39-98ea-442d-910e-b40a6e4b85c4\") " pod="openstack/barbican-worker-54b7554787-vqrt4" Oct 05 21:13:15 crc kubenswrapper[4754]: I1005 21:13:15.723193 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/346d8b39-98ea-442d-910e-b40a6e4b85c4-logs\") pod \"barbican-worker-54b7554787-vqrt4\" (UID: \"346d8b39-98ea-442d-910e-b40a6e4b85c4\") " pod="openstack/barbican-worker-54b7554787-vqrt4" Oct 05 21:13:15 crc kubenswrapper[4754]: I1005 21:13:15.723226 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/346d8b39-98ea-442d-910e-b40a6e4b85c4-config-data-custom\") pod \"barbican-worker-54b7554787-vqrt4\" (UID: \"346d8b39-98ea-442d-910e-b40a6e4b85c4\") " pod="openstack/barbican-worker-54b7554787-vqrt4" Oct 05 21:13:15 crc kubenswrapper[4754]: I1005 21:13:15.723283 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d9qlv\" (UniqueName: \"kubernetes.io/projected/346d8b39-98ea-442d-910e-b40a6e4b85c4-kube-api-access-d9qlv\") pod \"barbican-worker-54b7554787-vqrt4\" (UID: \"346d8b39-98ea-442d-910e-b40a6e4b85c4\") " pod="openstack/barbican-worker-54b7554787-vqrt4" Oct 05 21:13:15 crc kubenswrapper[4754]: I1005 21:13:15.723313 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/346d8b39-98ea-442d-910e-b40a6e4b85c4-config-data\") pod \"barbican-worker-54b7554787-vqrt4\" (UID: \"346d8b39-98ea-442d-910e-b40a6e4b85c4\") " pod="openstack/barbican-worker-54b7554787-vqrt4" Oct 05 21:13:15 crc kubenswrapper[4754]: I1005 21:13:15.724234 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Oct 05 21:13:15 crc kubenswrapper[4754]: I1005 21:13:15.725280 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/346d8b39-98ea-442d-910e-b40a6e4b85c4-logs\") pod \"barbican-worker-54b7554787-vqrt4\" (UID: \"346d8b39-98ea-442d-910e-b40a6e4b85c4\") " pod="openstack/barbican-worker-54b7554787-vqrt4" Oct 05 21:13:15 crc kubenswrapper[4754]: I1005 21:13:15.757306 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/346d8b39-98ea-442d-910e-b40a6e4b85c4-combined-ca-bundle\") pod \"barbican-worker-54b7554787-vqrt4\" (UID: \"346d8b39-98ea-442d-910e-b40a6e4b85c4\") " pod="openstack/barbican-worker-54b7554787-vqrt4" Oct 05 21:13:15 crc kubenswrapper[4754]: I1005 21:13:15.758108 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/346d8b39-98ea-442d-910e-b40a6e4b85c4-config-data-custom\") pod \"barbican-worker-54b7554787-vqrt4\" (UID: \"346d8b39-98ea-442d-910e-b40a6e4b85c4\") " pod="openstack/barbican-worker-54b7554787-vqrt4" Oct 05 21:13:15 crc kubenswrapper[4754]: I1005 21:13:15.760796 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/346d8b39-98ea-442d-910e-b40a6e4b85c4-config-data\") pod \"barbican-worker-54b7554787-vqrt4\" (UID: \"346d8b39-98ea-442d-910e-b40a6e4b85c4\") " pod="openstack/barbican-worker-54b7554787-vqrt4" Oct 05 21:13:15 crc kubenswrapper[4754]: I1005 21:13:15.761317 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d9qlv\" (UniqueName: \"kubernetes.io/projected/346d8b39-98ea-442d-910e-b40a6e4b85c4-kube-api-access-d9qlv\") pod \"barbican-worker-54b7554787-vqrt4\" (UID: \"346d8b39-98ea-442d-910e-b40a6e4b85c4\") " pod="openstack/barbican-worker-54b7554787-vqrt4" Oct 05 21:13:15 crc kubenswrapper[4754]: I1005 21:13:15.775552 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-9dc4bfc9d-cfm98"] Oct 05 21:13:15 crc kubenswrapper[4754]: I1005 21:13:15.811531 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-59d5ff467f-8sfrm"] Oct 05 21:13:15 crc kubenswrapper[4754]: I1005 21:13:15.824395 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2qn87\" (UniqueName: \"kubernetes.io/projected/358e33ff-b4ce-46e3-b077-e1df373f2a75-kube-api-access-2qn87\") pod \"barbican-keystone-listener-9dc4bfc9d-cfm98\" (UID: \"358e33ff-b4ce-46e3-b077-e1df373f2a75\") " pod="openstack/barbican-keystone-listener-9dc4bfc9d-cfm98" Oct 05 21:13:15 crc kubenswrapper[4754]: I1005 21:13:15.824444 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/358e33ff-b4ce-46e3-b077-e1df373f2a75-combined-ca-bundle\") pod \"barbican-keystone-listener-9dc4bfc9d-cfm98\" (UID: \"358e33ff-b4ce-46e3-b077-e1df373f2a75\") " pod="openstack/barbican-keystone-listener-9dc4bfc9d-cfm98" Oct 05 21:13:15 crc kubenswrapper[4754]: I1005 21:13:15.824483 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3ba3d75f-7b8c-490f-b950-4b11fad2573c-dns-swift-storage-0\") pod \"dnsmasq-dns-59d5ff467f-8sfrm\" (UID: \"3ba3d75f-7b8c-490f-b950-4b11fad2573c\") " pod="openstack/dnsmasq-dns-59d5ff467f-8sfrm" Oct 05 21:13:15 crc kubenswrapper[4754]: I1005 21:13:15.824539 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3ba3d75f-7b8c-490f-b950-4b11fad2573c-ovsdbserver-sb\") pod \"dnsmasq-dns-59d5ff467f-8sfrm\" (UID: \"3ba3d75f-7b8c-490f-b950-4b11fad2573c\") " pod="openstack/dnsmasq-dns-59d5ff467f-8sfrm" Oct 05 21:13:15 crc kubenswrapper[4754]: I1005 21:13:15.824556 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3ba3d75f-7b8c-490f-b950-4b11fad2573c-dns-svc\") pod \"dnsmasq-dns-59d5ff467f-8sfrm\" (UID: \"3ba3d75f-7b8c-490f-b950-4b11fad2573c\") " pod="openstack/dnsmasq-dns-59d5ff467f-8sfrm" Oct 05 21:13:15 crc kubenswrapper[4754]: I1005 21:13:15.824580 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7w2vt\" (UniqueName: \"kubernetes.io/projected/3ba3d75f-7b8c-490f-b950-4b11fad2573c-kube-api-access-7w2vt\") pod \"dnsmasq-dns-59d5ff467f-8sfrm\" (UID: \"3ba3d75f-7b8c-490f-b950-4b11fad2573c\") " pod="openstack/dnsmasq-dns-59d5ff467f-8sfrm" Oct 05 21:13:15 crc kubenswrapper[4754]: I1005 21:13:15.824602 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/358e33ff-b4ce-46e3-b077-e1df373f2a75-logs\") pod \"barbican-keystone-listener-9dc4bfc9d-cfm98\" (UID: \"358e33ff-b4ce-46e3-b077-e1df373f2a75\") " pod="openstack/barbican-keystone-listener-9dc4bfc9d-cfm98" Oct 05 21:13:15 crc kubenswrapper[4754]: I1005 21:13:15.824640 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3ba3d75f-7b8c-490f-b950-4b11fad2573c-config\") pod \"dnsmasq-dns-59d5ff467f-8sfrm\" (UID: \"3ba3d75f-7b8c-490f-b950-4b11fad2573c\") " pod="openstack/dnsmasq-dns-59d5ff467f-8sfrm" Oct 05 21:13:15 crc kubenswrapper[4754]: I1005 21:13:15.824679 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3ba3d75f-7b8c-490f-b950-4b11fad2573c-ovsdbserver-nb\") pod \"dnsmasq-dns-59d5ff467f-8sfrm\" (UID: \"3ba3d75f-7b8c-490f-b950-4b11fad2573c\") " pod="openstack/dnsmasq-dns-59d5ff467f-8sfrm" Oct 05 21:13:15 crc kubenswrapper[4754]: I1005 21:13:15.824699 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/358e33ff-b4ce-46e3-b077-e1df373f2a75-config-data-custom\") pod \"barbican-keystone-listener-9dc4bfc9d-cfm98\" (UID: \"358e33ff-b4ce-46e3-b077-e1df373f2a75\") " pod="openstack/barbican-keystone-listener-9dc4bfc9d-cfm98" Oct 05 21:13:15 crc kubenswrapper[4754]: I1005 21:13:15.824715 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/358e33ff-b4ce-46e3-b077-e1df373f2a75-config-data\") pod \"barbican-keystone-listener-9dc4bfc9d-cfm98\" (UID: \"358e33ff-b4ce-46e3-b077-e1df373f2a75\") " pod="openstack/barbican-keystone-listener-9dc4bfc9d-cfm98" Oct 05 21:13:15 crc kubenswrapper[4754]: I1005 21:13:15.948277 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3ba3d75f-7b8c-490f-b950-4b11fad2573c-ovsdbserver-nb\") pod \"dnsmasq-dns-59d5ff467f-8sfrm\" (UID: \"3ba3d75f-7b8c-490f-b950-4b11fad2573c\") " pod="openstack/dnsmasq-dns-59d5ff467f-8sfrm" Oct 05 21:13:15 crc kubenswrapper[4754]: I1005 21:13:15.948389 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/358e33ff-b4ce-46e3-b077-e1df373f2a75-config-data-custom\") pod \"barbican-keystone-listener-9dc4bfc9d-cfm98\" (UID: \"358e33ff-b4ce-46e3-b077-e1df373f2a75\") " pod="openstack/barbican-keystone-listener-9dc4bfc9d-cfm98" Oct 05 21:13:15 crc kubenswrapper[4754]: I1005 21:13:15.948429 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/358e33ff-b4ce-46e3-b077-e1df373f2a75-config-data\") pod \"barbican-keystone-listener-9dc4bfc9d-cfm98\" (UID: \"358e33ff-b4ce-46e3-b077-e1df373f2a75\") " pod="openstack/barbican-keystone-listener-9dc4bfc9d-cfm98" Oct 05 21:13:15 crc kubenswrapper[4754]: I1005 21:13:15.948602 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2qn87\" (UniqueName: \"kubernetes.io/projected/358e33ff-b4ce-46e3-b077-e1df373f2a75-kube-api-access-2qn87\") pod \"barbican-keystone-listener-9dc4bfc9d-cfm98\" (UID: \"358e33ff-b4ce-46e3-b077-e1df373f2a75\") " pod="openstack/barbican-keystone-listener-9dc4bfc9d-cfm98" Oct 05 21:13:15 crc kubenswrapper[4754]: I1005 21:13:15.948666 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/358e33ff-b4ce-46e3-b077-e1df373f2a75-combined-ca-bundle\") pod \"barbican-keystone-listener-9dc4bfc9d-cfm98\" (UID: \"358e33ff-b4ce-46e3-b077-e1df373f2a75\") " pod="openstack/barbican-keystone-listener-9dc4bfc9d-cfm98" Oct 05 21:13:15 crc kubenswrapper[4754]: I1005 21:13:15.948796 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3ba3d75f-7b8c-490f-b950-4b11fad2573c-dns-swift-storage-0\") pod \"dnsmasq-dns-59d5ff467f-8sfrm\" (UID: \"3ba3d75f-7b8c-490f-b950-4b11fad2573c\") " pod="openstack/dnsmasq-dns-59d5ff467f-8sfrm" Oct 05 21:13:15 crc kubenswrapper[4754]: I1005 21:13:15.948897 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3ba3d75f-7b8c-490f-b950-4b11fad2573c-ovsdbserver-sb\") pod \"dnsmasq-dns-59d5ff467f-8sfrm\" (UID: \"3ba3d75f-7b8c-490f-b950-4b11fad2573c\") " pod="openstack/dnsmasq-dns-59d5ff467f-8sfrm" Oct 05 21:13:15 crc kubenswrapper[4754]: I1005 21:13:15.948941 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3ba3d75f-7b8c-490f-b950-4b11fad2573c-dns-svc\") pod \"dnsmasq-dns-59d5ff467f-8sfrm\" (UID: \"3ba3d75f-7b8c-490f-b950-4b11fad2573c\") " pod="openstack/dnsmasq-dns-59d5ff467f-8sfrm" Oct 05 21:13:15 crc kubenswrapper[4754]: I1005 21:13:15.949023 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7w2vt\" (UniqueName: \"kubernetes.io/projected/3ba3d75f-7b8c-490f-b950-4b11fad2573c-kube-api-access-7w2vt\") pod \"dnsmasq-dns-59d5ff467f-8sfrm\" (UID: \"3ba3d75f-7b8c-490f-b950-4b11fad2573c\") " pod="openstack/dnsmasq-dns-59d5ff467f-8sfrm" Oct 05 21:13:15 crc kubenswrapper[4754]: I1005 21:13:15.949090 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/358e33ff-b4ce-46e3-b077-e1df373f2a75-logs\") pod \"barbican-keystone-listener-9dc4bfc9d-cfm98\" (UID: \"358e33ff-b4ce-46e3-b077-e1df373f2a75\") " pod="openstack/barbican-keystone-listener-9dc4bfc9d-cfm98" Oct 05 21:13:15 crc kubenswrapper[4754]: I1005 21:13:15.949312 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3ba3d75f-7b8c-490f-b950-4b11fad2573c-config\") pod \"dnsmasq-dns-59d5ff467f-8sfrm\" (UID: \"3ba3d75f-7b8c-490f-b950-4b11fad2573c\") " pod="openstack/dnsmasq-dns-59d5ff467f-8sfrm" Oct 05 21:13:15 crc kubenswrapper[4754]: I1005 21:13:15.951392 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/358e33ff-b4ce-46e3-b077-e1df373f2a75-logs\") pod \"barbican-keystone-listener-9dc4bfc9d-cfm98\" (UID: \"358e33ff-b4ce-46e3-b077-e1df373f2a75\") " pod="openstack/barbican-keystone-listener-9dc4bfc9d-cfm98" Oct 05 21:13:15 crc kubenswrapper[4754]: I1005 21:13:15.951860 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3ba3d75f-7b8c-490f-b950-4b11fad2573c-dns-swift-storage-0\") pod \"dnsmasq-dns-59d5ff467f-8sfrm\" (UID: \"3ba3d75f-7b8c-490f-b950-4b11fad2573c\") " pod="openstack/dnsmasq-dns-59d5ff467f-8sfrm" Oct 05 21:13:15 crc kubenswrapper[4754]: I1005 21:13:15.952657 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3ba3d75f-7b8c-490f-b950-4b11fad2573c-ovsdbserver-sb\") pod \"dnsmasq-dns-59d5ff467f-8sfrm\" (UID: \"3ba3d75f-7b8c-490f-b950-4b11fad2573c\") " pod="openstack/dnsmasq-dns-59d5ff467f-8sfrm" Oct 05 21:13:15 crc kubenswrapper[4754]: I1005 21:13:15.953381 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3ba3d75f-7b8c-490f-b950-4b11fad2573c-dns-svc\") pod \"dnsmasq-dns-59d5ff467f-8sfrm\" (UID: \"3ba3d75f-7b8c-490f-b950-4b11fad2573c\") " pod="openstack/dnsmasq-dns-59d5ff467f-8sfrm" Oct 05 21:13:15 crc kubenswrapper[4754]: I1005 21:13:15.955231 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3ba3d75f-7b8c-490f-b950-4b11fad2573c-ovsdbserver-nb\") pod \"dnsmasq-dns-59d5ff467f-8sfrm\" (UID: \"3ba3d75f-7b8c-490f-b950-4b11fad2573c\") " pod="openstack/dnsmasq-dns-59d5ff467f-8sfrm" Oct 05 21:13:15 crc kubenswrapper[4754]: I1005 21:13:15.958884 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-54b7554787-vqrt4" Oct 05 21:13:15 crc kubenswrapper[4754]: I1005 21:13:15.963730 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3ba3d75f-7b8c-490f-b950-4b11fad2573c-config\") pod \"dnsmasq-dns-59d5ff467f-8sfrm\" (UID: \"3ba3d75f-7b8c-490f-b950-4b11fad2573c\") " pod="openstack/dnsmasq-dns-59d5ff467f-8sfrm" Oct 05 21:13:16 crc kubenswrapper[4754]: I1005 21:13:15.991237 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2qn87\" (UniqueName: \"kubernetes.io/projected/358e33ff-b4ce-46e3-b077-e1df373f2a75-kube-api-access-2qn87\") pod \"barbican-keystone-listener-9dc4bfc9d-cfm98\" (UID: \"358e33ff-b4ce-46e3-b077-e1df373f2a75\") " pod="openstack/barbican-keystone-listener-9dc4bfc9d-cfm98" Oct 05 21:13:16 crc kubenswrapper[4754]: I1005 21:13:16.035183 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/358e33ff-b4ce-46e3-b077-e1df373f2a75-config-data\") pod \"barbican-keystone-listener-9dc4bfc9d-cfm98\" (UID: \"358e33ff-b4ce-46e3-b077-e1df373f2a75\") " pod="openstack/barbican-keystone-listener-9dc4bfc9d-cfm98" Oct 05 21:13:16 crc kubenswrapper[4754]: I1005 21:13:16.045163 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/358e33ff-b4ce-46e3-b077-e1df373f2a75-config-data-custom\") pod \"barbican-keystone-listener-9dc4bfc9d-cfm98\" (UID: \"358e33ff-b4ce-46e3-b077-e1df373f2a75\") " pod="openstack/barbican-keystone-listener-9dc4bfc9d-cfm98" Oct 05 21:13:16 crc kubenswrapper[4754]: I1005 21:13:16.082457 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/358e33ff-b4ce-46e3-b077-e1df373f2a75-combined-ca-bundle\") pod \"barbican-keystone-listener-9dc4bfc9d-cfm98\" (UID: \"358e33ff-b4ce-46e3-b077-e1df373f2a75\") " pod="openstack/barbican-keystone-listener-9dc4bfc9d-cfm98" Oct 05 21:13:16 crc kubenswrapper[4754]: I1005 21:13:16.095041 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7w2vt\" (UniqueName: \"kubernetes.io/projected/3ba3d75f-7b8c-490f-b950-4b11fad2573c-kube-api-access-7w2vt\") pod \"dnsmasq-dns-59d5ff467f-8sfrm\" (UID: \"3ba3d75f-7b8c-490f-b950-4b11fad2573c\") " pod="openstack/dnsmasq-dns-59d5ff467f-8sfrm" Oct 05 21:13:16 crc kubenswrapper[4754]: I1005 21:13:16.121813 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59d5ff467f-8sfrm" Oct 05 21:13:16 crc kubenswrapper[4754]: I1005 21:13:16.147565 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-5fb78978fb-2mr8x"] Oct 05 21:13:16 crc kubenswrapper[4754]: I1005 21:13:16.149204 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5fb78978fb-2mr8x" Oct 05 21:13:16 crc kubenswrapper[4754]: I1005 21:13:16.156876 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Oct 05 21:13:16 crc kubenswrapper[4754]: I1005 21:13:16.159661 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-9dc4bfc9d-cfm98" Oct 05 21:13:16 crc kubenswrapper[4754]: I1005 21:13:16.177766 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-5fb78978fb-2mr8x"] Oct 05 21:13:16 crc kubenswrapper[4754]: I1005 21:13:16.292941 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rq47t\" (UniqueName: \"kubernetes.io/projected/a0655389-4b2f-4973-b572-88df7ee28344-kube-api-access-rq47t\") pod \"barbican-api-5fb78978fb-2mr8x\" (UID: \"a0655389-4b2f-4973-b572-88df7ee28344\") " pod="openstack/barbican-api-5fb78978fb-2mr8x" Oct 05 21:13:16 crc kubenswrapper[4754]: I1005 21:13:16.293540 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a0655389-4b2f-4973-b572-88df7ee28344-config-data-custom\") pod \"barbican-api-5fb78978fb-2mr8x\" (UID: \"a0655389-4b2f-4973-b572-88df7ee28344\") " pod="openstack/barbican-api-5fb78978fb-2mr8x" Oct 05 21:13:16 crc kubenswrapper[4754]: I1005 21:13:16.293636 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0655389-4b2f-4973-b572-88df7ee28344-config-data\") pod \"barbican-api-5fb78978fb-2mr8x\" (UID: \"a0655389-4b2f-4973-b572-88df7ee28344\") " pod="openstack/barbican-api-5fb78978fb-2mr8x" Oct 05 21:13:16 crc kubenswrapper[4754]: I1005 21:13:16.294142 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a0655389-4b2f-4973-b572-88df7ee28344-logs\") pod \"barbican-api-5fb78978fb-2mr8x\" (UID: \"a0655389-4b2f-4973-b572-88df7ee28344\") " pod="openstack/barbican-api-5fb78978fb-2mr8x" Oct 05 21:13:16 crc kubenswrapper[4754]: I1005 21:13:16.294564 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0655389-4b2f-4973-b572-88df7ee28344-combined-ca-bundle\") pod \"barbican-api-5fb78978fb-2mr8x\" (UID: \"a0655389-4b2f-4973-b572-88df7ee28344\") " pod="openstack/barbican-api-5fb78978fb-2mr8x" Oct 05 21:13:16 crc kubenswrapper[4754]: I1005 21:13:16.399138 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rq47t\" (UniqueName: \"kubernetes.io/projected/a0655389-4b2f-4973-b572-88df7ee28344-kube-api-access-rq47t\") pod \"barbican-api-5fb78978fb-2mr8x\" (UID: \"a0655389-4b2f-4973-b572-88df7ee28344\") " pod="openstack/barbican-api-5fb78978fb-2mr8x" Oct 05 21:13:16 crc kubenswrapper[4754]: I1005 21:13:16.399205 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a0655389-4b2f-4973-b572-88df7ee28344-config-data-custom\") pod \"barbican-api-5fb78978fb-2mr8x\" (UID: \"a0655389-4b2f-4973-b572-88df7ee28344\") " pod="openstack/barbican-api-5fb78978fb-2mr8x" Oct 05 21:13:16 crc kubenswrapper[4754]: I1005 21:13:16.399224 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0655389-4b2f-4973-b572-88df7ee28344-config-data\") pod \"barbican-api-5fb78978fb-2mr8x\" (UID: \"a0655389-4b2f-4973-b572-88df7ee28344\") " pod="openstack/barbican-api-5fb78978fb-2mr8x" Oct 05 21:13:16 crc kubenswrapper[4754]: I1005 21:13:16.399296 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a0655389-4b2f-4973-b572-88df7ee28344-logs\") pod \"barbican-api-5fb78978fb-2mr8x\" (UID: \"a0655389-4b2f-4973-b572-88df7ee28344\") " pod="openstack/barbican-api-5fb78978fb-2mr8x" Oct 05 21:13:16 crc kubenswrapper[4754]: I1005 21:13:16.399381 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0655389-4b2f-4973-b572-88df7ee28344-combined-ca-bundle\") pod \"barbican-api-5fb78978fb-2mr8x\" (UID: \"a0655389-4b2f-4973-b572-88df7ee28344\") " pod="openstack/barbican-api-5fb78978fb-2mr8x" Oct 05 21:13:16 crc kubenswrapper[4754]: I1005 21:13:16.404572 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a0655389-4b2f-4973-b572-88df7ee28344-logs\") pod \"barbican-api-5fb78978fb-2mr8x\" (UID: \"a0655389-4b2f-4973-b572-88df7ee28344\") " pod="openstack/barbican-api-5fb78978fb-2mr8x" Oct 05 21:13:16 crc kubenswrapper[4754]: I1005 21:13:16.413386 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a0655389-4b2f-4973-b572-88df7ee28344-config-data-custom\") pod \"barbican-api-5fb78978fb-2mr8x\" (UID: \"a0655389-4b2f-4973-b572-88df7ee28344\") " pod="openstack/barbican-api-5fb78978fb-2mr8x" Oct 05 21:13:16 crc kubenswrapper[4754]: I1005 21:13:16.420374 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0655389-4b2f-4973-b572-88df7ee28344-config-data\") pod \"barbican-api-5fb78978fb-2mr8x\" (UID: \"a0655389-4b2f-4973-b572-88df7ee28344\") " pod="openstack/barbican-api-5fb78978fb-2mr8x" Oct 05 21:13:16 crc kubenswrapper[4754]: I1005 21:13:16.428677 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0655389-4b2f-4973-b572-88df7ee28344-combined-ca-bundle\") pod \"barbican-api-5fb78978fb-2mr8x\" (UID: \"a0655389-4b2f-4973-b572-88df7ee28344\") " pod="openstack/barbican-api-5fb78978fb-2mr8x" Oct 05 21:13:16 crc kubenswrapper[4754]: I1005 21:13:16.451057 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9fbf0155-8666-4b0c-a321-8f6c788b529d","Type":"ContainerStarted","Data":"2e3e7b10d0983827e68056feff99b72e1a1bb822b837d9c738e1730f1e6b70e4"} Oct 05 21:13:16 crc kubenswrapper[4754]: I1005 21:13:16.457050 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rq47t\" (UniqueName: \"kubernetes.io/projected/a0655389-4b2f-4973-b572-88df7ee28344-kube-api-access-rq47t\") pod \"barbican-api-5fb78978fb-2mr8x\" (UID: \"a0655389-4b2f-4973-b572-88df7ee28344\") " pod="openstack/barbican-api-5fb78978fb-2mr8x" Oct 05 21:13:16 crc kubenswrapper[4754]: I1005 21:13:16.506457 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5fb78978fb-2mr8x" Oct 05 21:13:16 crc kubenswrapper[4754]: I1005 21:13:16.956303 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-54b7554787-vqrt4"] Oct 05 21:13:17 crc kubenswrapper[4754]: I1005 21:13:17.206785 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-59d5ff467f-8sfrm"] Oct 05 21:13:17 crc kubenswrapper[4754]: I1005 21:13:17.322298 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-5fb78978fb-2mr8x"] Oct 05 21:13:17 crc kubenswrapper[4754]: I1005 21:13:17.340467 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-9dc4bfc9d-cfm98"] Oct 05 21:13:17 crc kubenswrapper[4754]: I1005 21:13:17.463195 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59d5ff467f-8sfrm" event={"ID":"3ba3d75f-7b8c-490f-b950-4b11fad2573c","Type":"ContainerStarted","Data":"f2d70caed95a6844b6110b82857e137d7878f5779a72ac9128bf58c089c22a71"} Oct 05 21:13:17 crc kubenswrapper[4754]: I1005 21:13:17.463961 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-9dc4bfc9d-cfm98" event={"ID":"358e33ff-b4ce-46e3-b077-e1df373f2a75","Type":"ContainerStarted","Data":"fa2c3e44a0cd0612a44c821e2403ae2d9d8566eb11239b014edc5a54e36b4e31"} Oct 05 21:13:17 crc kubenswrapper[4754]: I1005 21:13:17.466166 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5fb78978fb-2mr8x" event={"ID":"a0655389-4b2f-4973-b572-88df7ee28344","Type":"ContainerStarted","Data":"7905d5d472685f664cbeb567f15d32bbde1d4a02eb2b026d1f81786c2145c5bc"} Oct 05 21:13:17 crc kubenswrapper[4754]: I1005 21:13:17.467886 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-54b7554787-vqrt4" event={"ID":"346d8b39-98ea-442d-910e-b40a6e4b85c4","Type":"ContainerStarted","Data":"e6b699e69f864a919c3fc1f3e62fdba4525fdd69f30c32f83dddadd196b8e978"} Oct 05 21:13:17 crc kubenswrapper[4754]: I1005 21:13:17.471174 4754 generic.go:334] "Generic (PLEG): container finished" podID="94d0ae23-e26d-4bca-b7f3-e4af07ac2722" containerID="907c4f77f15040be8d5f56e9c1085880e54be3a78590fa12671e33449f1ded81" exitCode=0 Oct 05 21:13:17 crc kubenswrapper[4754]: I1005 21:13:17.471268 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-xncdg" event={"ID":"94d0ae23-e26d-4bca-b7f3-e4af07ac2722","Type":"ContainerDied","Data":"907c4f77f15040be8d5f56e9c1085880e54be3a78590fa12671e33449f1ded81"} Oct 05 21:13:17 crc kubenswrapper[4754]: I1005 21:13:17.480238 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-xlw68" event={"ID":"94c73d9b-e224-4788-995d-be11dcee7b3c","Type":"ContainerStarted","Data":"beb03566dca75372fd1c3a5f05aae0ee0555748b120f181ee2ca7eb6e2d97a71"} Oct 05 21:13:17 crc kubenswrapper[4754]: I1005 21:13:17.506663 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-xlw68" podStartSLOduration=4.075997524 podStartE2EDuration="1m36.506643219s" podCreationTimestamp="2025-10-05 21:11:41 +0000 UTC" firstStartedPulling="2025-10-05 21:11:44.063391948 +0000 UTC m=+1027.967510659" lastFinishedPulling="2025-10-05 21:13:16.494037654 +0000 UTC m=+1120.398156354" observedRunningTime="2025-10-05 21:13:17.503757234 +0000 UTC m=+1121.407875944" watchObservedRunningTime="2025-10-05 21:13:17.506643219 +0000 UTC m=+1121.410761929" Oct 05 21:13:17 crc kubenswrapper[4754]: E1005 21:13:17.929919 4754 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3ba3d75f_7b8c_490f_b950_4b11fad2573c.slice/crio-conmon-aa4560e4bee403604979bda67fd34b6f05aa1466c510822d70fc137f09672782.scope\": RecentStats: unable to find data in memory cache]" Oct 05 21:13:18 crc kubenswrapper[4754]: I1005 21:13:18.546388 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9fbf0155-8666-4b0c-a321-8f6c788b529d","Type":"ContainerStarted","Data":"80d1fd901cea8412d079b32fc3b0d132e3b3755b31b3438ab736600c319c8d61"} Oct 05 21:13:18 crc kubenswrapper[4754]: I1005 21:13:18.547813 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 05 21:13:18 crc kubenswrapper[4754]: I1005 21:13:18.557410 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5fb78978fb-2mr8x" event={"ID":"a0655389-4b2f-4973-b572-88df7ee28344","Type":"ContainerStarted","Data":"e09d95f4d0fe4bccba2be305d0ec39651d39f0c07e8fbb497b7f9bad6ff21f9b"} Oct 05 21:13:18 crc kubenswrapper[4754]: I1005 21:13:18.557460 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5fb78978fb-2mr8x" event={"ID":"a0655389-4b2f-4973-b572-88df7ee28344","Type":"ContainerStarted","Data":"7e2f47e1eba20b443bc6c1598258c83ac78c624aa918294f7f18a7ff0e5b1770"} Oct 05 21:13:18 crc kubenswrapper[4754]: I1005 21:13:18.557737 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-5fb78978fb-2mr8x" Oct 05 21:13:18 crc kubenswrapper[4754]: I1005 21:13:18.558216 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-5fb78978fb-2mr8x" Oct 05 21:13:18 crc kubenswrapper[4754]: I1005 21:13:18.566739 4754 generic.go:334] "Generic (PLEG): container finished" podID="3ba3d75f-7b8c-490f-b950-4b11fad2573c" containerID="aa4560e4bee403604979bda67fd34b6f05aa1466c510822d70fc137f09672782" exitCode=0 Oct 05 21:13:18 crc kubenswrapper[4754]: I1005 21:13:18.567568 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59d5ff467f-8sfrm" event={"ID":"3ba3d75f-7b8c-490f-b950-4b11fad2573c","Type":"ContainerDied","Data":"aa4560e4bee403604979bda67fd34b6f05aa1466c510822d70fc137f09672782"} Oct 05 21:13:18 crc kubenswrapper[4754]: I1005 21:13:18.613098 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.221234747 podStartE2EDuration="6.613079073s" podCreationTimestamp="2025-10-05 21:13:12 +0000 UTC" firstStartedPulling="2025-10-05 21:13:13.339096876 +0000 UTC m=+1117.243215576" lastFinishedPulling="2025-10-05 21:13:17.730941192 +0000 UTC m=+1121.635059902" observedRunningTime="2025-10-05 21:13:18.611022059 +0000 UTC m=+1122.515140769" watchObservedRunningTime="2025-10-05 21:13:18.613079073 +0000 UTC m=+1122.517197783" Oct 05 21:13:18 crc kubenswrapper[4754]: I1005 21:13:18.744166 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-5fb78978fb-2mr8x" podStartSLOduration=2.744140342 podStartE2EDuration="2.744140342s" podCreationTimestamp="2025-10-05 21:13:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 21:13:18.710925478 +0000 UTC m=+1122.615044178" watchObservedRunningTime="2025-10-05 21:13:18.744140342 +0000 UTC m=+1122.648259052" Oct 05 21:13:19 crc kubenswrapper[4754]: I1005 21:13:19.053866 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-xncdg" Oct 05 21:13:19 crc kubenswrapper[4754]: I1005 21:13:19.176973 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/94d0ae23-e26d-4bca-b7f3-e4af07ac2722-config-data\") pod \"94d0ae23-e26d-4bca-b7f3-e4af07ac2722\" (UID: \"94d0ae23-e26d-4bca-b7f3-e4af07ac2722\") " Oct 05 21:13:19 crc kubenswrapper[4754]: I1005 21:13:19.177046 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94d0ae23-e26d-4bca-b7f3-e4af07ac2722-combined-ca-bundle\") pod \"94d0ae23-e26d-4bca-b7f3-e4af07ac2722\" (UID: \"94d0ae23-e26d-4bca-b7f3-e4af07ac2722\") " Oct 05 21:13:19 crc kubenswrapper[4754]: I1005 21:13:19.177252 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/94d0ae23-e26d-4bca-b7f3-e4af07ac2722-etc-machine-id\") pod \"94d0ae23-e26d-4bca-b7f3-e4af07ac2722\" (UID: \"94d0ae23-e26d-4bca-b7f3-e4af07ac2722\") " Oct 05 21:13:19 crc kubenswrapper[4754]: I1005 21:13:19.177343 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/94d0ae23-e26d-4bca-b7f3-e4af07ac2722-db-sync-config-data\") pod \"94d0ae23-e26d-4bca-b7f3-e4af07ac2722\" (UID: \"94d0ae23-e26d-4bca-b7f3-e4af07ac2722\") " Oct 05 21:13:19 crc kubenswrapper[4754]: I1005 21:13:19.177384 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/94d0ae23-e26d-4bca-b7f3-e4af07ac2722-scripts\") pod \"94d0ae23-e26d-4bca-b7f3-e4af07ac2722\" (UID: \"94d0ae23-e26d-4bca-b7f3-e4af07ac2722\") " Oct 05 21:13:19 crc kubenswrapper[4754]: I1005 21:13:19.177460 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tvn5l\" (UniqueName: \"kubernetes.io/projected/94d0ae23-e26d-4bca-b7f3-e4af07ac2722-kube-api-access-tvn5l\") pod \"94d0ae23-e26d-4bca-b7f3-e4af07ac2722\" (UID: \"94d0ae23-e26d-4bca-b7f3-e4af07ac2722\") " Oct 05 21:13:19 crc kubenswrapper[4754]: I1005 21:13:19.178115 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/94d0ae23-e26d-4bca-b7f3-e4af07ac2722-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "94d0ae23-e26d-4bca-b7f3-e4af07ac2722" (UID: "94d0ae23-e26d-4bca-b7f3-e4af07ac2722"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 05 21:13:19 crc kubenswrapper[4754]: I1005 21:13:19.185906 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/94d0ae23-e26d-4bca-b7f3-e4af07ac2722-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "94d0ae23-e26d-4bca-b7f3-e4af07ac2722" (UID: "94d0ae23-e26d-4bca-b7f3-e4af07ac2722"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:13:19 crc kubenswrapper[4754]: I1005 21:13:19.191170 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/94d0ae23-e26d-4bca-b7f3-e4af07ac2722-scripts" (OuterVolumeSpecName: "scripts") pod "94d0ae23-e26d-4bca-b7f3-e4af07ac2722" (UID: "94d0ae23-e26d-4bca-b7f3-e4af07ac2722"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:13:19 crc kubenswrapper[4754]: I1005 21:13:19.193227 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/94d0ae23-e26d-4bca-b7f3-e4af07ac2722-kube-api-access-tvn5l" (OuterVolumeSpecName: "kube-api-access-tvn5l") pod "94d0ae23-e26d-4bca-b7f3-e4af07ac2722" (UID: "94d0ae23-e26d-4bca-b7f3-e4af07ac2722"). InnerVolumeSpecName "kube-api-access-tvn5l". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:13:19 crc kubenswrapper[4754]: I1005 21:13:19.237333 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/94d0ae23-e26d-4bca-b7f3-e4af07ac2722-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "94d0ae23-e26d-4bca-b7f3-e4af07ac2722" (UID: "94d0ae23-e26d-4bca-b7f3-e4af07ac2722"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:13:19 crc kubenswrapper[4754]: I1005 21:13:19.266357 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/94d0ae23-e26d-4bca-b7f3-e4af07ac2722-config-data" (OuterVolumeSpecName: "config-data") pod "94d0ae23-e26d-4bca-b7f3-e4af07ac2722" (UID: "94d0ae23-e26d-4bca-b7f3-e4af07ac2722"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:13:19 crc kubenswrapper[4754]: I1005 21:13:19.280169 4754 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/94d0ae23-e26d-4bca-b7f3-e4af07ac2722-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 05 21:13:19 crc kubenswrapper[4754]: I1005 21:13:19.280212 4754 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/94d0ae23-e26d-4bca-b7f3-e4af07ac2722-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 21:13:19 crc kubenswrapper[4754]: I1005 21:13:19.280228 4754 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/94d0ae23-e26d-4bca-b7f3-e4af07ac2722-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 21:13:19 crc kubenswrapper[4754]: I1005 21:13:19.280241 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tvn5l\" (UniqueName: \"kubernetes.io/projected/94d0ae23-e26d-4bca-b7f3-e4af07ac2722-kube-api-access-tvn5l\") on node \"crc\" DevicePath \"\"" Oct 05 21:13:19 crc kubenswrapper[4754]: I1005 21:13:19.280256 4754 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/94d0ae23-e26d-4bca-b7f3-e4af07ac2722-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 21:13:19 crc kubenswrapper[4754]: I1005 21:13:19.280280 4754 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94d0ae23-e26d-4bca-b7f3-e4af07ac2722-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 21:13:19 crc kubenswrapper[4754]: I1005 21:13:19.428141 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-55bcb5dcdd-jhgrw" Oct 05 21:13:19 crc kubenswrapper[4754]: I1005 21:13:19.608418 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-xncdg" event={"ID":"94d0ae23-e26d-4bca-b7f3-e4af07ac2722","Type":"ContainerDied","Data":"77b307198d532f13d65299e2dad4c683500be12a880a83e15d30c5537d8c355f"} Oct 05 21:13:19 crc kubenswrapper[4754]: I1005 21:13:19.608477 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="77b307198d532f13d65299e2dad4c683500be12a880a83e15d30c5537d8c355f" Oct 05 21:13:19 crc kubenswrapper[4754]: I1005 21:13:19.608592 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-xncdg" Oct 05 21:13:19 crc kubenswrapper[4754]: I1005 21:13:19.643576 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59d5ff467f-8sfrm" event={"ID":"3ba3d75f-7b8c-490f-b950-4b11fad2573c","Type":"ContainerStarted","Data":"591b748c000024e5fbfc6c9f33557e705a206aef0da8a9757f00d9f49c6dce1f"} Oct 05 21:13:19 crc kubenswrapper[4754]: I1005 21:13:19.865753 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-59d5ff467f-8sfrm" podStartSLOduration=4.865736255 podStartE2EDuration="4.865736255s" podCreationTimestamp="2025-10-05 21:13:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 21:13:19.696742578 +0000 UTC m=+1123.600861288" watchObservedRunningTime="2025-10-05 21:13:19.865736255 +0000 UTC m=+1123.769854965" Oct 05 21:13:19 crc kubenswrapper[4754]: I1005 21:13:19.876293 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-765c6857c8-q7glx"] Oct 05 21:13:19 crc kubenswrapper[4754]: E1005 21:13:19.876718 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94d0ae23-e26d-4bca-b7f3-e4af07ac2722" containerName="cinder-db-sync" Oct 05 21:13:19 crc kubenswrapper[4754]: I1005 21:13:19.876734 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="94d0ae23-e26d-4bca-b7f3-e4af07ac2722" containerName="cinder-db-sync" Oct 05 21:13:19 crc kubenswrapper[4754]: I1005 21:13:19.876942 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="94d0ae23-e26d-4bca-b7f3-e4af07ac2722" containerName="cinder-db-sync" Oct 05 21:13:19 crc kubenswrapper[4754]: I1005 21:13:19.877849 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-765c6857c8-q7glx" Oct 05 21:13:19 crc kubenswrapper[4754]: I1005 21:13:19.901207 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Oct 05 21:13:19 crc kubenswrapper[4754]: I1005 21:13:19.901528 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Oct 05 21:13:19 crc kubenswrapper[4754]: I1005 21:13:19.974277 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-765c6857c8-q7glx"] Oct 05 21:13:20 crc kubenswrapper[4754]: I1005 21:13:20.015290 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd23550d-8815-4211-b8be-10ee8ab95c1f-config-data\") pod \"barbican-api-765c6857c8-q7glx\" (UID: \"bd23550d-8815-4211-b8be-10ee8ab95c1f\") " pod="openstack/barbican-api-765c6857c8-q7glx" Oct 05 21:13:20 crc kubenswrapper[4754]: I1005 21:13:20.015548 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bd23550d-8815-4211-b8be-10ee8ab95c1f-logs\") pod \"barbican-api-765c6857c8-q7glx\" (UID: \"bd23550d-8815-4211-b8be-10ee8ab95c1f\") " pod="openstack/barbican-api-765c6857c8-q7glx" Oct 05 21:13:20 crc kubenswrapper[4754]: I1005 21:13:20.015602 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l4d2h\" (UniqueName: \"kubernetes.io/projected/bd23550d-8815-4211-b8be-10ee8ab95c1f-kube-api-access-l4d2h\") pod \"barbican-api-765c6857c8-q7glx\" (UID: \"bd23550d-8815-4211-b8be-10ee8ab95c1f\") " pod="openstack/barbican-api-765c6857c8-q7glx" Oct 05 21:13:20 crc kubenswrapper[4754]: I1005 21:13:20.015694 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bd23550d-8815-4211-b8be-10ee8ab95c1f-public-tls-certs\") pod \"barbican-api-765c6857c8-q7glx\" (UID: \"bd23550d-8815-4211-b8be-10ee8ab95c1f\") " pod="openstack/barbican-api-765c6857c8-q7glx" Oct 05 21:13:20 crc kubenswrapper[4754]: I1005 21:13:20.015943 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd23550d-8815-4211-b8be-10ee8ab95c1f-combined-ca-bundle\") pod \"barbican-api-765c6857c8-q7glx\" (UID: \"bd23550d-8815-4211-b8be-10ee8ab95c1f\") " pod="openstack/barbican-api-765c6857c8-q7glx" Oct 05 21:13:20 crc kubenswrapper[4754]: I1005 21:13:20.016082 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bd23550d-8815-4211-b8be-10ee8ab95c1f-config-data-custom\") pod \"barbican-api-765c6857c8-q7glx\" (UID: \"bd23550d-8815-4211-b8be-10ee8ab95c1f\") " pod="openstack/barbican-api-765c6857c8-q7glx" Oct 05 21:13:20 crc kubenswrapper[4754]: I1005 21:13:20.016193 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bd23550d-8815-4211-b8be-10ee8ab95c1f-internal-tls-certs\") pod \"barbican-api-765c6857c8-q7glx\" (UID: \"bd23550d-8815-4211-b8be-10ee8ab95c1f\") " pod="openstack/barbican-api-765c6857c8-q7glx" Oct 05 21:13:20 crc kubenswrapper[4754]: I1005 21:13:20.119779 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bd23550d-8815-4211-b8be-10ee8ab95c1f-config-data-custom\") pod \"barbican-api-765c6857c8-q7glx\" (UID: \"bd23550d-8815-4211-b8be-10ee8ab95c1f\") " pod="openstack/barbican-api-765c6857c8-q7glx" Oct 05 21:13:20 crc kubenswrapper[4754]: I1005 21:13:20.119872 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bd23550d-8815-4211-b8be-10ee8ab95c1f-internal-tls-certs\") pod \"barbican-api-765c6857c8-q7glx\" (UID: \"bd23550d-8815-4211-b8be-10ee8ab95c1f\") " pod="openstack/barbican-api-765c6857c8-q7glx" Oct 05 21:13:20 crc kubenswrapper[4754]: I1005 21:13:20.119913 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd23550d-8815-4211-b8be-10ee8ab95c1f-config-data\") pod \"barbican-api-765c6857c8-q7glx\" (UID: \"bd23550d-8815-4211-b8be-10ee8ab95c1f\") " pod="openstack/barbican-api-765c6857c8-q7glx" Oct 05 21:13:20 crc kubenswrapper[4754]: I1005 21:13:20.119958 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bd23550d-8815-4211-b8be-10ee8ab95c1f-logs\") pod \"barbican-api-765c6857c8-q7glx\" (UID: \"bd23550d-8815-4211-b8be-10ee8ab95c1f\") " pod="openstack/barbican-api-765c6857c8-q7glx" Oct 05 21:13:20 crc kubenswrapper[4754]: I1005 21:13:20.119977 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l4d2h\" (UniqueName: \"kubernetes.io/projected/bd23550d-8815-4211-b8be-10ee8ab95c1f-kube-api-access-l4d2h\") pod \"barbican-api-765c6857c8-q7glx\" (UID: \"bd23550d-8815-4211-b8be-10ee8ab95c1f\") " pod="openstack/barbican-api-765c6857c8-q7glx" Oct 05 21:13:20 crc kubenswrapper[4754]: I1005 21:13:20.120008 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bd23550d-8815-4211-b8be-10ee8ab95c1f-public-tls-certs\") pod \"barbican-api-765c6857c8-q7glx\" (UID: \"bd23550d-8815-4211-b8be-10ee8ab95c1f\") " pod="openstack/barbican-api-765c6857c8-q7glx" Oct 05 21:13:20 crc kubenswrapper[4754]: I1005 21:13:20.120095 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd23550d-8815-4211-b8be-10ee8ab95c1f-combined-ca-bundle\") pod \"barbican-api-765c6857c8-q7glx\" (UID: \"bd23550d-8815-4211-b8be-10ee8ab95c1f\") " pod="openstack/barbican-api-765c6857c8-q7glx" Oct 05 21:13:20 crc kubenswrapper[4754]: I1005 21:13:20.123874 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bd23550d-8815-4211-b8be-10ee8ab95c1f-logs\") pod \"barbican-api-765c6857c8-q7glx\" (UID: \"bd23550d-8815-4211-b8be-10ee8ab95c1f\") " pod="openstack/barbican-api-765c6857c8-q7glx" Oct 05 21:13:20 crc kubenswrapper[4754]: I1005 21:13:20.141475 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bd23550d-8815-4211-b8be-10ee8ab95c1f-config-data-custom\") pod \"barbican-api-765c6857c8-q7glx\" (UID: \"bd23550d-8815-4211-b8be-10ee8ab95c1f\") " pod="openstack/barbican-api-765c6857c8-q7glx" Oct 05 21:13:20 crc kubenswrapper[4754]: I1005 21:13:20.144564 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd23550d-8815-4211-b8be-10ee8ab95c1f-combined-ca-bundle\") pod \"barbican-api-765c6857c8-q7glx\" (UID: \"bd23550d-8815-4211-b8be-10ee8ab95c1f\") " pod="openstack/barbican-api-765c6857c8-q7glx" Oct 05 21:13:20 crc kubenswrapper[4754]: I1005 21:13:20.149206 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bd23550d-8815-4211-b8be-10ee8ab95c1f-internal-tls-certs\") pod \"barbican-api-765c6857c8-q7glx\" (UID: \"bd23550d-8815-4211-b8be-10ee8ab95c1f\") " pod="openstack/barbican-api-765c6857c8-q7glx" Oct 05 21:13:20 crc kubenswrapper[4754]: I1005 21:13:20.151064 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd23550d-8815-4211-b8be-10ee8ab95c1f-config-data\") pod \"barbican-api-765c6857c8-q7glx\" (UID: \"bd23550d-8815-4211-b8be-10ee8ab95c1f\") " pod="openstack/barbican-api-765c6857c8-q7glx" Oct 05 21:13:20 crc kubenswrapper[4754]: I1005 21:13:20.153312 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bd23550d-8815-4211-b8be-10ee8ab95c1f-public-tls-certs\") pod \"barbican-api-765c6857c8-q7glx\" (UID: \"bd23550d-8815-4211-b8be-10ee8ab95c1f\") " pod="openstack/barbican-api-765c6857c8-q7glx" Oct 05 21:13:20 crc kubenswrapper[4754]: I1005 21:13:20.172163 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l4d2h\" (UniqueName: \"kubernetes.io/projected/bd23550d-8815-4211-b8be-10ee8ab95c1f-kube-api-access-l4d2h\") pod \"barbican-api-765c6857c8-q7glx\" (UID: \"bd23550d-8815-4211-b8be-10ee8ab95c1f\") " pod="openstack/barbican-api-765c6857c8-q7glx" Oct 05 21:13:20 crc kubenswrapper[4754]: I1005 21:13:20.204155 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-765c6857c8-q7glx" Oct 05 21:13:20 crc kubenswrapper[4754]: I1005 21:13:20.222029 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Oct 05 21:13:20 crc kubenswrapper[4754]: I1005 21:13:20.224178 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 05 21:13:20 crc kubenswrapper[4754]: I1005 21:13:20.257672 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Oct 05 21:13:20 crc kubenswrapper[4754]: I1005 21:13:20.258172 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Oct 05 21:13:20 crc kubenswrapper[4754]: I1005 21:13:20.258301 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Oct 05 21:13:20 crc kubenswrapper[4754]: I1005 21:13:20.258355 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-s6bxr" Oct 05 21:13:20 crc kubenswrapper[4754]: I1005 21:13:20.285060 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 05 21:13:20 crc kubenswrapper[4754]: I1005 21:13:20.425540 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e757f47-326c-4174-9c28-c577a314f70b-config-data\") pod \"cinder-scheduler-0\" (UID: \"5e757f47-326c-4174-9c28-c577a314f70b\") " pod="openstack/cinder-scheduler-0" Oct 05 21:13:20 crc kubenswrapper[4754]: I1005 21:13:20.425599 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5e757f47-326c-4174-9c28-c577a314f70b-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"5e757f47-326c-4174-9c28-c577a314f70b\") " pod="openstack/cinder-scheduler-0" Oct 05 21:13:20 crc kubenswrapper[4754]: I1005 21:13:20.425635 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e757f47-326c-4174-9c28-c577a314f70b-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"5e757f47-326c-4174-9c28-c577a314f70b\") " pod="openstack/cinder-scheduler-0" Oct 05 21:13:20 crc kubenswrapper[4754]: I1005 21:13:20.425726 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5e757f47-326c-4174-9c28-c577a314f70b-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"5e757f47-326c-4174-9c28-c577a314f70b\") " pod="openstack/cinder-scheduler-0" Oct 05 21:13:20 crc kubenswrapper[4754]: I1005 21:13:20.425899 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bmvlt\" (UniqueName: \"kubernetes.io/projected/5e757f47-326c-4174-9c28-c577a314f70b-kube-api-access-bmvlt\") pod \"cinder-scheduler-0\" (UID: \"5e757f47-326c-4174-9c28-c577a314f70b\") " pod="openstack/cinder-scheduler-0" Oct 05 21:13:20 crc kubenswrapper[4754]: I1005 21:13:20.425939 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5e757f47-326c-4174-9c28-c577a314f70b-scripts\") pod \"cinder-scheduler-0\" (UID: \"5e757f47-326c-4174-9c28-c577a314f70b\") " pod="openstack/cinder-scheduler-0" Oct 05 21:13:20 crc kubenswrapper[4754]: I1005 21:13:20.479961 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-59d5ff467f-8sfrm"] Oct 05 21:13:20 crc kubenswrapper[4754]: I1005 21:13:20.531456 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e757f47-326c-4174-9c28-c577a314f70b-config-data\") pod \"cinder-scheduler-0\" (UID: \"5e757f47-326c-4174-9c28-c577a314f70b\") " pod="openstack/cinder-scheduler-0" Oct 05 21:13:20 crc kubenswrapper[4754]: I1005 21:13:20.531523 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5e757f47-326c-4174-9c28-c577a314f70b-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"5e757f47-326c-4174-9c28-c577a314f70b\") " pod="openstack/cinder-scheduler-0" Oct 05 21:13:20 crc kubenswrapper[4754]: I1005 21:13:20.531563 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e757f47-326c-4174-9c28-c577a314f70b-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"5e757f47-326c-4174-9c28-c577a314f70b\") " pod="openstack/cinder-scheduler-0" Oct 05 21:13:20 crc kubenswrapper[4754]: I1005 21:13:20.531588 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5e757f47-326c-4174-9c28-c577a314f70b-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"5e757f47-326c-4174-9c28-c577a314f70b\") " pod="openstack/cinder-scheduler-0" Oct 05 21:13:20 crc kubenswrapper[4754]: I1005 21:13:20.531651 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bmvlt\" (UniqueName: \"kubernetes.io/projected/5e757f47-326c-4174-9c28-c577a314f70b-kube-api-access-bmvlt\") pod \"cinder-scheduler-0\" (UID: \"5e757f47-326c-4174-9c28-c577a314f70b\") " pod="openstack/cinder-scheduler-0" Oct 05 21:13:20 crc kubenswrapper[4754]: I1005 21:13:20.531688 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5e757f47-326c-4174-9c28-c577a314f70b-scripts\") pod \"cinder-scheduler-0\" (UID: \"5e757f47-326c-4174-9c28-c577a314f70b\") " pod="openstack/cinder-scheduler-0" Oct 05 21:13:20 crc kubenswrapper[4754]: I1005 21:13:20.541008 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e757f47-326c-4174-9c28-c577a314f70b-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"5e757f47-326c-4174-9c28-c577a314f70b\") " pod="openstack/cinder-scheduler-0" Oct 05 21:13:20 crc kubenswrapper[4754]: I1005 21:13:20.542728 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5e757f47-326c-4174-9c28-c577a314f70b-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"5e757f47-326c-4174-9c28-c577a314f70b\") " pod="openstack/cinder-scheduler-0" Oct 05 21:13:20 crc kubenswrapper[4754]: I1005 21:13:20.559957 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-69c986f6d7-fksbs"] Oct 05 21:13:20 crc kubenswrapper[4754]: I1005 21:13:20.562210 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5e757f47-326c-4174-9c28-c577a314f70b-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"5e757f47-326c-4174-9c28-c577a314f70b\") " pod="openstack/cinder-scheduler-0" Oct 05 21:13:20 crc kubenswrapper[4754]: I1005 21:13:20.562730 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-69c986f6d7-fksbs" Oct 05 21:13:20 crc kubenswrapper[4754]: I1005 21:13:20.568240 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5e757f47-326c-4174-9c28-c577a314f70b-scripts\") pod \"cinder-scheduler-0\" (UID: \"5e757f47-326c-4174-9c28-c577a314f70b\") " pod="openstack/cinder-scheduler-0" Oct 05 21:13:20 crc kubenswrapper[4754]: I1005 21:13:20.570768 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e757f47-326c-4174-9c28-c577a314f70b-config-data\") pod \"cinder-scheduler-0\" (UID: \"5e757f47-326c-4174-9c28-c577a314f70b\") " pod="openstack/cinder-scheduler-0" Oct 05 21:13:20 crc kubenswrapper[4754]: I1005 21:13:20.589124 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bmvlt\" (UniqueName: \"kubernetes.io/projected/5e757f47-326c-4174-9c28-c577a314f70b-kube-api-access-bmvlt\") pod \"cinder-scheduler-0\" (UID: \"5e757f47-326c-4174-9c28-c577a314f70b\") " pod="openstack/cinder-scheduler-0" Oct 05 21:13:20 crc kubenswrapper[4754]: I1005 21:13:20.599034 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-69c986f6d7-fksbs"] Oct 05 21:13:20 crc kubenswrapper[4754]: I1005 21:13:20.633040 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c79a3ebb-fbdd-4680-aff0-27fbd7d8a7eb-config\") pod \"dnsmasq-dns-69c986f6d7-fksbs\" (UID: \"c79a3ebb-fbdd-4680-aff0-27fbd7d8a7eb\") " pod="openstack/dnsmasq-dns-69c986f6d7-fksbs" Oct 05 21:13:20 crc kubenswrapper[4754]: I1005 21:13:20.633083 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c79a3ebb-fbdd-4680-aff0-27fbd7d8a7eb-ovsdbserver-nb\") pod \"dnsmasq-dns-69c986f6d7-fksbs\" (UID: \"c79a3ebb-fbdd-4680-aff0-27fbd7d8a7eb\") " pod="openstack/dnsmasq-dns-69c986f6d7-fksbs" Oct 05 21:13:20 crc kubenswrapper[4754]: I1005 21:13:20.633140 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c79a3ebb-fbdd-4680-aff0-27fbd7d8a7eb-dns-swift-storage-0\") pod \"dnsmasq-dns-69c986f6d7-fksbs\" (UID: \"c79a3ebb-fbdd-4680-aff0-27fbd7d8a7eb\") " pod="openstack/dnsmasq-dns-69c986f6d7-fksbs" Oct 05 21:13:20 crc kubenswrapper[4754]: I1005 21:13:20.633170 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nkvdp\" (UniqueName: \"kubernetes.io/projected/c79a3ebb-fbdd-4680-aff0-27fbd7d8a7eb-kube-api-access-nkvdp\") pod \"dnsmasq-dns-69c986f6d7-fksbs\" (UID: \"c79a3ebb-fbdd-4680-aff0-27fbd7d8a7eb\") " pod="openstack/dnsmasq-dns-69c986f6d7-fksbs" Oct 05 21:13:20 crc kubenswrapper[4754]: I1005 21:13:20.633214 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c79a3ebb-fbdd-4680-aff0-27fbd7d8a7eb-ovsdbserver-sb\") pod \"dnsmasq-dns-69c986f6d7-fksbs\" (UID: \"c79a3ebb-fbdd-4680-aff0-27fbd7d8a7eb\") " pod="openstack/dnsmasq-dns-69c986f6d7-fksbs" Oct 05 21:13:20 crc kubenswrapper[4754]: I1005 21:13:20.633237 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c79a3ebb-fbdd-4680-aff0-27fbd7d8a7eb-dns-svc\") pod \"dnsmasq-dns-69c986f6d7-fksbs\" (UID: \"c79a3ebb-fbdd-4680-aff0-27fbd7d8a7eb\") " pod="openstack/dnsmasq-dns-69c986f6d7-fksbs" Oct 05 21:13:20 crc kubenswrapper[4754]: I1005 21:13:20.662686 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-68c6b67864-9msm5" podUID="27cd6be5-f5eb-4158-bdd1-37f4df3d530d" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.148:8443/dashboard/auth/login/?next=/dashboard/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 05 21:13:20 crc kubenswrapper[4754]: I1005 21:13:20.667846 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-59d5ff467f-8sfrm" Oct 05 21:13:20 crc kubenswrapper[4754]: I1005 21:13:20.731248 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-574454d6cb-vbnk6" podUID="39d1aac3-14a1-45bf-9a9d-ba4cdb0afcb1" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.149:8443/dashboard/auth/login/?next=/dashboard/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 05 21:13:20 crc kubenswrapper[4754]: I1005 21:13:20.736656 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c79a3ebb-fbdd-4680-aff0-27fbd7d8a7eb-ovsdbserver-sb\") pod \"dnsmasq-dns-69c986f6d7-fksbs\" (UID: \"c79a3ebb-fbdd-4680-aff0-27fbd7d8a7eb\") " pod="openstack/dnsmasq-dns-69c986f6d7-fksbs" Oct 05 21:13:20 crc kubenswrapper[4754]: I1005 21:13:20.736721 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c79a3ebb-fbdd-4680-aff0-27fbd7d8a7eb-dns-svc\") pod \"dnsmasq-dns-69c986f6d7-fksbs\" (UID: \"c79a3ebb-fbdd-4680-aff0-27fbd7d8a7eb\") " pod="openstack/dnsmasq-dns-69c986f6d7-fksbs" Oct 05 21:13:20 crc kubenswrapper[4754]: I1005 21:13:20.736937 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c79a3ebb-fbdd-4680-aff0-27fbd7d8a7eb-config\") pod \"dnsmasq-dns-69c986f6d7-fksbs\" (UID: \"c79a3ebb-fbdd-4680-aff0-27fbd7d8a7eb\") " pod="openstack/dnsmasq-dns-69c986f6d7-fksbs" Oct 05 21:13:20 crc kubenswrapper[4754]: I1005 21:13:20.736972 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c79a3ebb-fbdd-4680-aff0-27fbd7d8a7eb-ovsdbserver-nb\") pod \"dnsmasq-dns-69c986f6d7-fksbs\" (UID: \"c79a3ebb-fbdd-4680-aff0-27fbd7d8a7eb\") " pod="openstack/dnsmasq-dns-69c986f6d7-fksbs" Oct 05 21:13:20 crc kubenswrapper[4754]: I1005 21:13:20.737183 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c79a3ebb-fbdd-4680-aff0-27fbd7d8a7eb-dns-swift-storage-0\") pod \"dnsmasq-dns-69c986f6d7-fksbs\" (UID: \"c79a3ebb-fbdd-4680-aff0-27fbd7d8a7eb\") " pod="openstack/dnsmasq-dns-69c986f6d7-fksbs" Oct 05 21:13:20 crc kubenswrapper[4754]: I1005 21:13:20.737209 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nkvdp\" (UniqueName: \"kubernetes.io/projected/c79a3ebb-fbdd-4680-aff0-27fbd7d8a7eb-kube-api-access-nkvdp\") pod \"dnsmasq-dns-69c986f6d7-fksbs\" (UID: \"c79a3ebb-fbdd-4680-aff0-27fbd7d8a7eb\") " pod="openstack/dnsmasq-dns-69c986f6d7-fksbs" Oct 05 21:13:20 crc kubenswrapper[4754]: I1005 21:13:20.738153 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c79a3ebb-fbdd-4680-aff0-27fbd7d8a7eb-ovsdbserver-sb\") pod \"dnsmasq-dns-69c986f6d7-fksbs\" (UID: \"c79a3ebb-fbdd-4680-aff0-27fbd7d8a7eb\") " pod="openstack/dnsmasq-dns-69c986f6d7-fksbs" Oct 05 21:13:20 crc kubenswrapper[4754]: I1005 21:13:20.738554 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c79a3ebb-fbdd-4680-aff0-27fbd7d8a7eb-dns-svc\") pod \"dnsmasq-dns-69c986f6d7-fksbs\" (UID: \"c79a3ebb-fbdd-4680-aff0-27fbd7d8a7eb\") " pod="openstack/dnsmasq-dns-69c986f6d7-fksbs" Oct 05 21:13:20 crc kubenswrapper[4754]: I1005 21:13:20.740370 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c79a3ebb-fbdd-4680-aff0-27fbd7d8a7eb-ovsdbserver-nb\") pod \"dnsmasq-dns-69c986f6d7-fksbs\" (UID: \"c79a3ebb-fbdd-4680-aff0-27fbd7d8a7eb\") " pod="openstack/dnsmasq-dns-69c986f6d7-fksbs" Oct 05 21:13:20 crc kubenswrapper[4754]: I1005 21:13:20.740459 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c79a3ebb-fbdd-4680-aff0-27fbd7d8a7eb-config\") pod \"dnsmasq-dns-69c986f6d7-fksbs\" (UID: \"c79a3ebb-fbdd-4680-aff0-27fbd7d8a7eb\") " pod="openstack/dnsmasq-dns-69c986f6d7-fksbs" Oct 05 21:13:20 crc kubenswrapper[4754]: I1005 21:13:20.746803 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c79a3ebb-fbdd-4680-aff0-27fbd7d8a7eb-dns-swift-storage-0\") pod \"dnsmasq-dns-69c986f6d7-fksbs\" (UID: \"c79a3ebb-fbdd-4680-aff0-27fbd7d8a7eb\") " pod="openstack/dnsmasq-dns-69c986f6d7-fksbs" Oct 05 21:13:20 crc kubenswrapper[4754]: I1005 21:13:20.747992 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Oct 05 21:13:20 crc kubenswrapper[4754]: I1005 21:13:20.750468 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 05 21:13:20 crc kubenswrapper[4754]: I1005 21:13:20.756798 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Oct 05 21:13:20 crc kubenswrapper[4754]: I1005 21:13:20.772414 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nkvdp\" (UniqueName: \"kubernetes.io/projected/c79a3ebb-fbdd-4680-aff0-27fbd7d8a7eb-kube-api-access-nkvdp\") pod \"dnsmasq-dns-69c986f6d7-fksbs\" (UID: \"c79a3ebb-fbdd-4680-aff0-27fbd7d8a7eb\") " pod="openstack/dnsmasq-dns-69c986f6d7-fksbs" Oct 05 21:13:20 crc kubenswrapper[4754]: I1005 21:13:20.845461 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 05 21:13:20 crc kubenswrapper[4754]: I1005 21:13:20.869391 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 05 21:13:20 crc kubenswrapper[4754]: I1005 21:13:20.947176 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0e4c84d5-ce7d-4b40-9b7c-980bc6b16c90-config-data-custom\") pod \"cinder-api-0\" (UID: \"0e4c84d5-ce7d-4b40-9b7c-980bc6b16c90\") " pod="openstack/cinder-api-0" Oct 05 21:13:20 crc kubenswrapper[4754]: I1005 21:13:20.947223 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e4c84d5-ce7d-4b40-9b7c-980bc6b16c90-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"0e4c84d5-ce7d-4b40-9b7c-980bc6b16c90\") " pod="openstack/cinder-api-0" Oct 05 21:13:20 crc kubenswrapper[4754]: I1005 21:13:20.947259 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gnckd\" (UniqueName: \"kubernetes.io/projected/0e4c84d5-ce7d-4b40-9b7c-980bc6b16c90-kube-api-access-gnckd\") pod \"cinder-api-0\" (UID: \"0e4c84d5-ce7d-4b40-9b7c-980bc6b16c90\") " pod="openstack/cinder-api-0" Oct 05 21:13:20 crc kubenswrapper[4754]: I1005 21:13:20.947283 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0e4c84d5-ce7d-4b40-9b7c-980bc6b16c90-etc-machine-id\") pod \"cinder-api-0\" (UID: \"0e4c84d5-ce7d-4b40-9b7c-980bc6b16c90\") " pod="openstack/cinder-api-0" Oct 05 21:13:20 crc kubenswrapper[4754]: I1005 21:13:20.947301 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0e4c84d5-ce7d-4b40-9b7c-980bc6b16c90-scripts\") pod \"cinder-api-0\" (UID: \"0e4c84d5-ce7d-4b40-9b7c-980bc6b16c90\") " pod="openstack/cinder-api-0" Oct 05 21:13:20 crc kubenswrapper[4754]: I1005 21:13:20.947325 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e4c84d5-ce7d-4b40-9b7c-980bc6b16c90-config-data\") pod \"cinder-api-0\" (UID: \"0e4c84d5-ce7d-4b40-9b7c-980bc6b16c90\") " pod="openstack/cinder-api-0" Oct 05 21:13:20 crc kubenswrapper[4754]: I1005 21:13:20.947357 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0e4c84d5-ce7d-4b40-9b7c-980bc6b16c90-logs\") pod \"cinder-api-0\" (UID: \"0e4c84d5-ce7d-4b40-9b7c-980bc6b16c90\") " pod="openstack/cinder-api-0" Oct 05 21:13:20 crc kubenswrapper[4754]: I1005 21:13:20.973177 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-69c986f6d7-fksbs" Oct 05 21:13:21 crc kubenswrapper[4754]: I1005 21:13:21.048905 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0e4c84d5-ce7d-4b40-9b7c-980bc6b16c90-config-data-custom\") pod \"cinder-api-0\" (UID: \"0e4c84d5-ce7d-4b40-9b7c-980bc6b16c90\") " pod="openstack/cinder-api-0" Oct 05 21:13:21 crc kubenswrapper[4754]: I1005 21:13:21.048974 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e4c84d5-ce7d-4b40-9b7c-980bc6b16c90-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"0e4c84d5-ce7d-4b40-9b7c-980bc6b16c90\") " pod="openstack/cinder-api-0" Oct 05 21:13:21 crc kubenswrapper[4754]: I1005 21:13:21.049008 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gnckd\" (UniqueName: \"kubernetes.io/projected/0e4c84d5-ce7d-4b40-9b7c-980bc6b16c90-kube-api-access-gnckd\") pod \"cinder-api-0\" (UID: \"0e4c84d5-ce7d-4b40-9b7c-980bc6b16c90\") " pod="openstack/cinder-api-0" Oct 05 21:13:21 crc kubenswrapper[4754]: I1005 21:13:21.049056 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0e4c84d5-ce7d-4b40-9b7c-980bc6b16c90-etc-machine-id\") pod \"cinder-api-0\" (UID: \"0e4c84d5-ce7d-4b40-9b7c-980bc6b16c90\") " pod="openstack/cinder-api-0" Oct 05 21:13:21 crc kubenswrapper[4754]: I1005 21:13:21.049074 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0e4c84d5-ce7d-4b40-9b7c-980bc6b16c90-scripts\") pod \"cinder-api-0\" (UID: \"0e4c84d5-ce7d-4b40-9b7c-980bc6b16c90\") " pod="openstack/cinder-api-0" Oct 05 21:13:21 crc kubenswrapper[4754]: I1005 21:13:21.049109 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e4c84d5-ce7d-4b40-9b7c-980bc6b16c90-config-data\") pod \"cinder-api-0\" (UID: \"0e4c84d5-ce7d-4b40-9b7c-980bc6b16c90\") " pod="openstack/cinder-api-0" Oct 05 21:13:21 crc kubenswrapper[4754]: I1005 21:13:21.049177 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0e4c84d5-ce7d-4b40-9b7c-980bc6b16c90-logs\") pod \"cinder-api-0\" (UID: \"0e4c84d5-ce7d-4b40-9b7c-980bc6b16c90\") " pod="openstack/cinder-api-0" Oct 05 21:13:21 crc kubenswrapper[4754]: I1005 21:13:21.049957 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0e4c84d5-ce7d-4b40-9b7c-980bc6b16c90-etc-machine-id\") pod \"cinder-api-0\" (UID: \"0e4c84d5-ce7d-4b40-9b7c-980bc6b16c90\") " pod="openstack/cinder-api-0" Oct 05 21:13:21 crc kubenswrapper[4754]: I1005 21:13:21.050877 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0e4c84d5-ce7d-4b40-9b7c-980bc6b16c90-logs\") pod \"cinder-api-0\" (UID: \"0e4c84d5-ce7d-4b40-9b7c-980bc6b16c90\") " pod="openstack/cinder-api-0" Oct 05 21:13:21 crc kubenswrapper[4754]: I1005 21:13:21.055890 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e4c84d5-ce7d-4b40-9b7c-980bc6b16c90-config-data\") pod \"cinder-api-0\" (UID: \"0e4c84d5-ce7d-4b40-9b7c-980bc6b16c90\") " pod="openstack/cinder-api-0" Oct 05 21:13:21 crc kubenswrapper[4754]: I1005 21:13:21.056692 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0e4c84d5-ce7d-4b40-9b7c-980bc6b16c90-scripts\") pod \"cinder-api-0\" (UID: \"0e4c84d5-ce7d-4b40-9b7c-980bc6b16c90\") " pod="openstack/cinder-api-0" Oct 05 21:13:21 crc kubenswrapper[4754]: I1005 21:13:21.058577 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0e4c84d5-ce7d-4b40-9b7c-980bc6b16c90-config-data-custom\") pod \"cinder-api-0\" (UID: \"0e4c84d5-ce7d-4b40-9b7c-980bc6b16c90\") " pod="openstack/cinder-api-0" Oct 05 21:13:21 crc kubenswrapper[4754]: I1005 21:13:21.070799 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e4c84d5-ce7d-4b40-9b7c-980bc6b16c90-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"0e4c84d5-ce7d-4b40-9b7c-980bc6b16c90\") " pod="openstack/cinder-api-0" Oct 05 21:13:21 crc kubenswrapper[4754]: I1005 21:13:21.086369 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gnckd\" (UniqueName: \"kubernetes.io/projected/0e4c84d5-ce7d-4b40-9b7c-980bc6b16c90-kube-api-access-gnckd\") pod \"cinder-api-0\" (UID: \"0e4c84d5-ce7d-4b40-9b7c-980bc6b16c90\") " pod="openstack/cinder-api-0" Oct 05 21:13:21 crc kubenswrapper[4754]: I1005 21:13:21.163724 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 05 21:13:21 crc kubenswrapper[4754]: I1005 21:13:21.678574 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-59d5ff467f-8sfrm" podUID="3ba3d75f-7b8c-490f-b950-4b11fad2573c" containerName="dnsmasq-dns" containerID="cri-o://591b748c000024e5fbfc6c9f33557e705a206aef0da8a9757f00d9f49c6dce1f" gracePeriod=10 Oct 05 21:13:22 crc kubenswrapper[4754]: I1005 21:13:22.554735 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59d5ff467f-8sfrm" Oct 05 21:13:22 crc kubenswrapper[4754]: I1005 21:13:22.685549 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3ba3d75f-7b8c-490f-b950-4b11fad2573c-ovsdbserver-nb\") pod \"3ba3d75f-7b8c-490f-b950-4b11fad2573c\" (UID: \"3ba3d75f-7b8c-490f-b950-4b11fad2573c\") " Oct 05 21:13:22 crc kubenswrapper[4754]: I1005 21:13:22.685710 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3ba3d75f-7b8c-490f-b950-4b11fad2573c-dns-swift-storage-0\") pod \"3ba3d75f-7b8c-490f-b950-4b11fad2573c\" (UID: \"3ba3d75f-7b8c-490f-b950-4b11fad2573c\") " Oct 05 21:13:22 crc kubenswrapper[4754]: I1005 21:13:22.685776 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3ba3d75f-7b8c-490f-b950-4b11fad2573c-config\") pod \"3ba3d75f-7b8c-490f-b950-4b11fad2573c\" (UID: \"3ba3d75f-7b8c-490f-b950-4b11fad2573c\") " Oct 05 21:13:22 crc kubenswrapper[4754]: I1005 21:13:22.685795 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3ba3d75f-7b8c-490f-b950-4b11fad2573c-dns-svc\") pod \"3ba3d75f-7b8c-490f-b950-4b11fad2573c\" (UID: \"3ba3d75f-7b8c-490f-b950-4b11fad2573c\") " Oct 05 21:13:22 crc kubenswrapper[4754]: I1005 21:13:22.685869 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3ba3d75f-7b8c-490f-b950-4b11fad2573c-ovsdbserver-sb\") pod \"3ba3d75f-7b8c-490f-b950-4b11fad2573c\" (UID: \"3ba3d75f-7b8c-490f-b950-4b11fad2573c\") " Oct 05 21:13:22 crc kubenswrapper[4754]: I1005 21:13:22.685901 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7w2vt\" (UniqueName: \"kubernetes.io/projected/3ba3d75f-7b8c-490f-b950-4b11fad2573c-kube-api-access-7w2vt\") pod \"3ba3d75f-7b8c-490f-b950-4b11fad2573c\" (UID: \"3ba3d75f-7b8c-490f-b950-4b11fad2573c\") " Oct 05 21:13:22 crc kubenswrapper[4754]: I1005 21:13:22.722305 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ba3d75f-7b8c-490f-b950-4b11fad2573c-kube-api-access-7w2vt" (OuterVolumeSpecName: "kube-api-access-7w2vt") pod "3ba3d75f-7b8c-490f-b950-4b11fad2573c" (UID: "3ba3d75f-7b8c-490f-b950-4b11fad2573c"). InnerVolumeSpecName "kube-api-access-7w2vt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:13:22 crc kubenswrapper[4754]: I1005 21:13:22.730030 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 05 21:13:22 crc kubenswrapper[4754]: I1005 21:13:22.737031 4754 generic.go:334] "Generic (PLEG): container finished" podID="3ba3d75f-7b8c-490f-b950-4b11fad2573c" containerID="591b748c000024e5fbfc6c9f33557e705a206aef0da8a9757f00d9f49c6dce1f" exitCode=0 Oct 05 21:13:22 crc kubenswrapper[4754]: I1005 21:13:22.737086 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59d5ff467f-8sfrm" event={"ID":"3ba3d75f-7b8c-490f-b950-4b11fad2573c","Type":"ContainerDied","Data":"591b748c000024e5fbfc6c9f33557e705a206aef0da8a9757f00d9f49c6dce1f"} Oct 05 21:13:22 crc kubenswrapper[4754]: I1005 21:13:22.737120 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59d5ff467f-8sfrm" event={"ID":"3ba3d75f-7b8c-490f-b950-4b11fad2573c","Type":"ContainerDied","Data":"f2d70caed95a6844b6110b82857e137d7878f5779a72ac9128bf58c089c22a71"} Oct 05 21:13:22 crc kubenswrapper[4754]: I1005 21:13:22.737136 4754 scope.go:117] "RemoveContainer" containerID="591b748c000024e5fbfc6c9f33557e705a206aef0da8a9757f00d9f49c6dce1f" Oct 05 21:13:22 crc kubenswrapper[4754]: I1005 21:13:22.737285 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59d5ff467f-8sfrm" Oct 05 21:13:22 crc kubenswrapper[4754]: I1005 21:13:22.789030 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7w2vt\" (UniqueName: \"kubernetes.io/projected/3ba3d75f-7b8c-490f-b950-4b11fad2573c-kube-api-access-7w2vt\") on node \"crc\" DevicePath \"\"" Oct 05 21:13:22 crc kubenswrapper[4754]: I1005 21:13:22.862559 4754 scope.go:117] "RemoveContainer" containerID="aa4560e4bee403604979bda67fd34b6f05aa1466c510822d70fc137f09672782" Oct 05 21:13:22 crc kubenswrapper[4754]: I1005 21:13:22.966792 4754 scope.go:117] "RemoveContainer" containerID="591b748c000024e5fbfc6c9f33557e705a206aef0da8a9757f00d9f49c6dce1f" Oct 05 21:13:22 crc kubenswrapper[4754]: E1005 21:13:22.970774 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"591b748c000024e5fbfc6c9f33557e705a206aef0da8a9757f00d9f49c6dce1f\": container with ID starting with 591b748c000024e5fbfc6c9f33557e705a206aef0da8a9757f00d9f49c6dce1f not found: ID does not exist" containerID="591b748c000024e5fbfc6c9f33557e705a206aef0da8a9757f00d9f49c6dce1f" Oct 05 21:13:22 crc kubenswrapper[4754]: I1005 21:13:22.970812 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"591b748c000024e5fbfc6c9f33557e705a206aef0da8a9757f00d9f49c6dce1f"} err="failed to get container status \"591b748c000024e5fbfc6c9f33557e705a206aef0da8a9757f00d9f49c6dce1f\": rpc error: code = NotFound desc = could not find container \"591b748c000024e5fbfc6c9f33557e705a206aef0da8a9757f00d9f49c6dce1f\": container with ID starting with 591b748c000024e5fbfc6c9f33557e705a206aef0da8a9757f00d9f49c6dce1f not found: ID does not exist" Oct 05 21:13:22 crc kubenswrapper[4754]: I1005 21:13:22.970842 4754 scope.go:117] "RemoveContainer" containerID="aa4560e4bee403604979bda67fd34b6f05aa1466c510822d70fc137f09672782" Oct 05 21:13:22 crc kubenswrapper[4754]: E1005 21:13:22.974621 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aa4560e4bee403604979bda67fd34b6f05aa1466c510822d70fc137f09672782\": container with ID starting with aa4560e4bee403604979bda67fd34b6f05aa1466c510822d70fc137f09672782 not found: ID does not exist" containerID="aa4560e4bee403604979bda67fd34b6f05aa1466c510822d70fc137f09672782" Oct 05 21:13:22 crc kubenswrapper[4754]: I1005 21:13:22.974708 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aa4560e4bee403604979bda67fd34b6f05aa1466c510822d70fc137f09672782"} err="failed to get container status \"aa4560e4bee403604979bda67fd34b6f05aa1466c510822d70fc137f09672782\": rpc error: code = NotFound desc = could not find container \"aa4560e4bee403604979bda67fd34b6f05aa1466c510822d70fc137f09672782\": container with ID starting with aa4560e4bee403604979bda67fd34b6f05aa1466c510822d70fc137f09672782 not found: ID does not exist" Oct 05 21:13:23 crc kubenswrapper[4754]: I1005 21:13:23.030974 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-765c6857c8-q7glx"] Oct 05 21:13:23 crc kubenswrapper[4754]: I1005 21:13:23.168898 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3ba3d75f-7b8c-490f-b950-4b11fad2573c-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "3ba3d75f-7b8c-490f-b950-4b11fad2573c" (UID: "3ba3d75f-7b8c-490f-b950-4b11fad2573c"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 21:13:23 crc kubenswrapper[4754]: I1005 21:13:23.196882 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3ba3d75f-7b8c-490f-b950-4b11fad2573c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "3ba3d75f-7b8c-490f-b950-4b11fad2573c" (UID: "3ba3d75f-7b8c-490f-b950-4b11fad2573c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 21:13:23 crc kubenswrapper[4754]: I1005 21:13:23.229935 4754 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3ba3d75f-7b8c-490f-b950-4b11fad2573c-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 05 21:13:23 crc kubenswrapper[4754]: I1005 21:13:23.229971 4754 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3ba3d75f-7b8c-490f-b950-4b11fad2573c-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 05 21:13:23 crc kubenswrapper[4754]: I1005 21:13:23.266104 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3ba3d75f-7b8c-490f-b950-4b11fad2573c-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "3ba3d75f-7b8c-490f-b950-4b11fad2573c" (UID: "3ba3d75f-7b8c-490f-b950-4b11fad2573c"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 21:13:23 crc kubenswrapper[4754]: I1005 21:13:23.307236 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3ba3d75f-7b8c-490f-b950-4b11fad2573c-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "3ba3d75f-7b8c-490f-b950-4b11fad2573c" (UID: "3ba3d75f-7b8c-490f-b950-4b11fad2573c"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 21:13:23 crc kubenswrapper[4754]: I1005 21:13:23.335202 4754 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3ba3d75f-7b8c-490f-b950-4b11fad2573c-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 05 21:13:23 crc kubenswrapper[4754]: I1005 21:13:23.335242 4754 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3ba3d75f-7b8c-490f-b950-4b11fad2573c-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 05 21:13:23 crc kubenswrapper[4754]: I1005 21:13:23.446939 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-69c986f6d7-fksbs"] Oct 05 21:13:23 crc kubenswrapper[4754]: I1005 21:13:23.544944 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3ba3d75f-7b8c-490f-b950-4b11fad2573c-config" (OuterVolumeSpecName: "config") pod "3ba3d75f-7b8c-490f-b950-4b11fad2573c" (UID: "3ba3d75f-7b8c-490f-b950-4b11fad2573c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 21:13:23 crc kubenswrapper[4754]: I1005 21:13:23.643968 4754 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3ba3d75f-7b8c-490f-b950-4b11fad2573c-config\") on node \"crc\" DevicePath \"\"" Oct 05 21:13:23 crc kubenswrapper[4754]: I1005 21:13:23.669146 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 05 21:13:23 crc kubenswrapper[4754]: I1005 21:13:23.710544 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-59d5ff467f-8sfrm"] Oct 05 21:13:23 crc kubenswrapper[4754]: I1005 21:13:23.731398 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-59d5ff467f-8sfrm"] Oct 05 21:13:23 crc kubenswrapper[4754]: I1005 21:13:23.813363 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"0e4c84d5-ce7d-4b40-9b7c-980bc6b16c90","Type":"ContainerStarted","Data":"81e325c0a77688d8937756abb6992b384f46301927253d461f8c7925e40ce31e"} Oct 05 21:13:23 crc kubenswrapper[4754]: I1005 21:13:23.814489 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-54b7554787-vqrt4" event={"ID":"346d8b39-98ea-442d-910e-b40a6e4b85c4","Type":"ContainerStarted","Data":"9d10aa2437c9c2221eb4fcea2447e757af5ad5b3b21f8ff0ba439f355d4148c3"} Oct 05 21:13:23 crc kubenswrapper[4754]: I1005 21:13:23.822756 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-765c6857c8-q7glx" event={"ID":"bd23550d-8815-4211-b8be-10ee8ab95c1f","Type":"ContainerStarted","Data":"99638c99a11850d8b3a25c39e5ee9f9519e8845cca9c3d2142c3a03580fe9330"} Oct 05 21:13:23 crc kubenswrapper[4754]: I1005 21:13:23.838441 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-69c986f6d7-fksbs" event={"ID":"c79a3ebb-fbdd-4680-aff0-27fbd7d8a7eb","Type":"ContainerStarted","Data":"ce0e640bbc5d27b424afee479a96ad561119313962435a85170dfa605e5931a7"} Oct 05 21:13:24 crc kubenswrapper[4754]: I1005 21:13:24.122082 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Oct 05 21:13:24 crc kubenswrapper[4754]: E1005 21:13:24.123013 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ba3d75f-7b8c-490f-b950-4b11fad2573c" containerName="init" Oct 05 21:13:24 crc kubenswrapper[4754]: I1005 21:13:24.123043 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ba3d75f-7b8c-490f-b950-4b11fad2573c" containerName="init" Oct 05 21:13:24 crc kubenswrapper[4754]: E1005 21:13:24.123092 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ba3d75f-7b8c-490f-b950-4b11fad2573c" containerName="dnsmasq-dns" Oct 05 21:13:24 crc kubenswrapper[4754]: I1005 21:13:24.123103 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ba3d75f-7b8c-490f-b950-4b11fad2573c" containerName="dnsmasq-dns" Oct 05 21:13:24 crc kubenswrapper[4754]: I1005 21:13:24.123358 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ba3d75f-7b8c-490f-b950-4b11fad2573c" containerName="dnsmasq-dns" Oct 05 21:13:24 crc kubenswrapper[4754]: I1005 21:13:24.124286 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 05 21:13:24 crc kubenswrapper[4754]: I1005 21:13:24.129771 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-skdf5" Oct 05 21:13:24 crc kubenswrapper[4754]: I1005 21:13:24.130404 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Oct 05 21:13:24 crc kubenswrapper[4754]: I1005 21:13:24.135202 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Oct 05 21:13:24 crc kubenswrapper[4754]: I1005 21:13:24.138237 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 05 21:13:24 crc kubenswrapper[4754]: I1005 21:13:24.165429 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/95ef109d-c48b-40ea-b491-3e730d4d651b-openstack-config-secret\") pod \"openstackclient\" (UID: \"95ef109d-c48b-40ea-b491-3e730d4d651b\") " pod="openstack/openstackclient" Oct 05 21:13:24 crc kubenswrapper[4754]: I1005 21:13:24.165514 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/95ef109d-c48b-40ea-b491-3e730d4d651b-openstack-config\") pod \"openstackclient\" (UID: \"95ef109d-c48b-40ea-b491-3e730d4d651b\") " pod="openstack/openstackclient" Oct 05 21:13:24 crc kubenswrapper[4754]: I1005 21:13:24.165568 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95ef109d-c48b-40ea-b491-3e730d4d651b-combined-ca-bundle\") pod \"openstackclient\" (UID: \"95ef109d-c48b-40ea-b491-3e730d4d651b\") " pod="openstack/openstackclient" Oct 05 21:13:24 crc kubenswrapper[4754]: I1005 21:13:24.165590 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x2t4w\" (UniqueName: \"kubernetes.io/projected/95ef109d-c48b-40ea-b491-3e730d4d651b-kube-api-access-x2t4w\") pod \"openstackclient\" (UID: \"95ef109d-c48b-40ea-b491-3e730d4d651b\") " pod="openstack/openstackclient" Oct 05 21:13:24 crc kubenswrapper[4754]: I1005 21:13:24.271662 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95ef109d-c48b-40ea-b491-3e730d4d651b-combined-ca-bundle\") pod \"openstackclient\" (UID: \"95ef109d-c48b-40ea-b491-3e730d4d651b\") " pod="openstack/openstackclient" Oct 05 21:13:24 crc kubenswrapper[4754]: I1005 21:13:24.271701 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x2t4w\" (UniqueName: \"kubernetes.io/projected/95ef109d-c48b-40ea-b491-3e730d4d651b-kube-api-access-x2t4w\") pod \"openstackclient\" (UID: \"95ef109d-c48b-40ea-b491-3e730d4d651b\") " pod="openstack/openstackclient" Oct 05 21:13:24 crc kubenswrapper[4754]: I1005 21:13:24.271809 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/95ef109d-c48b-40ea-b491-3e730d4d651b-openstack-config-secret\") pod \"openstackclient\" (UID: \"95ef109d-c48b-40ea-b491-3e730d4d651b\") " pod="openstack/openstackclient" Oct 05 21:13:24 crc kubenswrapper[4754]: I1005 21:13:24.271833 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/95ef109d-c48b-40ea-b491-3e730d4d651b-openstack-config\") pod \"openstackclient\" (UID: \"95ef109d-c48b-40ea-b491-3e730d4d651b\") " pod="openstack/openstackclient" Oct 05 21:13:24 crc kubenswrapper[4754]: I1005 21:13:24.281032 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/95ef109d-c48b-40ea-b491-3e730d4d651b-openstack-config\") pod \"openstackclient\" (UID: \"95ef109d-c48b-40ea-b491-3e730d4d651b\") " pod="openstack/openstackclient" Oct 05 21:13:24 crc kubenswrapper[4754]: I1005 21:13:24.286682 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/95ef109d-c48b-40ea-b491-3e730d4d651b-openstack-config-secret\") pod \"openstackclient\" (UID: \"95ef109d-c48b-40ea-b491-3e730d4d651b\") " pod="openstack/openstackclient" Oct 05 21:13:24 crc kubenswrapper[4754]: I1005 21:13:24.293840 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95ef109d-c48b-40ea-b491-3e730d4d651b-combined-ca-bundle\") pod \"openstackclient\" (UID: \"95ef109d-c48b-40ea-b491-3e730d4d651b\") " pod="openstack/openstackclient" Oct 05 21:13:24 crc kubenswrapper[4754]: I1005 21:13:24.293875 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x2t4w\" (UniqueName: \"kubernetes.io/projected/95ef109d-c48b-40ea-b491-3e730d4d651b-kube-api-access-x2t4w\") pod \"openstackclient\" (UID: \"95ef109d-c48b-40ea-b491-3e730d4d651b\") " pod="openstack/openstackclient" Oct 05 21:13:24 crc kubenswrapper[4754]: I1005 21:13:24.428301 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 05 21:13:24 crc kubenswrapper[4754]: I1005 21:13:24.478913 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 05 21:13:24 crc kubenswrapper[4754]: I1005 21:13:24.875307 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ba3d75f-7b8c-490f-b950-4b11fad2573c" path="/var/lib/kubelet/pods/3ba3d75f-7b8c-490f-b950-4b11fad2573c/volumes" Oct 05 21:13:24 crc kubenswrapper[4754]: I1005 21:13:24.928438 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-9dc4bfc9d-cfm98" event={"ID":"358e33ff-b4ce-46e3-b077-e1df373f2a75","Type":"ContainerStarted","Data":"6578759701caf722665406a0e088cbbc66101a235ccb2e0889ca7abeb9c61153"} Oct 05 21:13:24 crc kubenswrapper[4754]: I1005 21:13:24.972551 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"0e4c84d5-ce7d-4b40-9b7c-980bc6b16c90","Type":"ContainerStarted","Data":"f0c9444516b4ed2de97be88fcc993a6d95e861794de12525343dfdf8876fd1fb"} Oct 05 21:13:24 crc kubenswrapper[4754]: I1005 21:13:24.988817 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-54b7554787-vqrt4" event={"ID":"346d8b39-98ea-442d-910e-b40a6e4b85c4","Type":"ContainerStarted","Data":"e9519c652ec8b0a82f4a8470b7ff94e630b92a076c5f5816565cae1bfe9ab630"} Oct 05 21:13:24 crc kubenswrapper[4754]: I1005 21:13:24.991777 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-765c6857c8-q7glx" event={"ID":"bd23550d-8815-4211-b8be-10ee8ab95c1f","Type":"ContainerStarted","Data":"93674f9b31332a3e4c58889b1f5129cb25a3015d04e4bf62424cac3e52347e41"} Oct 05 21:13:24 crc kubenswrapper[4754]: I1005 21:13:24.993474 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-69c986f6d7-fksbs" event={"ID":"c79a3ebb-fbdd-4680-aff0-27fbd7d8a7eb","Type":"ContainerStarted","Data":"21ba4cc396f6c99a169e2dfa92b344ae6aa2976e2c3fc8e4fd96e75d9f6ca6c0"} Oct 05 21:13:25 crc kubenswrapper[4754]: I1005 21:13:25.001134 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"5e757f47-326c-4174-9c28-c577a314f70b","Type":"ContainerStarted","Data":"d2230dbff5c9e90a0a5d763f1160225660a5953534e2327fdd30590657c1a469"} Oct 05 21:13:25 crc kubenswrapper[4754]: I1005 21:13:25.065107 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-54b7554787-vqrt4" podStartSLOduration=4.734809231 podStartE2EDuration="10.065085469s" podCreationTimestamp="2025-10-05 21:13:15 +0000 UTC" firstStartedPulling="2025-10-05 21:13:16.960828007 +0000 UTC m=+1120.864946717" lastFinishedPulling="2025-10-05 21:13:22.291104245 +0000 UTC m=+1126.195222955" observedRunningTime="2025-10-05 21:13:25.014238971 +0000 UTC m=+1128.918357681" watchObservedRunningTime="2025-10-05 21:13:25.065085469 +0000 UTC m=+1128.969204179" Oct 05 21:13:25 crc kubenswrapper[4754]: I1005 21:13:25.071551 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 05 21:13:26 crc kubenswrapper[4754]: I1005 21:13:26.039067 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"95ef109d-c48b-40ea-b491-3e730d4d651b","Type":"ContainerStarted","Data":"ce6f8526fc32f8facbda0cdbe2746716165f11d9aba8063d5ca06477479186a0"} Oct 05 21:13:26 crc kubenswrapper[4754]: I1005 21:13:26.054892 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-765c6857c8-q7glx" event={"ID":"bd23550d-8815-4211-b8be-10ee8ab95c1f","Type":"ContainerStarted","Data":"fcb022fd4d7c766e94bd9648f71e997601f5416ee92589702a6231b6699fb36a"} Oct 05 21:13:26 crc kubenswrapper[4754]: I1005 21:13:26.057108 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-765c6857c8-q7glx" Oct 05 21:13:26 crc kubenswrapper[4754]: I1005 21:13:26.057133 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-765c6857c8-q7glx" Oct 05 21:13:26 crc kubenswrapper[4754]: I1005 21:13:26.069551 4754 generic.go:334] "Generic (PLEG): container finished" podID="c79a3ebb-fbdd-4680-aff0-27fbd7d8a7eb" containerID="21ba4cc396f6c99a169e2dfa92b344ae6aa2976e2c3fc8e4fd96e75d9f6ca6c0" exitCode=0 Oct 05 21:13:26 crc kubenswrapper[4754]: I1005 21:13:26.069639 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-69c986f6d7-fksbs" event={"ID":"c79a3ebb-fbdd-4680-aff0-27fbd7d8a7eb","Type":"ContainerDied","Data":"21ba4cc396f6c99a169e2dfa92b344ae6aa2976e2c3fc8e4fd96e75d9f6ca6c0"} Oct 05 21:13:26 crc kubenswrapper[4754]: I1005 21:13:26.069685 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-69c986f6d7-fksbs" event={"ID":"c79a3ebb-fbdd-4680-aff0-27fbd7d8a7eb","Type":"ContainerStarted","Data":"7bc4934b991c3e3b09d48b6d180324d91500beee9d23471553fa9e0b511ea99d"} Oct 05 21:13:26 crc kubenswrapper[4754]: I1005 21:13:26.070754 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-69c986f6d7-fksbs" Oct 05 21:13:26 crc kubenswrapper[4754]: I1005 21:13:26.084760 4754 generic.go:334] "Generic (PLEG): container finished" podID="94c73d9b-e224-4788-995d-be11dcee7b3c" containerID="beb03566dca75372fd1c3a5f05aae0ee0555748b120f181ee2ca7eb6e2d97a71" exitCode=0 Oct 05 21:13:26 crc kubenswrapper[4754]: I1005 21:13:26.084933 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-xlw68" event={"ID":"94c73d9b-e224-4788-995d-be11dcee7b3c","Type":"ContainerDied","Data":"beb03566dca75372fd1c3a5f05aae0ee0555748b120f181ee2ca7eb6e2d97a71"} Oct 05 21:13:26 crc kubenswrapper[4754]: I1005 21:13:26.091727 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-765c6857c8-q7glx" podStartSLOduration=7.091701812 podStartE2EDuration="7.091701812s" podCreationTimestamp="2025-10-05 21:13:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 21:13:26.08022059 +0000 UTC m=+1129.984339300" watchObservedRunningTime="2025-10-05 21:13:26.091701812 +0000 UTC m=+1129.995820522" Oct 05 21:13:26 crc kubenswrapper[4754]: I1005 21:13:26.126303 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-69c986f6d7-fksbs" podStartSLOduration=6.126272432 podStartE2EDuration="6.126272432s" podCreationTimestamp="2025-10-05 21:13:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 21:13:26.120602823 +0000 UTC m=+1130.024721533" watchObservedRunningTime="2025-10-05 21:13:26.126272432 +0000 UTC m=+1130.030391142" Oct 05 21:13:26 crc kubenswrapper[4754]: I1005 21:13:26.160742 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-9dc4bfc9d-cfm98" event={"ID":"358e33ff-b4ce-46e3-b077-e1df373f2a75","Type":"ContainerStarted","Data":"ab9d7b029b6ff3cd8ecbbc1bb90eaa31e9bd0f1d62682556fc86a7a7fe26d49a"} Oct 05 21:13:26 crc kubenswrapper[4754]: I1005 21:13:26.194785 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-9dc4bfc9d-cfm98" podStartSLOduration=6.184076166 podStartE2EDuration="11.194756414s" podCreationTimestamp="2025-10-05 21:13:15 +0000 UTC" firstStartedPulling="2025-10-05 21:13:17.352912334 +0000 UTC m=+1121.257031044" lastFinishedPulling="2025-10-05 21:13:22.363592582 +0000 UTC m=+1126.267711292" observedRunningTime="2025-10-05 21:13:26.192205247 +0000 UTC m=+1130.096323967" watchObservedRunningTime="2025-10-05 21:13:26.194756414 +0000 UTC m=+1130.098875124" Oct 05 21:13:27 crc kubenswrapper[4754]: I1005 21:13:27.210703 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"5e757f47-326c-4174-9c28-c577a314f70b","Type":"ContainerStarted","Data":"5f27bec07a559ab193b04f03e1f03973140ba82e9195d2e4952585a2e52e40e7"} Oct 05 21:13:27 crc kubenswrapper[4754]: I1005 21:13:27.246450 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"0e4c84d5-ce7d-4b40-9b7c-980bc6b16c90","Type":"ContainerStarted","Data":"7d8e02519c00ab8dab67118f5aed2e8b655977781ccf0af4ff4939c450a39aef"} Oct 05 21:13:27 crc kubenswrapper[4754]: I1005 21:13:27.246781 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="0e4c84d5-ce7d-4b40-9b7c-980bc6b16c90" containerName="cinder-api-log" containerID="cri-o://f0c9444516b4ed2de97be88fcc993a6d95e861794de12525343dfdf8876fd1fb" gracePeriod=30 Oct 05 21:13:27 crc kubenswrapper[4754]: I1005 21:13:27.250845 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="0e4c84d5-ce7d-4b40-9b7c-980bc6b16c90" containerName="cinder-api" containerID="cri-o://7d8e02519c00ab8dab67118f5aed2e8b655977781ccf0af4ff4939c450a39aef" gracePeriod=30 Oct 05 21:13:27 crc kubenswrapper[4754]: I1005 21:13:27.251090 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Oct 05 21:13:27 crc kubenswrapper[4754]: I1005 21:13:27.332090 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=7.3320629010000005 podStartE2EDuration="7.332062901s" podCreationTimestamp="2025-10-05 21:13:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 21:13:27.292999243 +0000 UTC m=+1131.197117953" watchObservedRunningTime="2025-10-05 21:13:27.332062901 +0000 UTC m=+1131.236181611" Oct 05 21:13:27 crc kubenswrapper[4754]: I1005 21:13:27.567748 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-5fb78978fb-2mr8x" podUID="a0655389-4b2f-4973-b572-88df7ee28344" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.158:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 05 21:13:27 crc kubenswrapper[4754]: I1005 21:13:27.927844 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-xlw68" Oct 05 21:13:28 crc kubenswrapper[4754]: I1005 21:13:28.036841 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/94c73d9b-e224-4788-995d-be11dcee7b3c-scripts\") pod \"94c73d9b-e224-4788-995d-be11dcee7b3c\" (UID: \"94c73d9b-e224-4788-995d-be11dcee7b3c\") " Oct 05 21:13:28 crc kubenswrapper[4754]: I1005 21:13:28.037026 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/94c73d9b-e224-4788-995d-be11dcee7b3c-config-data\") pod \"94c73d9b-e224-4788-995d-be11dcee7b3c\" (UID: \"94c73d9b-e224-4788-995d-be11dcee7b3c\") " Oct 05 21:13:28 crc kubenswrapper[4754]: I1005 21:13:28.037123 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v775j\" (UniqueName: \"kubernetes.io/projected/94c73d9b-e224-4788-995d-be11dcee7b3c-kube-api-access-v775j\") pod \"94c73d9b-e224-4788-995d-be11dcee7b3c\" (UID: \"94c73d9b-e224-4788-995d-be11dcee7b3c\") " Oct 05 21:13:28 crc kubenswrapper[4754]: I1005 21:13:28.037167 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94c73d9b-e224-4788-995d-be11dcee7b3c-combined-ca-bundle\") pod \"94c73d9b-e224-4788-995d-be11dcee7b3c\" (UID: \"94c73d9b-e224-4788-995d-be11dcee7b3c\") " Oct 05 21:13:28 crc kubenswrapper[4754]: I1005 21:13:28.037209 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/94c73d9b-e224-4788-995d-be11dcee7b3c-logs\") pod \"94c73d9b-e224-4788-995d-be11dcee7b3c\" (UID: \"94c73d9b-e224-4788-995d-be11dcee7b3c\") " Oct 05 21:13:28 crc kubenswrapper[4754]: I1005 21:13:28.037876 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/94c73d9b-e224-4788-995d-be11dcee7b3c-logs" (OuterVolumeSpecName: "logs") pod "94c73d9b-e224-4788-995d-be11dcee7b3c" (UID: "94c73d9b-e224-4788-995d-be11dcee7b3c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 21:13:28 crc kubenswrapper[4754]: I1005 21:13:28.085878 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/94c73d9b-e224-4788-995d-be11dcee7b3c-scripts" (OuterVolumeSpecName: "scripts") pod "94c73d9b-e224-4788-995d-be11dcee7b3c" (UID: "94c73d9b-e224-4788-995d-be11dcee7b3c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:13:28 crc kubenswrapper[4754]: I1005 21:13:28.089399 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/94c73d9b-e224-4788-995d-be11dcee7b3c-kube-api-access-v775j" (OuterVolumeSpecName: "kube-api-access-v775j") pod "94c73d9b-e224-4788-995d-be11dcee7b3c" (UID: "94c73d9b-e224-4788-995d-be11dcee7b3c"). InnerVolumeSpecName "kube-api-access-v775j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:13:28 crc kubenswrapper[4754]: I1005 21:13:28.130731 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/94c73d9b-e224-4788-995d-be11dcee7b3c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "94c73d9b-e224-4788-995d-be11dcee7b3c" (UID: "94c73d9b-e224-4788-995d-be11dcee7b3c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:13:28 crc kubenswrapper[4754]: I1005 21:13:28.139551 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v775j\" (UniqueName: \"kubernetes.io/projected/94c73d9b-e224-4788-995d-be11dcee7b3c-kube-api-access-v775j\") on node \"crc\" DevicePath \"\"" Oct 05 21:13:28 crc kubenswrapper[4754]: I1005 21:13:28.139592 4754 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94c73d9b-e224-4788-995d-be11dcee7b3c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 21:13:28 crc kubenswrapper[4754]: I1005 21:13:28.139606 4754 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/94c73d9b-e224-4788-995d-be11dcee7b3c-logs\") on node \"crc\" DevicePath \"\"" Oct 05 21:13:28 crc kubenswrapper[4754]: I1005 21:13:28.139621 4754 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/94c73d9b-e224-4788-995d-be11dcee7b3c-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 21:13:28 crc kubenswrapper[4754]: I1005 21:13:28.217569 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/94c73d9b-e224-4788-995d-be11dcee7b3c-config-data" (OuterVolumeSpecName: "config-data") pod "94c73d9b-e224-4788-995d-be11dcee7b3c" (UID: "94c73d9b-e224-4788-995d-be11dcee7b3c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:13:28 crc kubenswrapper[4754]: I1005 21:13:28.242754 4754 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/94c73d9b-e224-4788-995d-be11dcee7b3c-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 21:13:28 crc kubenswrapper[4754]: I1005 21:13:28.324989 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-66d84555d4-xzsp4"] Oct 05 21:13:28 crc kubenswrapper[4754]: E1005 21:13:28.325559 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94c73d9b-e224-4788-995d-be11dcee7b3c" containerName="placement-db-sync" Oct 05 21:13:28 crc kubenswrapper[4754]: I1005 21:13:28.325574 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="94c73d9b-e224-4788-995d-be11dcee7b3c" containerName="placement-db-sync" Oct 05 21:13:28 crc kubenswrapper[4754]: I1005 21:13:28.325795 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="94c73d9b-e224-4788-995d-be11dcee7b3c" containerName="placement-db-sync" Oct 05 21:13:28 crc kubenswrapper[4754]: I1005 21:13:28.327064 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-66d84555d4-xzsp4" Oct 05 21:13:28 crc kubenswrapper[4754]: I1005 21:13:28.346296 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-66d84555d4-xzsp4"] Oct 05 21:13:28 crc kubenswrapper[4754]: I1005 21:13:28.347248 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Oct 05 21:13:28 crc kubenswrapper[4754]: I1005 21:13:28.347564 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Oct 05 21:13:28 crc kubenswrapper[4754]: I1005 21:13:28.356071 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"5e757f47-326c-4174-9c28-c577a314f70b","Type":"ContainerStarted","Data":"a9fcf577241addd321a4099ecf2e6d85990ed2142aebe79888bfa7f0a64aeb46"} Oct 05 21:13:28 crc kubenswrapper[4754]: I1005 21:13:28.365734 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-xlw68" event={"ID":"94c73d9b-e224-4788-995d-be11dcee7b3c","Type":"ContainerDied","Data":"33fe11d04fe5e87797e874bbaac02c954c408a1820e8306bb640a1ce1251b93e"} Oct 05 21:13:28 crc kubenswrapper[4754]: I1005 21:13:28.365773 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="33fe11d04fe5e87797e874bbaac02c954c408a1820e8306bb640a1ce1251b93e" Oct 05 21:13:28 crc kubenswrapper[4754]: I1005 21:13:28.365835 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-xlw68" Oct 05 21:13:28 crc kubenswrapper[4754]: I1005 21:13:28.446340 4754 generic.go:334] "Generic (PLEG): container finished" podID="0e4c84d5-ce7d-4b40-9b7c-980bc6b16c90" containerID="f0c9444516b4ed2de97be88fcc993a6d95e861794de12525343dfdf8876fd1fb" exitCode=143 Oct 05 21:13:28 crc kubenswrapper[4754]: I1005 21:13:28.446715 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"0e4c84d5-ce7d-4b40-9b7c-980bc6b16c90","Type":"ContainerDied","Data":"f0c9444516b4ed2de97be88fcc993a6d95e861794de12525343dfdf8876fd1fb"} Oct 05 21:13:28 crc kubenswrapper[4754]: I1005 21:13:28.447410 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ac596dcc-74ee-4144-bd90-3e0967d0478d-logs\") pod \"placement-66d84555d4-xzsp4\" (UID: \"ac596dcc-74ee-4144-bd90-3e0967d0478d\") " pod="openstack/placement-66d84555d4-xzsp4" Oct 05 21:13:28 crc kubenswrapper[4754]: I1005 21:13:28.447633 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac596dcc-74ee-4144-bd90-3e0967d0478d-config-data\") pod \"placement-66d84555d4-xzsp4\" (UID: \"ac596dcc-74ee-4144-bd90-3e0967d0478d\") " pod="openstack/placement-66d84555d4-xzsp4" Oct 05 21:13:28 crc kubenswrapper[4754]: I1005 21:13:28.448240 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ac596dcc-74ee-4144-bd90-3e0967d0478d-internal-tls-certs\") pod \"placement-66d84555d4-xzsp4\" (UID: \"ac596dcc-74ee-4144-bd90-3e0967d0478d\") " pod="openstack/placement-66d84555d4-xzsp4" Oct 05 21:13:28 crc kubenswrapper[4754]: I1005 21:13:28.448323 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac596dcc-74ee-4144-bd90-3e0967d0478d-combined-ca-bundle\") pod \"placement-66d84555d4-xzsp4\" (UID: \"ac596dcc-74ee-4144-bd90-3e0967d0478d\") " pod="openstack/placement-66d84555d4-xzsp4" Oct 05 21:13:28 crc kubenswrapper[4754]: I1005 21:13:28.448563 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ac596dcc-74ee-4144-bd90-3e0967d0478d-public-tls-certs\") pod \"placement-66d84555d4-xzsp4\" (UID: \"ac596dcc-74ee-4144-bd90-3e0967d0478d\") " pod="openstack/placement-66d84555d4-xzsp4" Oct 05 21:13:28 crc kubenswrapper[4754]: I1005 21:13:28.448676 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ac596dcc-74ee-4144-bd90-3e0967d0478d-scripts\") pod \"placement-66d84555d4-xzsp4\" (UID: \"ac596dcc-74ee-4144-bd90-3e0967d0478d\") " pod="openstack/placement-66d84555d4-xzsp4" Oct 05 21:13:28 crc kubenswrapper[4754]: I1005 21:13:28.448698 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qrpql\" (UniqueName: \"kubernetes.io/projected/ac596dcc-74ee-4144-bd90-3e0967d0478d-kube-api-access-qrpql\") pod \"placement-66d84555d4-xzsp4\" (UID: \"ac596dcc-74ee-4144-bd90-3e0967d0478d\") " pod="openstack/placement-66d84555d4-xzsp4" Oct 05 21:13:28 crc kubenswrapper[4754]: I1005 21:13:28.465860 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=7.226196616 podStartE2EDuration="8.465839715s" podCreationTimestamp="2025-10-05 21:13:20 +0000 UTC" firstStartedPulling="2025-10-05 21:13:24.045704305 +0000 UTC m=+1127.949823015" lastFinishedPulling="2025-10-05 21:13:25.285347404 +0000 UTC m=+1129.189466114" observedRunningTime="2025-10-05 21:13:28.432228611 +0000 UTC m=+1132.336347321" watchObservedRunningTime="2025-10-05 21:13:28.465839715 +0000 UTC m=+1132.369958425" Oct 05 21:13:28 crc kubenswrapper[4754]: I1005 21:13:28.550325 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ac596dcc-74ee-4144-bd90-3e0967d0478d-logs\") pod \"placement-66d84555d4-xzsp4\" (UID: \"ac596dcc-74ee-4144-bd90-3e0967d0478d\") " pod="openstack/placement-66d84555d4-xzsp4" Oct 05 21:13:28 crc kubenswrapper[4754]: I1005 21:13:28.550441 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac596dcc-74ee-4144-bd90-3e0967d0478d-config-data\") pod \"placement-66d84555d4-xzsp4\" (UID: \"ac596dcc-74ee-4144-bd90-3e0967d0478d\") " pod="openstack/placement-66d84555d4-xzsp4" Oct 05 21:13:28 crc kubenswrapper[4754]: I1005 21:13:28.550925 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ac596dcc-74ee-4144-bd90-3e0967d0478d-logs\") pod \"placement-66d84555d4-xzsp4\" (UID: \"ac596dcc-74ee-4144-bd90-3e0967d0478d\") " pod="openstack/placement-66d84555d4-xzsp4" Oct 05 21:13:28 crc kubenswrapper[4754]: I1005 21:13:28.551318 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ac596dcc-74ee-4144-bd90-3e0967d0478d-internal-tls-certs\") pod \"placement-66d84555d4-xzsp4\" (UID: \"ac596dcc-74ee-4144-bd90-3e0967d0478d\") " pod="openstack/placement-66d84555d4-xzsp4" Oct 05 21:13:28 crc kubenswrapper[4754]: I1005 21:13:28.551662 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac596dcc-74ee-4144-bd90-3e0967d0478d-combined-ca-bundle\") pod \"placement-66d84555d4-xzsp4\" (UID: \"ac596dcc-74ee-4144-bd90-3e0967d0478d\") " pod="openstack/placement-66d84555d4-xzsp4" Oct 05 21:13:28 crc kubenswrapper[4754]: I1005 21:13:28.551890 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ac596dcc-74ee-4144-bd90-3e0967d0478d-public-tls-certs\") pod \"placement-66d84555d4-xzsp4\" (UID: \"ac596dcc-74ee-4144-bd90-3e0967d0478d\") " pod="openstack/placement-66d84555d4-xzsp4" Oct 05 21:13:28 crc kubenswrapper[4754]: I1005 21:13:28.551960 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ac596dcc-74ee-4144-bd90-3e0967d0478d-scripts\") pod \"placement-66d84555d4-xzsp4\" (UID: \"ac596dcc-74ee-4144-bd90-3e0967d0478d\") " pod="openstack/placement-66d84555d4-xzsp4" Oct 05 21:13:28 crc kubenswrapper[4754]: I1005 21:13:28.551983 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qrpql\" (UniqueName: \"kubernetes.io/projected/ac596dcc-74ee-4144-bd90-3e0967d0478d-kube-api-access-qrpql\") pod \"placement-66d84555d4-xzsp4\" (UID: \"ac596dcc-74ee-4144-bd90-3e0967d0478d\") " pod="openstack/placement-66d84555d4-xzsp4" Oct 05 21:13:28 crc kubenswrapper[4754]: I1005 21:13:28.565025 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ac596dcc-74ee-4144-bd90-3e0967d0478d-scripts\") pod \"placement-66d84555d4-xzsp4\" (UID: \"ac596dcc-74ee-4144-bd90-3e0967d0478d\") " pod="openstack/placement-66d84555d4-xzsp4" Oct 05 21:13:28 crc kubenswrapper[4754]: I1005 21:13:28.565247 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac596dcc-74ee-4144-bd90-3e0967d0478d-combined-ca-bundle\") pod \"placement-66d84555d4-xzsp4\" (UID: \"ac596dcc-74ee-4144-bd90-3e0967d0478d\") " pod="openstack/placement-66d84555d4-xzsp4" Oct 05 21:13:28 crc kubenswrapper[4754]: I1005 21:13:28.566443 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac596dcc-74ee-4144-bd90-3e0967d0478d-config-data\") pod \"placement-66d84555d4-xzsp4\" (UID: \"ac596dcc-74ee-4144-bd90-3e0967d0478d\") " pod="openstack/placement-66d84555d4-xzsp4" Oct 05 21:13:28 crc kubenswrapper[4754]: I1005 21:13:28.574294 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ac596dcc-74ee-4144-bd90-3e0967d0478d-internal-tls-certs\") pod \"placement-66d84555d4-xzsp4\" (UID: \"ac596dcc-74ee-4144-bd90-3e0967d0478d\") " pod="openstack/placement-66d84555d4-xzsp4" Oct 05 21:13:28 crc kubenswrapper[4754]: I1005 21:13:28.575069 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qrpql\" (UniqueName: \"kubernetes.io/projected/ac596dcc-74ee-4144-bd90-3e0967d0478d-kube-api-access-qrpql\") pod \"placement-66d84555d4-xzsp4\" (UID: \"ac596dcc-74ee-4144-bd90-3e0967d0478d\") " pod="openstack/placement-66d84555d4-xzsp4" Oct 05 21:13:28 crc kubenswrapper[4754]: I1005 21:13:28.585453 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ac596dcc-74ee-4144-bd90-3e0967d0478d-public-tls-certs\") pod \"placement-66d84555d4-xzsp4\" (UID: \"ac596dcc-74ee-4144-bd90-3e0967d0478d\") " pod="openstack/placement-66d84555d4-xzsp4" Oct 05 21:13:28 crc kubenswrapper[4754]: I1005 21:13:28.691456 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-66d84555d4-xzsp4" Oct 05 21:13:29 crc kubenswrapper[4754]: I1005 21:13:29.339788 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-66d84555d4-xzsp4"] Oct 05 21:13:29 crc kubenswrapper[4754]: I1005 21:13:29.458771 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-66d84555d4-xzsp4" event={"ID":"ac596dcc-74ee-4144-bd90-3e0967d0478d","Type":"ContainerStarted","Data":"06a74cf491de8608c03adc53ae1cb118645bfa223df163948ff12e38fe2cc1ab"} Oct 05 21:13:30 crc kubenswrapper[4754]: I1005 21:13:30.479106 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-66d84555d4-xzsp4" event={"ID":"ac596dcc-74ee-4144-bd90-3e0967d0478d","Type":"ContainerStarted","Data":"4b5f2b2a1aea8f8390d433e7bb1f532a97f83670ee97be9de5225993c0b01deb"} Oct 05 21:13:30 crc kubenswrapper[4754]: I1005 21:13:30.480047 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-66d84555d4-xzsp4" event={"ID":"ac596dcc-74ee-4144-bd90-3e0967d0478d","Type":"ContainerStarted","Data":"ec5b13ace872a9d771b7d2554fbab8327f868749c6e31d7ddd4c48af963b9ebc"} Oct 05 21:13:30 crc kubenswrapper[4754]: I1005 21:13:30.481388 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-66d84555d4-xzsp4" Oct 05 21:13:30 crc kubenswrapper[4754]: I1005 21:13:30.481439 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-66d84555d4-xzsp4" Oct 05 21:13:30 crc kubenswrapper[4754]: I1005 21:13:30.526122 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-66d84555d4-xzsp4" podStartSLOduration=2.526105908 podStartE2EDuration="2.526105908s" podCreationTimestamp="2025-10-05 21:13:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 21:13:30.52429519 +0000 UTC m=+1134.428413900" watchObservedRunningTime="2025-10-05 21:13:30.526105908 +0000 UTC m=+1134.430224618" Oct 05 21:13:30 crc kubenswrapper[4754]: I1005 21:13:30.553950 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-5fb78978fb-2mr8x" podUID="a0655389-4b2f-4973-b572-88df7ee28344" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.158:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 05 21:13:30 crc kubenswrapper[4754]: I1005 21:13:30.609930 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-68c6b67864-9msm5" podUID="27cd6be5-f5eb-4158-bdd1-37f4df3d530d" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.148:8443/dashboard/auth/login/?next=/dashboard/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 05 21:13:30 crc kubenswrapper[4754]: I1005 21:13:30.610064 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-68c6b67864-9msm5" Oct 05 21:13:30 crc kubenswrapper[4754]: I1005 21:13:30.611241 4754 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="horizon" containerStatusID={"Type":"cri-o","ID":"7d14664c74062c1ff554d9b5e811703f4db1749cb0131849ed4d6056fb07bc02"} pod="openstack/horizon-68c6b67864-9msm5" containerMessage="Container horizon failed startup probe, will be restarted" Oct 05 21:13:30 crc kubenswrapper[4754]: I1005 21:13:30.611291 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-68c6b67864-9msm5" podUID="27cd6be5-f5eb-4158-bdd1-37f4df3d530d" containerName="horizon" containerID="cri-o://7d14664c74062c1ff554d9b5e811703f4db1749cb0131849ed4d6056fb07bc02" gracePeriod=30 Oct 05 21:13:30 crc kubenswrapper[4754]: I1005 21:13:30.728769 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-574454d6cb-vbnk6" podUID="39d1aac3-14a1-45bf-9a9d-ba4cdb0afcb1" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.149:8443/dashboard/auth/login/?next=/dashboard/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 05 21:13:30 crc kubenswrapper[4754]: I1005 21:13:30.728912 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-574454d6cb-vbnk6" Oct 05 21:13:30 crc kubenswrapper[4754]: I1005 21:13:30.730066 4754 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="horizon" containerStatusID={"Type":"cri-o","ID":"7ebc5fb1dedef86d2a80ae2a6e400e3f44aaa47d3b94955c70a007eb48092a6a"} pod="openstack/horizon-574454d6cb-vbnk6" containerMessage="Container horizon failed startup probe, will be restarted" Oct 05 21:13:30 crc kubenswrapper[4754]: I1005 21:13:30.730122 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-574454d6cb-vbnk6" podUID="39d1aac3-14a1-45bf-9a9d-ba4cdb0afcb1" containerName="horizon" containerID="cri-o://7ebc5fb1dedef86d2a80ae2a6e400e3f44aaa47d3b94955c70a007eb48092a6a" gracePeriod=30 Oct 05 21:13:30 crc kubenswrapper[4754]: I1005 21:13:30.871444 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Oct 05 21:13:30 crc kubenswrapper[4754]: I1005 21:13:30.975656 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-69c986f6d7-fksbs" Oct 05 21:13:31 crc kubenswrapper[4754]: I1005 21:13:31.070230 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-lsj2f"] Oct 05 21:13:31 crc kubenswrapper[4754]: I1005 21:13:31.070513 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-8b5c85b87-lsj2f" podUID="f8012368-f6e9-43a6-aaa2-b228f05b79e3" containerName="dnsmasq-dns" containerID="cri-o://662c7f7244687f194dbaef3262108de7752198aac4491f90932c3b7f614ab2a3" gracePeriod=10 Oct 05 21:13:31 crc kubenswrapper[4754]: I1005 21:13:31.521786 4754 generic.go:334] "Generic (PLEG): container finished" podID="f8012368-f6e9-43a6-aaa2-b228f05b79e3" containerID="662c7f7244687f194dbaef3262108de7752198aac4491f90932c3b7f614ab2a3" exitCode=0 Oct 05 21:13:31 crc kubenswrapper[4754]: I1005 21:13:31.523096 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-lsj2f" event={"ID":"f8012368-f6e9-43a6-aaa2-b228f05b79e3","Type":"ContainerDied","Data":"662c7f7244687f194dbaef3262108de7752198aac4491f90932c3b7f614ab2a3"} Oct 05 21:13:31 crc kubenswrapper[4754]: I1005 21:13:31.596883 4754 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-5fb78978fb-2mr8x" podUID="a0655389-4b2f-4973-b572-88df7ee28344" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.158:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 05 21:13:31 crc kubenswrapper[4754]: I1005 21:13:31.597252 4754 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-5fb78978fb-2mr8x" podUID="a0655389-4b2f-4973-b572-88df7ee28344" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.158:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 05 21:13:31 crc kubenswrapper[4754]: I1005 21:13:31.889779 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8b5c85b87-lsj2f" Oct 05 21:13:32 crc kubenswrapper[4754]: I1005 21:13:32.049215 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f8012368-f6e9-43a6-aaa2-b228f05b79e3-dns-swift-storage-0\") pod \"f8012368-f6e9-43a6-aaa2-b228f05b79e3\" (UID: \"f8012368-f6e9-43a6-aaa2-b228f05b79e3\") " Oct 05 21:13:32 crc kubenswrapper[4754]: I1005 21:13:32.049330 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f8012368-f6e9-43a6-aaa2-b228f05b79e3-ovsdbserver-sb\") pod \"f8012368-f6e9-43a6-aaa2-b228f05b79e3\" (UID: \"f8012368-f6e9-43a6-aaa2-b228f05b79e3\") " Oct 05 21:13:32 crc kubenswrapper[4754]: I1005 21:13:32.049386 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f8012368-f6e9-43a6-aaa2-b228f05b79e3-dns-svc\") pod \"f8012368-f6e9-43a6-aaa2-b228f05b79e3\" (UID: \"f8012368-f6e9-43a6-aaa2-b228f05b79e3\") " Oct 05 21:13:32 crc kubenswrapper[4754]: I1005 21:13:32.049437 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f8012368-f6e9-43a6-aaa2-b228f05b79e3-config\") pod \"f8012368-f6e9-43a6-aaa2-b228f05b79e3\" (UID: \"f8012368-f6e9-43a6-aaa2-b228f05b79e3\") " Oct 05 21:13:32 crc kubenswrapper[4754]: I1005 21:13:32.049510 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s5hkv\" (UniqueName: \"kubernetes.io/projected/f8012368-f6e9-43a6-aaa2-b228f05b79e3-kube-api-access-s5hkv\") pod \"f8012368-f6e9-43a6-aaa2-b228f05b79e3\" (UID: \"f8012368-f6e9-43a6-aaa2-b228f05b79e3\") " Oct 05 21:13:32 crc kubenswrapper[4754]: I1005 21:13:32.049533 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f8012368-f6e9-43a6-aaa2-b228f05b79e3-ovsdbserver-nb\") pod \"f8012368-f6e9-43a6-aaa2-b228f05b79e3\" (UID: \"f8012368-f6e9-43a6-aaa2-b228f05b79e3\") " Oct 05 21:13:32 crc kubenswrapper[4754]: I1005 21:13:32.071792 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f8012368-f6e9-43a6-aaa2-b228f05b79e3-kube-api-access-s5hkv" (OuterVolumeSpecName: "kube-api-access-s5hkv") pod "f8012368-f6e9-43a6-aaa2-b228f05b79e3" (UID: "f8012368-f6e9-43a6-aaa2-b228f05b79e3"). InnerVolumeSpecName "kube-api-access-s5hkv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:13:32 crc kubenswrapper[4754]: I1005 21:13:32.152180 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s5hkv\" (UniqueName: \"kubernetes.io/projected/f8012368-f6e9-43a6-aaa2-b228f05b79e3-kube-api-access-s5hkv\") on node \"crc\" DevicePath \"\"" Oct 05 21:13:32 crc kubenswrapper[4754]: I1005 21:13:32.184185 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f8012368-f6e9-43a6-aaa2-b228f05b79e3-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "f8012368-f6e9-43a6-aaa2-b228f05b79e3" (UID: "f8012368-f6e9-43a6-aaa2-b228f05b79e3"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 21:13:32 crc kubenswrapper[4754]: I1005 21:13:32.184387 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f8012368-f6e9-43a6-aaa2-b228f05b79e3-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "f8012368-f6e9-43a6-aaa2-b228f05b79e3" (UID: "f8012368-f6e9-43a6-aaa2-b228f05b79e3"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 21:13:32 crc kubenswrapper[4754]: I1005 21:13:32.242202 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f8012368-f6e9-43a6-aaa2-b228f05b79e3-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f8012368-f6e9-43a6-aaa2-b228f05b79e3" (UID: "f8012368-f6e9-43a6-aaa2-b228f05b79e3"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 21:13:32 crc kubenswrapper[4754]: I1005 21:13:32.248004 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f8012368-f6e9-43a6-aaa2-b228f05b79e3-config" (OuterVolumeSpecName: "config") pod "f8012368-f6e9-43a6-aaa2-b228f05b79e3" (UID: "f8012368-f6e9-43a6-aaa2-b228f05b79e3"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 21:13:32 crc kubenswrapper[4754]: I1005 21:13:32.253998 4754 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f8012368-f6e9-43a6-aaa2-b228f05b79e3-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 05 21:13:32 crc kubenswrapper[4754]: I1005 21:13:32.254031 4754 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f8012368-f6e9-43a6-aaa2-b228f05b79e3-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 05 21:13:32 crc kubenswrapper[4754]: I1005 21:13:32.254041 4754 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f8012368-f6e9-43a6-aaa2-b228f05b79e3-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 05 21:13:32 crc kubenswrapper[4754]: I1005 21:13:32.254055 4754 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f8012368-f6e9-43a6-aaa2-b228f05b79e3-config\") on node \"crc\" DevicePath \"\"" Oct 05 21:13:32 crc kubenswrapper[4754]: I1005 21:13:32.255171 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f8012368-f6e9-43a6-aaa2-b228f05b79e3-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "f8012368-f6e9-43a6-aaa2-b228f05b79e3" (UID: "f8012368-f6e9-43a6-aaa2-b228f05b79e3"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 21:13:32 crc kubenswrapper[4754]: I1005 21:13:32.355528 4754 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f8012368-f6e9-43a6-aaa2-b228f05b79e3-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 05 21:13:32 crc kubenswrapper[4754]: I1005 21:13:32.534645 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-lsj2f" event={"ID":"f8012368-f6e9-43a6-aaa2-b228f05b79e3","Type":"ContainerDied","Data":"0d4460efe4cb5abf2ebf32ff09520226a7c4892dcc794f577c86d187e041c4c1"} Oct 05 21:13:32 crc kubenswrapper[4754]: I1005 21:13:32.534702 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8b5c85b87-lsj2f" Oct 05 21:13:32 crc kubenswrapper[4754]: I1005 21:13:32.534763 4754 scope.go:117] "RemoveContainer" containerID="662c7f7244687f194dbaef3262108de7752198aac4491f90932c3b7f614ab2a3" Oct 05 21:13:32 crc kubenswrapper[4754]: I1005 21:13:32.577610 4754 scope.go:117] "RemoveContainer" containerID="540f320a309cdb6e6ec64a15c4d6c93e9e2b9eae2b0d9eef21cf099c1c26fd88" Oct 05 21:13:32 crc kubenswrapper[4754]: I1005 21:13:32.604451 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-lsj2f"] Oct 05 21:13:32 crc kubenswrapper[4754]: I1005 21:13:32.614325 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-lsj2f"] Oct 05 21:13:32 crc kubenswrapper[4754]: I1005 21:13:32.615744 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-5fb78978fb-2mr8x" podUID="a0655389-4b2f-4973-b572-88df7ee28344" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.158:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 05 21:13:32 crc kubenswrapper[4754]: I1005 21:13:32.853983 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f8012368-f6e9-43a6-aaa2-b228f05b79e3" path="/var/lib/kubelet/pods/f8012368-f6e9-43a6-aaa2-b228f05b79e3/volumes" Oct 05 21:13:33 crc kubenswrapper[4754]: I1005 21:13:33.559456 4754 generic.go:334] "Generic (PLEG): container finished" podID="5594b283-6565-4ab3-89bd-9769992f6a20" containerID="35e7ec9f9fdd5504e8cead852c32e929a754fb8f8f40cf4c89ee0c93643b374b" exitCode=0 Oct 05 21:13:33 crc kubenswrapper[4754]: I1005 21:13:33.559643 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-drwtj" event={"ID":"5594b283-6565-4ab3-89bd-9769992f6a20","Type":"ContainerDied","Data":"35e7ec9f9fdd5504e8cead852c32e929a754fb8f8f40cf4c89ee0c93643b374b"} Oct 05 21:13:34 crc kubenswrapper[4754]: I1005 21:13:34.219718 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-765c6857c8-q7glx" podUID="bd23550d-8815-4211-b8be-10ee8ab95c1f" containerName="barbican-api-log" probeResult="failure" output="Get \"https://10.217.0.159:9311/healthcheck\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 05 21:13:35 crc kubenswrapper[4754]: I1005 21:13:35.250730 4754 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-765c6857c8-q7glx" podUID="bd23550d-8815-4211-b8be-10ee8ab95c1f" containerName="barbican-api-log" probeResult="failure" output="Get \"https://10.217.0.159:9311/healthcheck\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 05 21:13:35 crc kubenswrapper[4754]: I1005 21:13:35.251613 4754 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-765c6857c8-q7glx" podUID="bd23550d-8815-4211-b8be-10ee8ab95c1f" containerName="barbican-api" probeResult="failure" output="Get \"https://10.217.0.159:9311/healthcheck\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 05 21:13:35 crc kubenswrapper[4754]: I1005 21:13:35.252176 4754 patch_prober.go:28] interesting pod/machine-config-daemon-b2h9k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 21:13:35 crc kubenswrapper[4754]: I1005 21:13:35.252205 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 21:13:35 crc kubenswrapper[4754]: I1005 21:13:35.287226 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-drwtj" Oct 05 21:13:35 crc kubenswrapper[4754]: I1005 21:13:35.413591 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5594b283-6565-4ab3-89bd-9769992f6a20-combined-ca-bundle\") pod \"5594b283-6565-4ab3-89bd-9769992f6a20\" (UID: \"5594b283-6565-4ab3-89bd-9769992f6a20\") " Oct 05 21:13:35 crc kubenswrapper[4754]: I1005 21:13:35.413789 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2rknr\" (UniqueName: \"kubernetes.io/projected/5594b283-6565-4ab3-89bd-9769992f6a20-kube-api-access-2rknr\") pod \"5594b283-6565-4ab3-89bd-9769992f6a20\" (UID: \"5594b283-6565-4ab3-89bd-9769992f6a20\") " Oct 05 21:13:35 crc kubenswrapper[4754]: I1005 21:13:35.413858 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/5594b283-6565-4ab3-89bd-9769992f6a20-config\") pod \"5594b283-6565-4ab3-89bd-9769992f6a20\" (UID: \"5594b283-6565-4ab3-89bd-9769992f6a20\") " Oct 05 21:13:35 crc kubenswrapper[4754]: I1005 21:13:35.442745 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5594b283-6565-4ab3-89bd-9769992f6a20-kube-api-access-2rknr" (OuterVolumeSpecName: "kube-api-access-2rknr") pod "5594b283-6565-4ab3-89bd-9769992f6a20" (UID: "5594b283-6565-4ab3-89bd-9769992f6a20"). InnerVolumeSpecName "kube-api-access-2rknr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:13:35 crc kubenswrapper[4754]: I1005 21:13:35.455872 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5594b283-6565-4ab3-89bd-9769992f6a20-config" (OuterVolumeSpecName: "config") pod "5594b283-6565-4ab3-89bd-9769992f6a20" (UID: "5594b283-6565-4ab3-89bd-9769992f6a20"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:13:35 crc kubenswrapper[4754]: I1005 21:13:35.464533 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5594b283-6565-4ab3-89bd-9769992f6a20-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5594b283-6565-4ab3-89bd-9769992f6a20" (UID: "5594b283-6565-4ab3-89bd-9769992f6a20"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:13:35 crc kubenswrapper[4754]: I1005 21:13:35.517514 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2rknr\" (UniqueName: \"kubernetes.io/projected/5594b283-6565-4ab3-89bd-9769992f6a20-kube-api-access-2rknr\") on node \"crc\" DevicePath \"\"" Oct 05 21:13:35 crc kubenswrapper[4754]: I1005 21:13:35.517548 4754 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/5594b283-6565-4ab3-89bd-9769992f6a20-config\") on node \"crc\" DevicePath \"\"" Oct 05 21:13:35 crc kubenswrapper[4754]: I1005 21:13:35.517560 4754 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5594b283-6565-4ab3-89bd-9769992f6a20-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 21:13:35 crc kubenswrapper[4754]: I1005 21:13:35.579168 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-drwtj" event={"ID":"5594b283-6565-4ab3-89bd-9769992f6a20","Type":"ContainerDied","Data":"a31e7bf693a7dec198c99da63f8bd527f44e35c07981f8741533f043e5bae8c5"} Oct 05 21:13:35 crc kubenswrapper[4754]: I1005 21:13:35.579214 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a31e7bf693a7dec198c99da63f8bd527f44e35c07981f8741533f043e5bae8c5" Oct 05 21:13:35 crc kubenswrapper[4754]: I1005 21:13:35.579287 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-drwtj" Oct 05 21:13:35 crc kubenswrapper[4754]: I1005 21:13:35.595232 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-5fb78978fb-2mr8x" podUID="a0655389-4b2f-4973-b572-88df7ee28344" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.158:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 05 21:13:35 crc kubenswrapper[4754]: I1005 21:13:35.827750 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-5fb78978fb-2mr8x" Oct 05 21:13:35 crc kubenswrapper[4754]: I1005 21:13:35.866002 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-dvhns"] Oct 05 21:13:35 crc kubenswrapper[4754]: E1005 21:13:35.866411 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8012368-f6e9-43a6-aaa2-b228f05b79e3" containerName="dnsmasq-dns" Oct 05 21:13:35 crc kubenswrapper[4754]: I1005 21:13:35.866423 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8012368-f6e9-43a6-aaa2-b228f05b79e3" containerName="dnsmasq-dns" Oct 05 21:13:35 crc kubenswrapper[4754]: E1005 21:13:35.866449 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8012368-f6e9-43a6-aaa2-b228f05b79e3" containerName="init" Oct 05 21:13:35 crc kubenswrapper[4754]: I1005 21:13:35.866454 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8012368-f6e9-43a6-aaa2-b228f05b79e3" containerName="init" Oct 05 21:13:35 crc kubenswrapper[4754]: E1005 21:13:35.866473 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5594b283-6565-4ab3-89bd-9769992f6a20" containerName="neutron-db-sync" Oct 05 21:13:35 crc kubenswrapper[4754]: I1005 21:13:35.866479 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="5594b283-6565-4ab3-89bd-9769992f6a20" containerName="neutron-db-sync" Oct 05 21:13:35 crc kubenswrapper[4754]: I1005 21:13:35.866661 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="5594b283-6565-4ab3-89bd-9769992f6a20" containerName="neutron-db-sync" Oct 05 21:13:35 crc kubenswrapper[4754]: I1005 21:13:35.866688 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8012368-f6e9-43a6-aaa2-b228f05b79e3" containerName="dnsmasq-dns" Oct 05 21:13:35 crc kubenswrapper[4754]: I1005 21:13:35.867616 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5784cf869f-dvhns" Oct 05 21:13:35 crc kubenswrapper[4754]: I1005 21:13:35.887829 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-dvhns"] Oct 05 21:13:35 crc kubenswrapper[4754]: I1005 21:13:35.907111 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/cinder-scheduler-0" podUID="5e757f47-326c-4174-9c28-c577a314f70b" containerName="cinder-scheduler" probeResult="failure" output="Get \"http://10.217.0.160:8080/\": dial tcp 10.217.0.160:8080: connect: connection refused" Oct 05 21:13:35 crc kubenswrapper[4754]: I1005 21:13:35.925709 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g84v8\" (UniqueName: \"kubernetes.io/projected/8a5ef4bc-52a1-4ab7-bc00-5bae67d18232-kube-api-access-g84v8\") pod \"dnsmasq-dns-5784cf869f-dvhns\" (UID: \"8a5ef4bc-52a1-4ab7-bc00-5bae67d18232\") " pod="openstack/dnsmasq-dns-5784cf869f-dvhns" Oct 05 21:13:35 crc kubenswrapper[4754]: I1005 21:13:35.925834 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8a5ef4bc-52a1-4ab7-bc00-5bae67d18232-ovsdbserver-nb\") pod \"dnsmasq-dns-5784cf869f-dvhns\" (UID: \"8a5ef4bc-52a1-4ab7-bc00-5bae67d18232\") " pod="openstack/dnsmasq-dns-5784cf869f-dvhns" Oct 05 21:13:35 crc kubenswrapper[4754]: I1005 21:13:35.925895 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8a5ef4bc-52a1-4ab7-bc00-5bae67d18232-config\") pod \"dnsmasq-dns-5784cf869f-dvhns\" (UID: \"8a5ef4bc-52a1-4ab7-bc00-5bae67d18232\") " pod="openstack/dnsmasq-dns-5784cf869f-dvhns" Oct 05 21:13:35 crc kubenswrapper[4754]: I1005 21:13:35.925941 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8a5ef4bc-52a1-4ab7-bc00-5bae67d18232-dns-swift-storage-0\") pod \"dnsmasq-dns-5784cf869f-dvhns\" (UID: \"8a5ef4bc-52a1-4ab7-bc00-5bae67d18232\") " pod="openstack/dnsmasq-dns-5784cf869f-dvhns" Oct 05 21:13:35 crc kubenswrapper[4754]: I1005 21:13:35.926004 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8a5ef4bc-52a1-4ab7-bc00-5bae67d18232-dns-svc\") pod \"dnsmasq-dns-5784cf869f-dvhns\" (UID: \"8a5ef4bc-52a1-4ab7-bc00-5bae67d18232\") " pod="openstack/dnsmasq-dns-5784cf869f-dvhns" Oct 05 21:13:35 crc kubenswrapper[4754]: I1005 21:13:35.926050 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8a5ef4bc-52a1-4ab7-bc00-5bae67d18232-ovsdbserver-sb\") pod \"dnsmasq-dns-5784cf869f-dvhns\" (UID: \"8a5ef4bc-52a1-4ab7-bc00-5bae67d18232\") " pod="openstack/dnsmasq-dns-5784cf869f-dvhns" Oct 05 21:13:35 crc kubenswrapper[4754]: I1005 21:13:35.950903 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-77ccdf5dfb-zv8nn"] Oct 05 21:13:35 crc kubenswrapper[4754]: I1005 21:13:35.952651 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-77ccdf5dfb-zv8nn" Oct 05 21:13:35 crc kubenswrapper[4754]: I1005 21:13:35.969208 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Oct 05 21:13:35 crc kubenswrapper[4754]: I1005 21:13:35.969380 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Oct 05 21:13:35 crc kubenswrapper[4754]: I1005 21:13:35.969433 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Oct 05 21:13:35 crc kubenswrapper[4754]: I1005 21:13:35.970802 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-zvvtc" Oct 05 21:13:35 crc kubenswrapper[4754]: I1005 21:13:35.979045 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-77ccdf5dfb-zv8nn"] Oct 05 21:13:36 crc kubenswrapper[4754]: I1005 21:13:36.037170 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jhxrj\" (UniqueName: \"kubernetes.io/projected/a42df2f2-2289-4a9e-9ab4-156d3c8d2145-kube-api-access-jhxrj\") pod \"neutron-77ccdf5dfb-zv8nn\" (UID: \"a42df2f2-2289-4a9e-9ab4-156d3c8d2145\") " pod="openstack/neutron-77ccdf5dfb-zv8nn" Oct 05 21:13:36 crc kubenswrapper[4754]: I1005 21:13:36.037297 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/a42df2f2-2289-4a9e-9ab4-156d3c8d2145-config\") pod \"neutron-77ccdf5dfb-zv8nn\" (UID: \"a42df2f2-2289-4a9e-9ab4-156d3c8d2145\") " pod="openstack/neutron-77ccdf5dfb-zv8nn" Oct 05 21:13:36 crc kubenswrapper[4754]: I1005 21:13:36.037329 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/a42df2f2-2289-4a9e-9ab4-156d3c8d2145-httpd-config\") pod \"neutron-77ccdf5dfb-zv8nn\" (UID: \"a42df2f2-2289-4a9e-9ab4-156d3c8d2145\") " pod="openstack/neutron-77ccdf5dfb-zv8nn" Oct 05 21:13:36 crc kubenswrapper[4754]: I1005 21:13:36.037377 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a42df2f2-2289-4a9e-9ab4-156d3c8d2145-combined-ca-bundle\") pod \"neutron-77ccdf5dfb-zv8nn\" (UID: \"a42df2f2-2289-4a9e-9ab4-156d3c8d2145\") " pod="openstack/neutron-77ccdf5dfb-zv8nn" Oct 05 21:13:36 crc kubenswrapper[4754]: I1005 21:13:36.037440 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g84v8\" (UniqueName: \"kubernetes.io/projected/8a5ef4bc-52a1-4ab7-bc00-5bae67d18232-kube-api-access-g84v8\") pod \"dnsmasq-dns-5784cf869f-dvhns\" (UID: \"8a5ef4bc-52a1-4ab7-bc00-5bae67d18232\") " pod="openstack/dnsmasq-dns-5784cf869f-dvhns" Oct 05 21:13:36 crc kubenswrapper[4754]: I1005 21:13:36.037489 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8a5ef4bc-52a1-4ab7-bc00-5bae67d18232-ovsdbserver-nb\") pod \"dnsmasq-dns-5784cf869f-dvhns\" (UID: \"8a5ef4bc-52a1-4ab7-bc00-5bae67d18232\") " pod="openstack/dnsmasq-dns-5784cf869f-dvhns" Oct 05 21:13:36 crc kubenswrapper[4754]: I1005 21:13:36.037540 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/a42df2f2-2289-4a9e-9ab4-156d3c8d2145-ovndb-tls-certs\") pod \"neutron-77ccdf5dfb-zv8nn\" (UID: \"a42df2f2-2289-4a9e-9ab4-156d3c8d2145\") " pod="openstack/neutron-77ccdf5dfb-zv8nn" Oct 05 21:13:36 crc kubenswrapper[4754]: I1005 21:13:36.037591 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8a5ef4bc-52a1-4ab7-bc00-5bae67d18232-config\") pod \"dnsmasq-dns-5784cf869f-dvhns\" (UID: \"8a5ef4bc-52a1-4ab7-bc00-5bae67d18232\") " pod="openstack/dnsmasq-dns-5784cf869f-dvhns" Oct 05 21:13:36 crc kubenswrapper[4754]: I1005 21:13:36.037648 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8a5ef4bc-52a1-4ab7-bc00-5bae67d18232-dns-swift-storage-0\") pod \"dnsmasq-dns-5784cf869f-dvhns\" (UID: \"8a5ef4bc-52a1-4ab7-bc00-5bae67d18232\") " pod="openstack/dnsmasq-dns-5784cf869f-dvhns" Oct 05 21:13:36 crc kubenswrapper[4754]: I1005 21:13:36.037719 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8a5ef4bc-52a1-4ab7-bc00-5bae67d18232-dns-svc\") pod \"dnsmasq-dns-5784cf869f-dvhns\" (UID: \"8a5ef4bc-52a1-4ab7-bc00-5bae67d18232\") " pod="openstack/dnsmasq-dns-5784cf869f-dvhns" Oct 05 21:13:36 crc kubenswrapper[4754]: I1005 21:13:36.037753 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8a5ef4bc-52a1-4ab7-bc00-5bae67d18232-ovsdbserver-sb\") pod \"dnsmasq-dns-5784cf869f-dvhns\" (UID: \"8a5ef4bc-52a1-4ab7-bc00-5bae67d18232\") " pod="openstack/dnsmasq-dns-5784cf869f-dvhns" Oct 05 21:13:36 crc kubenswrapper[4754]: I1005 21:13:36.039110 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8a5ef4bc-52a1-4ab7-bc00-5bae67d18232-ovsdbserver-sb\") pod \"dnsmasq-dns-5784cf869f-dvhns\" (UID: \"8a5ef4bc-52a1-4ab7-bc00-5bae67d18232\") " pod="openstack/dnsmasq-dns-5784cf869f-dvhns" Oct 05 21:13:36 crc kubenswrapper[4754]: I1005 21:13:36.044635 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8a5ef4bc-52a1-4ab7-bc00-5bae67d18232-ovsdbserver-nb\") pod \"dnsmasq-dns-5784cf869f-dvhns\" (UID: \"8a5ef4bc-52a1-4ab7-bc00-5bae67d18232\") " pod="openstack/dnsmasq-dns-5784cf869f-dvhns" Oct 05 21:13:36 crc kubenswrapper[4754]: I1005 21:13:36.045741 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8a5ef4bc-52a1-4ab7-bc00-5bae67d18232-config\") pod \"dnsmasq-dns-5784cf869f-dvhns\" (UID: \"8a5ef4bc-52a1-4ab7-bc00-5bae67d18232\") " pod="openstack/dnsmasq-dns-5784cf869f-dvhns" Oct 05 21:13:36 crc kubenswrapper[4754]: I1005 21:13:36.046479 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8a5ef4bc-52a1-4ab7-bc00-5bae67d18232-dns-swift-storage-0\") pod \"dnsmasq-dns-5784cf869f-dvhns\" (UID: \"8a5ef4bc-52a1-4ab7-bc00-5bae67d18232\") " pod="openstack/dnsmasq-dns-5784cf869f-dvhns" Oct 05 21:13:36 crc kubenswrapper[4754]: I1005 21:13:36.047215 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8a5ef4bc-52a1-4ab7-bc00-5bae67d18232-dns-svc\") pod \"dnsmasq-dns-5784cf869f-dvhns\" (UID: \"8a5ef4bc-52a1-4ab7-bc00-5bae67d18232\") " pod="openstack/dnsmasq-dns-5784cf869f-dvhns" Oct 05 21:13:36 crc kubenswrapper[4754]: I1005 21:13:36.112525 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g84v8\" (UniqueName: \"kubernetes.io/projected/8a5ef4bc-52a1-4ab7-bc00-5bae67d18232-kube-api-access-g84v8\") pod \"dnsmasq-dns-5784cf869f-dvhns\" (UID: \"8a5ef4bc-52a1-4ab7-bc00-5bae67d18232\") " pod="openstack/dnsmasq-dns-5784cf869f-dvhns" Oct 05 21:13:36 crc kubenswrapper[4754]: I1005 21:13:36.139454 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jhxrj\" (UniqueName: \"kubernetes.io/projected/a42df2f2-2289-4a9e-9ab4-156d3c8d2145-kube-api-access-jhxrj\") pod \"neutron-77ccdf5dfb-zv8nn\" (UID: \"a42df2f2-2289-4a9e-9ab4-156d3c8d2145\") " pod="openstack/neutron-77ccdf5dfb-zv8nn" Oct 05 21:13:36 crc kubenswrapper[4754]: I1005 21:13:36.139583 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/a42df2f2-2289-4a9e-9ab4-156d3c8d2145-config\") pod \"neutron-77ccdf5dfb-zv8nn\" (UID: \"a42df2f2-2289-4a9e-9ab4-156d3c8d2145\") " pod="openstack/neutron-77ccdf5dfb-zv8nn" Oct 05 21:13:36 crc kubenswrapper[4754]: I1005 21:13:36.139607 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/a42df2f2-2289-4a9e-9ab4-156d3c8d2145-httpd-config\") pod \"neutron-77ccdf5dfb-zv8nn\" (UID: \"a42df2f2-2289-4a9e-9ab4-156d3c8d2145\") " pod="openstack/neutron-77ccdf5dfb-zv8nn" Oct 05 21:13:36 crc kubenswrapper[4754]: I1005 21:13:36.139651 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a42df2f2-2289-4a9e-9ab4-156d3c8d2145-combined-ca-bundle\") pod \"neutron-77ccdf5dfb-zv8nn\" (UID: \"a42df2f2-2289-4a9e-9ab4-156d3c8d2145\") " pod="openstack/neutron-77ccdf5dfb-zv8nn" Oct 05 21:13:36 crc kubenswrapper[4754]: I1005 21:13:36.139766 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/a42df2f2-2289-4a9e-9ab4-156d3c8d2145-ovndb-tls-certs\") pod \"neutron-77ccdf5dfb-zv8nn\" (UID: \"a42df2f2-2289-4a9e-9ab4-156d3c8d2145\") " pod="openstack/neutron-77ccdf5dfb-zv8nn" Oct 05 21:13:36 crc kubenswrapper[4754]: I1005 21:13:36.148466 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/a42df2f2-2289-4a9e-9ab4-156d3c8d2145-config\") pod \"neutron-77ccdf5dfb-zv8nn\" (UID: \"a42df2f2-2289-4a9e-9ab4-156d3c8d2145\") " pod="openstack/neutron-77ccdf5dfb-zv8nn" Oct 05 21:13:36 crc kubenswrapper[4754]: I1005 21:13:36.148844 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/a42df2f2-2289-4a9e-9ab4-156d3c8d2145-httpd-config\") pod \"neutron-77ccdf5dfb-zv8nn\" (UID: \"a42df2f2-2289-4a9e-9ab4-156d3c8d2145\") " pod="openstack/neutron-77ccdf5dfb-zv8nn" Oct 05 21:13:36 crc kubenswrapper[4754]: I1005 21:13:36.150548 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/a42df2f2-2289-4a9e-9ab4-156d3c8d2145-ovndb-tls-certs\") pod \"neutron-77ccdf5dfb-zv8nn\" (UID: \"a42df2f2-2289-4a9e-9ab4-156d3c8d2145\") " pod="openstack/neutron-77ccdf5dfb-zv8nn" Oct 05 21:13:36 crc kubenswrapper[4754]: I1005 21:13:36.155284 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a42df2f2-2289-4a9e-9ab4-156d3c8d2145-combined-ca-bundle\") pod \"neutron-77ccdf5dfb-zv8nn\" (UID: \"a42df2f2-2289-4a9e-9ab4-156d3c8d2145\") " pod="openstack/neutron-77ccdf5dfb-zv8nn" Oct 05 21:13:36 crc kubenswrapper[4754]: I1005 21:13:36.167716 4754 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cinder-api-0" podUID="0e4c84d5-ce7d-4b40-9b7c-980bc6b16c90" containerName="cinder-api" probeResult="failure" output="Get \"http://10.217.0.162:8776/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 05 21:13:36 crc kubenswrapper[4754]: I1005 21:13:36.216201 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5784cf869f-dvhns" Oct 05 21:13:36 crc kubenswrapper[4754]: I1005 21:13:36.217779 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jhxrj\" (UniqueName: \"kubernetes.io/projected/a42df2f2-2289-4a9e-9ab4-156d3c8d2145-kube-api-access-jhxrj\") pod \"neutron-77ccdf5dfb-zv8nn\" (UID: \"a42df2f2-2289-4a9e-9ab4-156d3c8d2145\") " pod="openstack/neutron-77ccdf5dfb-zv8nn" Oct 05 21:13:36 crc kubenswrapper[4754]: I1005 21:13:36.305983 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-77ccdf5dfb-zv8nn" Oct 05 21:13:36 crc kubenswrapper[4754]: I1005 21:13:36.638995 4754 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-5fb78978fb-2mr8x" podUID="a0655389-4b2f-4973-b572-88df7ee28344" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.158:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 05 21:13:36 crc kubenswrapper[4754]: I1005 21:13:36.674162 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-5fb78978fb-2mr8x" Oct 05 21:13:37 crc kubenswrapper[4754]: I1005 21:13:37.040737 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-dvhns"] Oct 05 21:13:37 crc kubenswrapper[4754]: I1005 21:13:37.209235 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-765c6857c8-q7glx" podUID="bd23550d-8815-4211-b8be-10ee8ab95c1f" containerName="barbican-api" probeResult="failure" output="Get \"https://10.217.0.159:9311/healthcheck\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 05 21:13:37 crc kubenswrapper[4754]: I1005 21:13:37.444563 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-77ccdf5dfb-zv8nn"] Oct 05 21:13:37 crc kubenswrapper[4754]: I1005 21:13:37.641644 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-77ccdf5dfb-zv8nn" event={"ID":"a42df2f2-2289-4a9e-9ab4-156d3c8d2145","Type":"ContainerStarted","Data":"dbaeab7aae3fc7fc5e91dc81679ec8b0c29f7333c686562bfb7b18703ef34eef"} Oct 05 21:13:37 crc kubenswrapper[4754]: I1005 21:13:37.659847 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-dvhns" event={"ID":"8a5ef4bc-52a1-4ab7-bc00-5bae67d18232","Type":"ContainerStarted","Data":"40ca378de3b0f6e31c8a7ea645e6d5239dd825b948855c118d4eb90cb8278546"} Oct 05 21:13:38 crc kubenswrapper[4754]: I1005 21:13:38.685003 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-77ccdf5dfb-zv8nn" event={"ID":"a42df2f2-2289-4a9e-9ab4-156d3c8d2145","Type":"ContainerStarted","Data":"daa7d97295865858fb6daa372fbbc8824aebf09867c36045c8d5ba563ee53ed9"} Oct 05 21:13:38 crc kubenswrapper[4754]: I1005 21:13:38.685293 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-77ccdf5dfb-zv8nn" event={"ID":"a42df2f2-2289-4a9e-9ab4-156d3c8d2145","Type":"ContainerStarted","Data":"28d4be6a6eb29d90be12ae24e0bb51cb8fac12201eacdd04cc51c1bec5626a1d"} Oct 05 21:13:38 crc kubenswrapper[4754]: I1005 21:13:38.685744 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-77ccdf5dfb-zv8nn" Oct 05 21:13:38 crc kubenswrapper[4754]: I1005 21:13:38.688753 4754 generic.go:334] "Generic (PLEG): container finished" podID="8a5ef4bc-52a1-4ab7-bc00-5bae67d18232" containerID="d5a18ce4db7c6a5664879e2181b9630036565632163d929a740c53352dc23536" exitCode=0 Oct 05 21:13:38 crc kubenswrapper[4754]: I1005 21:13:38.688788 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-dvhns" event={"ID":"8a5ef4bc-52a1-4ab7-bc00-5bae67d18232","Type":"ContainerDied","Data":"d5a18ce4db7c6a5664879e2181b9630036565632163d929a740c53352dc23536"} Oct 05 21:13:38 crc kubenswrapper[4754]: I1005 21:13:38.863017 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-77ccdf5dfb-zv8nn" podStartSLOduration=3.862991391 podStartE2EDuration="3.862991391s" podCreationTimestamp="2025-10-05 21:13:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 21:13:38.718458458 +0000 UTC m=+1142.622577168" watchObservedRunningTime="2025-10-05 21:13:38.862991391 +0000 UTC m=+1142.767110101" Oct 05 21:13:39 crc kubenswrapper[4754]: I1005 21:13:39.224684 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-765c6857c8-q7glx" podUID="bd23550d-8815-4211-b8be-10ee8ab95c1f" containerName="barbican-api-log" probeResult="failure" output="Get \"https://10.217.0.159:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 05 21:13:39 crc kubenswrapper[4754]: I1005 21:13:39.706950 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-dvhns" event={"ID":"8a5ef4bc-52a1-4ab7-bc00-5bae67d18232","Type":"ContainerStarted","Data":"8232f51f3147300f4ec3dd8d79475ed0e37e904ce8b0d2ba0df89f73c1725c50"} Oct 05 21:13:39 crc kubenswrapper[4754]: I1005 21:13:39.742127 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5784cf869f-dvhns" podStartSLOduration=4.742108674 podStartE2EDuration="4.742108674s" podCreationTimestamp="2025-10-05 21:13:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 21:13:39.735750116 +0000 UTC m=+1143.639868836" watchObservedRunningTime="2025-10-05 21:13:39.742108674 +0000 UTC m=+1143.646227384" Oct 05 21:13:40 crc kubenswrapper[4754]: I1005 21:13:40.259677 4754 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-765c6857c8-q7glx" podUID="bd23550d-8815-4211-b8be-10ee8ab95c1f" containerName="barbican-api-log" probeResult="failure" output="Get \"https://10.217.0.159:9311/healthcheck\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 05 21:13:40 crc kubenswrapper[4754]: I1005 21:13:40.260053 4754 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-765c6857c8-q7glx" podUID="bd23550d-8815-4211-b8be-10ee8ab95c1f" containerName="barbican-api" probeResult="failure" output="Get \"https://10.217.0.159:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 05 21:13:40 crc kubenswrapper[4754]: I1005 21:13:40.268476 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-765c6857c8-q7glx" Oct 05 21:13:40 crc kubenswrapper[4754]: I1005 21:13:40.279836 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-765c6857c8-q7glx" Oct 05 21:13:40 crc kubenswrapper[4754]: I1005 21:13:40.349804 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-646d45cbfc-6sjtl"] Oct 05 21:13:40 crc kubenswrapper[4754]: I1005 21:13:40.419784 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-646d45cbfc-6sjtl"] Oct 05 21:13:40 crc kubenswrapper[4754]: I1005 21:13:40.419936 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-646d45cbfc-6sjtl" Oct 05 21:13:40 crc kubenswrapper[4754]: I1005 21:13:40.421129 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-5fb78978fb-2mr8x"] Oct 05 21:13:40 crc kubenswrapper[4754]: I1005 21:13:40.421382 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-5fb78978fb-2mr8x" podUID="a0655389-4b2f-4973-b572-88df7ee28344" containerName="barbican-api-log" containerID="cri-o://7e2f47e1eba20b443bc6c1598258c83ac78c624aa918294f7f18a7ff0e5b1770" gracePeriod=30 Oct 05 21:13:40 crc kubenswrapper[4754]: I1005 21:13:40.421514 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-5fb78978fb-2mr8x" podUID="a0655389-4b2f-4973-b572-88df7ee28344" containerName="barbican-api" containerID="cri-o://e09d95f4d0fe4bccba2be305d0ec39651d39f0c07e8fbb497b7f9bad6ff21f9b" gracePeriod=30 Oct 05 21:13:40 crc kubenswrapper[4754]: I1005 21:13:40.434658 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Oct 05 21:13:40 crc kubenswrapper[4754]: I1005 21:13:40.436875 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Oct 05 21:13:40 crc kubenswrapper[4754]: I1005 21:13:40.438816 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Oct 05 21:13:40 crc kubenswrapper[4754]: I1005 21:13:40.482711 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e4c2921a-fff3-4c97-98df-206a94054d24-run-httpd\") pod \"swift-proxy-646d45cbfc-6sjtl\" (UID: \"e4c2921a-fff3-4c97-98df-206a94054d24\") " pod="openstack/swift-proxy-646d45cbfc-6sjtl" Oct 05 21:13:40 crc kubenswrapper[4754]: I1005 21:13:40.482756 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dzdq9\" (UniqueName: \"kubernetes.io/projected/e4c2921a-fff3-4c97-98df-206a94054d24-kube-api-access-dzdq9\") pod \"swift-proxy-646d45cbfc-6sjtl\" (UID: \"e4c2921a-fff3-4c97-98df-206a94054d24\") " pod="openstack/swift-proxy-646d45cbfc-6sjtl" Oct 05 21:13:40 crc kubenswrapper[4754]: I1005 21:13:40.482821 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e4c2921a-fff3-4c97-98df-206a94054d24-public-tls-certs\") pod \"swift-proxy-646d45cbfc-6sjtl\" (UID: \"e4c2921a-fff3-4c97-98df-206a94054d24\") " pod="openstack/swift-proxy-646d45cbfc-6sjtl" Oct 05 21:13:40 crc kubenswrapper[4754]: I1005 21:13:40.482843 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e4c2921a-fff3-4c97-98df-206a94054d24-internal-tls-certs\") pod \"swift-proxy-646d45cbfc-6sjtl\" (UID: \"e4c2921a-fff3-4c97-98df-206a94054d24\") " pod="openstack/swift-proxy-646d45cbfc-6sjtl" Oct 05 21:13:40 crc kubenswrapper[4754]: I1005 21:13:40.482897 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4c2921a-fff3-4c97-98df-206a94054d24-combined-ca-bundle\") pod \"swift-proxy-646d45cbfc-6sjtl\" (UID: \"e4c2921a-fff3-4c97-98df-206a94054d24\") " pod="openstack/swift-proxy-646d45cbfc-6sjtl" Oct 05 21:13:40 crc kubenswrapper[4754]: I1005 21:13:40.482924 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e4c2921a-fff3-4c97-98df-206a94054d24-log-httpd\") pod \"swift-proxy-646d45cbfc-6sjtl\" (UID: \"e4c2921a-fff3-4c97-98df-206a94054d24\") " pod="openstack/swift-proxy-646d45cbfc-6sjtl" Oct 05 21:13:40 crc kubenswrapper[4754]: I1005 21:13:40.482939 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4c2921a-fff3-4c97-98df-206a94054d24-config-data\") pod \"swift-proxy-646d45cbfc-6sjtl\" (UID: \"e4c2921a-fff3-4c97-98df-206a94054d24\") " pod="openstack/swift-proxy-646d45cbfc-6sjtl" Oct 05 21:13:40 crc kubenswrapper[4754]: I1005 21:13:40.482967 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/e4c2921a-fff3-4c97-98df-206a94054d24-etc-swift\") pod \"swift-proxy-646d45cbfc-6sjtl\" (UID: \"e4c2921a-fff3-4c97-98df-206a94054d24\") " pod="openstack/swift-proxy-646d45cbfc-6sjtl" Oct 05 21:13:40 crc kubenswrapper[4754]: I1005 21:13:40.594130 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e4c2921a-fff3-4c97-98df-206a94054d24-log-httpd\") pod \"swift-proxy-646d45cbfc-6sjtl\" (UID: \"e4c2921a-fff3-4c97-98df-206a94054d24\") " pod="openstack/swift-proxy-646d45cbfc-6sjtl" Oct 05 21:13:40 crc kubenswrapper[4754]: I1005 21:13:40.594187 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4c2921a-fff3-4c97-98df-206a94054d24-config-data\") pod \"swift-proxy-646d45cbfc-6sjtl\" (UID: \"e4c2921a-fff3-4c97-98df-206a94054d24\") " pod="openstack/swift-proxy-646d45cbfc-6sjtl" Oct 05 21:13:40 crc kubenswrapper[4754]: I1005 21:13:40.594241 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/e4c2921a-fff3-4c97-98df-206a94054d24-etc-swift\") pod \"swift-proxy-646d45cbfc-6sjtl\" (UID: \"e4c2921a-fff3-4c97-98df-206a94054d24\") " pod="openstack/swift-proxy-646d45cbfc-6sjtl" Oct 05 21:13:40 crc kubenswrapper[4754]: I1005 21:13:40.594340 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e4c2921a-fff3-4c97-98df-206a94054d24-run-httpd\") pod \"swift-proxy-646d45cbfc-6sjtl\" (UID: \"e4c2921a-fff3-4c97-98df-206a94054d24\") " pod="openstack/swift-proxy-646d45cbfc-6sjtl" Oct 05 21:13:40 crc kubenswrapper[4754]: I1005 21:13:40.594371 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dzdq9\" (UniqueName: \"kubernetes.io/projected/e4c2921a-fff3-4c97-98df-206a94054d24-kube-api-access-dzdq9\") pod \"swift-proxy-646d45cbfc-6sjtl\" (UID: \"e4c2921a-fff3-4c97-98df-206a94054d24\") " pod="openstack/swift-proxy-646d45cbfc-6sjtl" Oct 05 21:13:40 crc kubenswrapper[4754]: I1005 21:13:40.594506 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e4c2921a-fff3-4c97-98df-206a94054d24-public-tls-certs\") pod \"swift-proxy-646d45cbfc-6sjtl\" (UID: \"e4c2921a-fff3-4c97-98df-206a94054d24\") " pod="openstack/swift-proxy-646d45cbfc-6sjtl" Oct 05 21:13:40 crc kubenswrapper[4754]: I1005 21:13:40.594526 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e4c2921a-fff3-4c97-98df-206a94054d24-internal-tls-certs\") pod \"swift-proxy-646d45cbfc-6sjtl\" (UID: \"e4c2921a-fff3-4c97-98df-206a94054d24\") " pod="openstack/swift-proxy-646d45cbfc-6sjtl" Oct 05 21:13:40 crc kubenswrapper[4754]: I1005 21:13:40.594650 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4c2921a-fff3-4c97-98df-206a94054d24-combined-ca-bundle\") pod \"swift-proxy-646d45cbfc-6sjtl\" (UID: \"e4c2921a-fff3-4c97-98df-206a94054d24\") " pod="openstack/swift-proxy-646d45cbfc-6sjtl" Oct 05 21:13:40 crc kubenswrapper[4754]: I1005 21:13:40.596845 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e4c2921a-fff3-4c97-98df-206a94054d24-run-httpd\") pod \"swift-proxy-646d45cbfc-6sjtl\" (UID: \"e4c2921a-fff3-4c97-98df-206a94054d24\") " pod="openstack/swift-proxy-646d45cbfc-6sjtl" Oct 05 21:13:40 crc kubenswrapper[4754]: I1005 21:13:40.597111 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e4c2921a-fff3-4c97-98df-206a94054d24-log-httpd\") pod \"swift-proxy-646d45cbfc-6sjtl\" (UID: \"e4c2921a-fff3-4c97-98df-206a94054d24\") " pod="openstack/swift-proxy-646d45cbfc-6sjtl" Oct 05 21:13:40 crc kubenswrapper[4754]: I1005 21:13:40.606544 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e4c2921a-fff3-4c97-98df-206a94054d24-public-tls-certs\") pod \"swift-proxy-646d45cbfc-6sjtl\" (UID: \"e4c2921a-fff3-4c97-98df-206a94054d24\") " pod="openstack/swift-proxy-646d45cbfc-6sjtl" Oct 05 21:13:40 crc kubenswrapper[4754]: I1005 21:13:40.609218 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4c2921a-fff3-4c97-98df-206a94054d24-combined-ca-bundle\") pod \"swift-proxy-646d45cbfc-6sjtl\" (UID: \"e4c2921a-fff3-4c97-98df-206a94054d24\") " pod="openstack/swift-proxy-646d45cbfc-6sjtl" Oct 05 21:13:40 crc kubenswrapper[4754]: I1005 21:13:40.613515 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4c2921a-fff3-4c97-98df-206a94054d24-config-data\") pod \"swift-proxy-646d45cbfc-6sjtl\" (UID: \"e4c2921a-fff3-4c97-98df-206a94054d24\") " pod="openstack/swift-proxy-646d45cbfc-6sjtl" Oct 05 21:13:40 crc kubenswrapper[4754]: I1005 21:13:40.616111 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e4c2921a-fff3-4c97-98df-206a94054d24-internal-tls-certs\") pod \"swift-proxy-646d45cbfc-6sjtl\" (UID: \"e4c2921a-fff3-4c97-98df-206a94054d24\") " pod="openstack/swift-proxy-646d45cbfc-6sjtl" Oct 05 21:13:40 crc kubenswrapper[4754]: I1005 21:13:40.636993 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/e4c2921a-fff3-4c97-98df-206a94054d24-etc-swift\") pod \"swift-proxy-646d45cbfc-6sjtl\" (UID: \"e4c2921a-fff3-4c97-98df-206a94054d24\") " pod="openstack/swift-proxy-646d45cbfc-6sjtl" Oct 05 21:13:40 crc kubenswrapper[4754]: I1005 21:13:40.651817 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dzdq9\" (UniqueName: \"kubernetes.io/projected/e4c2921a-fff3-4c97-98df-206a94054d24-kube-api-access-dzdq9\") pod \"swift-proxy-646d45cbfc-6sjtl\" (UID: \"e4c2921a-fff3-4c97-98df-206a94054d24\") " pod="openstack/swift-proxy-646d45cbfc-6sjtl" Oct 05 21:13:40 crc kubenswrapper[4754]: I1005 21:13:40.744113 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-646d45cbfc-6sjtl" Oct 05 21:13:40 crc kubenswrapper[4754]: I1005 21:13:40.769591 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-5d6f8484b5-vstqw"] Oct 05 21:13:40 crc kubenswrapper[4754]: I1005 21:13:40.771265 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5d6f8484b5-vstqw" Oct 05 21:13:40 crc kubenswrapper[4754]: I1005 21:13:40.775215 4754 generic.go:334] "Generic (PLEG): container finished" podID="a0655389-4b2f-4973-b572-88df7ee28344" containerID="7e2f47e1eba20b443bc6c1598258c83ac78c624aa918294f7f18a7ff0e5b1770" exitCode=143 Oct 05 21:13:40 crc kubenswrapper[4754]: I1005 21:13:40.776856 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5fb78978fb-2mr8x" event={"ID":"a0655389-4b2f-4973-b572-88df7ee28344","Type":"ContainerDied","Data":"7e2f47e1eba20b443bc6c1598258c83ac78c624aa918294f7f18a7ff0e5b1770"} Oct 05 21:13:40 crc kubenswrapper[4754]: I1005 21:13:40.777423 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5784cf869f-dvhns" Oct 05 21:13:40 crc kubenswrapper[4754]: I1005 21:13:40.780933 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Oct 05 21:13:40 crc kubenswrapper[4754]: I1005 21:13:40.781348 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Oct 05 21:13:40 crc kubenswrapper[4754]: I1005 21:13:40.791797 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5d6f8484b5-vstqw"] Oct 05 21:13:40 crc kubenswrapper[4754]: I1005 21:13:40.904208 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/cb41212f-892e-4fe5-ae36-7ba898943277-httpd-config\") pod \"neutron-5d6f8484b5-vstqw\" (UID: \"cb41212f-892e-4fe5-ae36-7ba898943277\") " pod="openstack/neutron-5d6f8484b5-vstqw" Oct 05 21:13:40 crc kubenswrapper[4754]: I1005 21:13:40.911722 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb41212f-892e-4fe5-ae36-7ba898943277-combined-ca-bundle\") pod \"neutron-5d6f8484b5-vstqw\" (UID: \"cb41212f-892e-4fe5-ae36-7ba898943277\") " pod="openstack/neutron-5d6f8484b5-vstqw" Oct 05 21:13:40 crc kubenswrapper[4754]: I1005 21:13:40.911838 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/cb41212f-892e-4fe5-ae36-7ba898943277-ovndb-tls-certs\") pod \"neutron-5d6f8484b5-vstqw\" (UID: \"cb41212f-892e-4fe5-ae36-7ba898943277\") " pod="openstack/neutron-5d6f8484b5-vstqw" Oct 05 21:13:40 crc kubenswrapper[4754]: I1005 21:13:40.911948 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cb41212f-892e-4fe5-ae36-7ba898943277-internal-tls-certs\") pod \"neutron-5d6f8484b5-vstqw\" (UID: \"cb41212f-892e-4fe5-ae36-7ba898943277\") " pod="openstack/neutron-5d6f8484b5-vstqw" Oct 05 21:13:40 crc kubenswrapper[4754]: I1005 21:13:40.912226 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/cb41212f-892e-4fe5-ae36-7ba898943277-config\") pod \"neutron-5d6f8484b5-vstqw\" (UID: \"cb41212f-892e-4fe5-ae36-7ba898943277\") " pod="openstack/neutron-5d6f8484b5-vstqw" Oct 05 21:13:40 crc kubenswrapper[4754]: I1005 21:13:40.912295 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cb41212f-892e-4fe5-ae36-7ba898943277-public-tls-certs\") pod \"neutron-5d6f8484b5-vstqw\" (UID: \"cb41212f-892e-4fe5-ae36-7ba898943277\") " pod="openstack/neutron-5d6f8484b5-vstqw" Oct 05 21:13:40 crc kubenswrapper[4754]: I1005 21:13:40.912334 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ljfxv\" (UniqueName: \"kubernetes.io/projected/cb41212f-892e-4fe5-ae36-7ba898943277-kube-api-access-ljfxv\") pod \"neutron-5d6f8484b5-vstqw\" (UID: \"cb41212f-892e-4fe5-ae36-7ba898943277\") " pod="openstack/neutron-5d6f8484b5-vstqw" Oct 05 21:13:41 crc kubenswrapper[4754]: I1005 21:13:41.014171 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/cb41212f-892e-4fe5-ae36-7ba898943277-config\") pod \"neutron-5d6f8484b5-vstqw\" (UID: \"cb41212f-892e-4fe5-ae36-7ba898943277\") " pod="openstack/neutron-5d6f8484b5-vstqw" Oct 05 21:13:41 crc kubenswrapper[4754]: I1005 21:13:41.014225 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cb41212f-892e-4fe5-ae36-7ba898943277-public-tls-certs\") pod \"neutron-5d6f8484b5-vstqw\" (UID: \"cb41212f-892e-4fe5-ae36-7ba898943277\") " pod="openstack/neutron-5d6f8484b5-vstqw" Oct 05 21:13:41 crc kubenswrapper[4754]: I1005 21:13:41.014253 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ljfxv\" (UniqueName: \"kubernetes.io/projected/cb41212f-892e-4fe5-ae36-7ba898943277-kube-api-access-ljfxv\") pod \"neutron-5d6f8484b5-vstqw\" (UID: \"cb41212f-892e-4fe5-ae36-7ba898943277\") " pod="openstack/neutron-5d6f8484b5-vstqw" Oct 05 21:13:41 crc kubenswrapper[4754]: I1005 21:13:41.014307 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/cb41212f-892e-4fe5-ae36-7ba898943277-httpd-config\") pod \"neutron-5d6f8484b5-vstqw\" (UID: \"cb41212f-892e-4fe5-ae36-7ba898943277\") " pod="openstack/neutron-5d6f8484b5-vstqw" Oct 05 21:13:41 crc kubenswrapper[4754]: I1005 21:13:41.014338 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb41212f-892e-4fe5-ae36-7ba898943277-combined-ca-bundle\") pod \"neutron-5d6f8484b5-vstqw\" (UID: \"cb41212f-892e-4fe5-ae36-7ba898943277\") " pod="openstack/neutron-5d6f8484b5-vstqw" Oct 05 21:13:41 crc kubenswrapper[4754]: I1005 21:13:41.014363 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/cb41212f-892e-4fe5-ae36-7ba898943277-ovndb-tls-certs\") pod \"neutron-5d6f8484b5-vstqw\" (UID: \"cb41212f-892e-4fe5-ae36-7ba898943277\") " pod="openstack/neutron-5d6f8484b5-vstqw" Oct 05 21:13:41 crc kubenswrapper[4754]: I1005 21:13:41.014397 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cb41212f-892e-4fe5-ae36-7ba898943277-internal-tls-certs\") pod \"neutron-5d6f8484b5-vstqw\" (UID: \"cb41212f-892e-4fe5-ae36-7ba898943277\") " pod="openstack/neutron-5d6f8484b5-vstqw" Oct 05 21:13:41 crc kubenswrapper[4754]: I1005 21:13:41.025045 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/cb41212f-892e-4fe5-ae36-7ba898943277-ovndb-tls-certs\") pod \"neutron-5d6f8484b5-vstqw\" (UID: \"cb41212f-892e-4fe5-ae36-7ba898943277\") " pod="openstack/neutron-5d6f8484b5-vstqw" Oct 05 21:13:41 crc kubenswrapper[4754]: I1005 21:13:41.025243 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/cb41212f-892e-4fe5-ae36-7ba898943277-config\") pod \"neutron-5d6f8484b5-vstqw\" (UID: \"cb41212f-892e-4fe5-ae36-7ba898943277\") " pod="openstack/neutron-5d6f8484b5-vstqw" Oct 05 21:13:41 crc kubenswrapper[4754]: I1005 21:13:41.026454 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/cb41212f-892e-4fe5-ae36-7ba898943277-httpd-config\") pod \"neutron-5d6f8484b5-vstqw\" (UID: \"cb41212f-892e-4fe5-ae36-7ba898943277\") " pod="openstack/neutron-5d6f8484b5-vstqw" Oct 05 21:13:41 crc kubenswrapper[4754]: I1005 21:13:41.027155 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cb41212f-892e-4fe5-ae36-7ba898943277-internal-tls-certs\") pod \"neutron-5d6f8484b5-vstqw\" (UID: \"cb41212f-892e-4fe5-ae36-7ba898943277\") " pod="openstack/neutron-5d6f8484b5-vstqw" Oct 05 21:13:41 crc kubenswrapper[4754]: I1005 21:13:41.029986 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb41212f-892e-4fe5-ae36-7ba898943277-combined-ca-bundle\") pod \"neutron-5d6f8484b5-vstqw\" (UID: \"cb41212f-892e-4fe5-ae36-7ba898943277\") " pod="openstack/neutron-5d6f8484b5-vstqw" Oct 05 21:13:41 crc kubenswrapper[4754]: I1005 21:13:41.034957 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ljfxv\" (UniqueName: \"kubernetes.io/projected/cb41212f-892e-4fe5-ae36-7ba898943277-kube-api-access-ljfxv\") pod \"neutron-5d6f8484b5-vstqw\" (UID: \"cb41212f-892e-4fe5-ae36-7ba898943277\") " pod="openstack/neutron-5d6f8484b5-vstqw" Oct 05 21:13:41 crc kubenswrapper[4754]: I1005 21:13:41.035661 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cb41212f-892e-4fe5-ae36-7ba898943277-public-tls-certs\") pod \"neutron-5d6f8484b5-vstqw\" (UID: \"cb41212f-892e-4fe5-ae36-7ba898943277\") " pod="openstack/neutron-5d6f8484b5-vstqw" Oct 05 21:13:41 crc kubenswrapper[4754]: I1005 21:13:41.110394 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5d6f8484b5-vstqw" Oct 05 21:13:41 crc kubenswrapper[4754]: I1005 21:13:41.214948 4754 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cinder-api-0" podUID="0e4c84d5-ce7d-4b40-9b7c-980bc6b16c90" containerName="cinder-api" probeResult="failure" output="Get \"http://10.217.0.162:8776/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 05 21:13:41 crc kubenswrapper[4754]: I1005 21:13:41.643792 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Oct 05 21:13:41 crc kubenswrapper[4754]: I1005 21:13:41.649363 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-646d45cbfc-6sjtl"] Oct 05 21:13:41 crc kubenswrapper[4754]: I1005 21:13:41.694023 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 05 21:13:41 crc kubenswrapper[4754]: I1005 21:13:41.843602 4754 generic.go:334] "Generic (PLEG): container finished" podID="39d1aac3-14a1-45bf-9a9d-ba4cdb0afcb1" containerID="7ebc5fb1dedef86d2a80ae2a6e400e3f44aaa47d3b94955c70a007eb48092a6a" exitCode=0 Oct 05 21:13:41 crc kubenswrapper[4754]: I1005 21:13:41.843708 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-574454d6cb-vbnk6" event={"ID":"39d1aac3-14a1-45bf-9a9d-ba4cdb0afcb1","Type":"ContainerDied","Data":"7ebc5fb1dedef86d2a80ae2a6e400e3f44aaa47d3b94955c70a007eb48092a6a"} Oct 05 21:13:41 crc kubenswrapper[4754]: I1005 21:13:41.843761 4754 scope.go:117] "RemoveContainer" containerID="4c8ff8949b109b6ab42fd3b43db16a006408ec8d790928e7785deedd5ef7d473" Oct 05 21:13:41 crc kubenswrapper[4754]: I1005 21:13:41.849708 4754 generic.go:334] "Generic (PLEG): container finished" podID="27cd6be5-f5eb-4158-bdd1-37f4df3d530d" containerID="7d14664c74062c1ff554d9b5e811703f4db1749cb0131849ed4d6056fb07bc02" exitCode=0 Oct 05 21:13:41 crc kubenswrapper[4754]: I1005 21:13:41.851535 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-68c6b67864-9msm5" event={"ID":"27cd6be5-f5eb-4158-bdd1-37f4df3d530d","Type":"ContainerDied","Data":"7d14664c74062c1ff554d9b5e811703f4db1749cb0131849ed4d6056fb07bc02"} Oct 05 21:13:41 crc kubenswrapper[4754]: I1005 21:13:41.851851 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="5e757f47-326c-4174-9c28-c577a314f70b" containerName="cinder-scheduler" containerID="cri-o://5f27bec07a559ab193b04f03e1f03973140ba82e9195d2e4952585a2e52e40e7" gracePeriod=30 Oct 05 21:13:41 crc kubenswrapper[4754]: I1005 21:13:41.852140 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="5e757f47-326c-4174-9c28-c577a314f70b" containerName="probe" containerID="cri-o://a9fcf577241addd321a4099ecf2e6d85990ed2142aebe79888bfa7f0a64aeb46" gracePeriod=30 Oct 05 21:13:41 crc kubenswrapper[4754]: I1005 21:13:41.869736 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5d6f8484b5-vstqw"] Oct 05 21:13:42 crc kubenswrapper[4754]: I1005 21:13:42.215797 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-765c6857c8-q7glx" podUID="bd23550d-8815-4211-b8be-10ee8ab95c1f" containerName="barbican-api" probeResult="failure" output="Get \"https://10.217.0.159:9311/healthcheck\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 05 21:13:42 crc kubenswrapper[4754]: I1005 21:13:42.795518 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 05 21:13:43 crc kubenswrapper[4754]: I1005 21:13:43.874283 4754 generic.go:334] "Generic (PLEG): container finished" podID="5e757f47-326c-4174-9c28-c577a314f70b" containerID="a9fcf577241addd321a4099ecf2e6d85990ed2142aebe79888bfa7f0a64aeb46" exitCode=0 Oct 05 21:13:43 crc kubenswrapper[4754]: I1005 21:13:43.874737 4754 generic.go:334] "Generic (PLEG): container finished" podID="5e757f47-326c-4174-9c28-c577a314f70b" containerID="5f27bec07a559ab193b04f03e1f03973140ba82e9195d2e4952585a2e52e40e7" exitCode=0 Oct 05 21:13:43 crc kubenswrapper[4754]: I1005 21:13:43.874379 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"5e757f47-326c-4174-9c28-c577a314f70b","Type":"ContainerDied","Data":"a9fcf577241addd321a4099ecf2e6d85990ed2142aebe79888bfa7f0a64aeb46"} Oct 05 21:13:43 crc kubenswrapper[4754]: I1005 21:13:43.874793 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"5e757f47-326c-4174-9c28-c577a314f70b","Type":"ContainerDied","Data":"5f27bec07a559ab193b04f03e1f03973140ba82e9195d2e4952585a2e52e40e7"} Oct 05 21:13:44 crc kubenswrapper[4754]: I1005 21:13:44.036123 4754 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-5fb78978fb-2mr8x" podUID="a0655389-4b2f-4973-b572-88df7ee28344" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.158:9311/healthcheck\": read tcp 10.217.0.2:59480->10.217.0.158:9311: read: connection reset by peer" Oct 05 21:13:44 crc kubenswrapper[4754]: I1005 21:13:44.036160 4754 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-5fb78978fb-2mr8x" podUID="a0655389-4b2f-4973-b572-88df7ee28344" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.158:9311/healthcheck\": read tcp 10.217.0.2:59496->10.217.0.158:9311: read: connection reset by peer" Oct 05 21:13:44 crc kubenswrapper[4754]: I1005 21:13:44.316742 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 05 21:13:44 crc kubenswrapper[4754]: I1005 21:13:44.317264 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9fbf0155-8666-4b0c-a321-8f6c788b529d" containerName="ceilometer-central-agent" containerID="cri-o://9cba681c08c510789981ad324899d5474bd4c967035d7a4f8ea21e4f81a5b44f" gracePeriod=30 Oct 05 21:13:44 crc kubenswrapper[4754]: I1005 21:13:44.317731 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9fbf0155-8666-4b0c-a321-8f6c788b529d" containerName="proxy-httpd" containerID="cri-o://80d1fd901cea8412d079b32fc3b0d132e3b3755b31b3438ab736600c319c8d61" gracePeriod=30 Oct 05 21:13:44 crc kubenswrapper[4754]: I1005 21:13:44.317771 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9fbf0155-8666-4b0c-a321-8f6c788b529d" containerName="sg-core" containerID="cri-o://2e3e7b10d0983827e68056feff99b72e1a1bb822b837d9c738e1730f1e6b70e4" gracePeriod=30 Oct 05 21:13:44 crc kubenswrapper[4754]: I1005 21:13:44.317802 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9fbf0155-8666-4b0c-a321-8f6c788b529d" containerName="ceilometer-notification-agent" containerID="cri-o://5cf22c8730a91a7041f82212493156f36557b87917c64440ab525727adffaab2" gracePeriod=30 Oct 05 21:13:44 crc kubenswrapper[4754]: I1005 21:13:44.896782 4754 generic.go:334] "Generic (PLEG): container finished" podID="9fbf0155-8666-4b0c-a321-8f6c788b529d" containerID="80d1fd901cea8412d079b32fc3b0d132e3b3755b31b3438ab736600c319c8d61" exitCode=0 Oct 05 21:13:44 crc kubenswrapper[4754]: I1005 21:13:44.896828 4754 generic.go:334] "Generic (PLEG): container finished" podID="9fbf0155-8666-4b0c-a321-8f6c788b529d" containerID="2e3e7b10d0983827e68056feff99b72e1a1bb822b837d9c738e1730f1e6b70e4" exitCode=2 Oct 05 21:13:44 crc kubenswrapper[4754]: I1005 21:13:44.896903 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9fbf0155-8666-4b0c-a321-8f6c788b529d","Type":"ContainerDied","Data":"80d1fd901cea8412d079b32fc3b0d132e3b3755b31b3438ab736600c319c8d61"} Oct 05 21:13:44 crc kubenswrapper[4754]: I1005 21:13:44.896933 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9fbf0155-8666-4b0c-a321-8f6c788b529d","Type":"ContainerDied","Data":"2e3e7b10d0983827e68056feff99b72e1a1bb822b837d9c738e1730f1e6b70e4"} Oct 05 21:13:44 crc kubenswrapper[4754]: I1005 21:13:44.900627 4754 generic.go:334] "Generic (PLEG): container finished" podID="a0655389-4b2f-4973-b572-88df7ee28344" containerID="e09d95f4d0fe4bccba2be305d0ec39651d39f0c07e8fbb497b7f9bad6ff21f9b" exitCode=0 Oct 05 21:13:44 crc kubenswrapper[4754]: I1005 21:13:44.900681 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5fb78978fb-2mr8x" event={"ID":"a0655389-4b2f-4973-b572-88df7ee28344","Type":"ContainerDied","Data":"e09d95f4d0fe4bccba2be305d0ec39651d39f0c07e8fbb497b7f9bad6ff21f9b"} Oct 05 21:13:45 crc kubenswrapper[4754]: I1005 21:13:45.019706 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Oct 05 21:13:45 crc kubenswrapper[4754]: I1005 21:13:45.915228 4754 generic.go:334] "Generic (PLEG): container finished" podID="9fbf0155-8666-4b0c-a321-8f6c788b529d" containerID="5cf22c8730a91a7041f82212493156f36557b87917c64440ab525727adffaab2" exitCode=0 Oct 05 21:13:45 crc kubenswrapper[4754]: I1005 21:13:45.915261 4754 generic.go:334] "Generic (PLEG): container finished" podID="9fbf0155-8666-4b0c-a321-8f6c788b529d" containerID="9cba681c08c510789981ad324899d5474bd4c967035d7a4f8ea21e4f81a5b44f" exitCode=0 Oct 05 21:13:45 crc kubenswrapper[4754]: I1005 21:13:45.915297 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9fbf0155-8666-4b0c-a321-8f6c788b529d","Type":"ContainerDied","Data":"5cf22c8730a91a7041f82212493156f36557b87917c64440ab525727adffaab2"} Oct 05 21:13:45 crc kubenswrapper[4754]: I1005 21:13:45.915326 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9fbf0155-8666-4b0c-a321-8f6c788b529d","Type":"ContainerDied","Data":"9cba681c08c510789981ad324899d5474bd4c967035d7a4f8ea21e4f81a5b44f"} Oct 05 21:13:46 crc kubenswrapper[4754]: I1005 21:13:46.218875 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5784cf869f-dvhns" Oct 05 21:13:46 crc kubenswrapper[4754]: I1005 21:13:46.382938 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-69c986f6d7-fksbs"] Oct 05 21:13:46 crc kubenswrapper[4754]: I1005 21:13:46.383282 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-69c986f6d7-fksbs" podUID="c79a3ebb-fbdd-4680-aff0-27fbd7d8a7eb" containerName="dnsmasq-dns" containerID="cri-o://7bc4934b991c3e3b09d48b6d180324d91500beee9d23471553fa9e0b511ea99d" gracePeriod=10 Oct 05 21:13:46 crc kubenswrapper[4754]: I1005 21:13:46.515826 4754 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-5fb78978fb-2mr8x" podUID="a0655389-4b2f-4973-b572-88df7ee28344" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.158:9311/healthcheck\": dial tcp 10.217.0.158:9311: connect: connection refused" Oct 05 21:13:46 crc kubenswrapper[4754]: I1005 21:13:46.515823 4754 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-5fb78978fb-2mr8x" podUID="a0655389-4b2f-4973-b572-88df7ee28344" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.158:9311/healthcheck\": dial tcp 10.217.0.158:9311: connect: connection refused" Oct 05 21:13:46 crc kubenswrapper[4754]: I1005 21:13:46.930131 4754 generic.go:334] "Generic (PLEG): container finished" podID="c79a3ebb-fbdd-4680-aff0-27fbd7d8a7eb" containerID="7bc4934b991c3e3b09d48b6d180324d91500beee9d23471553fa9e0b511ea99d" exitCode=0 Oct 05 21:13:46 crc kubenswrapper[4754]: I1005 21:13:46.930180 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-69c986f6d7-fksbs" event={"ID":"c79a3ebb-fbdd-4680-aff0-27fbd7d8a7eb","Type":"ContainerDied","Data":"7bc4934b991c3e3b09d48b6d180324d91500beee9d23471553fa9e0b511ea99d"} Oct 05 21:13:50 crc kubenswrapper[4754]: W1005 21:13:50.556756 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcb41212f_892e_4fe5_ae36_7ba898943277.slice/crio-bd13280e27d37a06db0f258b9ce6a1102c63c090338abf9d1ba158b492fafcb9 WatchSource:0}: Error finding container bd13280e27d37a06db0f258b9ce6a1102c63c090338abf9d1ba158b492fafcb9: Status 404 returned error can't find the container with id bd13280e27d37a06db0f258b9ce6a1102c63c090338abf9d1ba158b492fafcb9 Oct 05 21:13:50 crc kubenswrapper[4754]: I1005 21:13:50.968443 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-646d45cbfc-6sjtl" event={"ID":"e4c2921a-fff3-4c97-98df-206a94054d24","Type":"ContainerStarted","Data":"d6de56786727947b18ab324a38df7977a88e07917d56e833c491c418f581ac44"} Oct 05 21:13:50 crc kubenswrapper[4754]: I1005 21:13:50.970558 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5d6f8484b5-vstqw" event={"ID":"cb41212f-892e-4fe5-ae36-7ba898943277","Type":"ContainerStarted","Data":"bd13280e27d37a06db0f258b9ce6a1102c63c090338abf9d1ba158b492fafcb9"} Oct 05 21:13:50 crc kubenswrapper[4754]: I1005 21:13:50.974083 4754 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-69c986f6d7-fksbs" podUID="c79a3ebb-fbdd-4680-aff0-27fbd7d8a7eb" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.161:5353: connect: connection refused" Oct 05 21:13:51 crc kubenswrapper[4754]: I1005 21:13:51.258836 4754 scope.go:117] "RemoveContainer" containerID="550fc17784cbed0030f262eb4233588b6160edd99bfa3b03eaec56012f272639" Oct 05 21:13:51 crc kubenswrapper[4754]: I1005 21:13:51.281351 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5fb78978fb-2mr8x" Oct 05 21:13:51 crc kubenswrapper[4754]: I1005 21:13:51.379769 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a0655389-4b2f-4973-b572-88df7ee28344-config-data-custom\") pod \"a0655389-4b2f-4973-b572-88df7ee28344\" (UID: \"a0655389-4b2f-4973-b572-88df7ee28344\") " Oct 05 21:13:51 crc kubenswrapper[4754]: I1005 21:13:51.379834 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0655389-4b2f-4973-b572-88df7ee28344-config-data\") pod \"a0655389-4b2f-4973-b572-88df7ee28344\" (UID: \"a0655389-4b2f-4973-b572-88df7ee28344\") " Oct 05 21:13:51 crc kubenswrapper[4754]: I1005 21:13:51.380030 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a0655389-4b2f-4973-b572-88df7ee28344-logs\") pod \"a0655389-4b2f-4973-b572-88df7ee28344\" (UID: \"a0655389-4b2f-4973-b572-88df7ee28344\") " Oct 05 21:13:51 crc kubenswrapper[4754]: I1005 21:13:51.380088 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0655389-4b2f-4973-b572-88df7ee28344-combined-ca-bundle\") pod \"a0655389-4b2f-4973-b572-88df7ee28344\" (UID: \"a0655389-4b2f-4973-b572-88df7ee28344\") " Oct 05 21:13:51 crc kubenswrapper[4754]: I1005 21:13:51.380122 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rq47t\" (UniqueName: \"kubernetes.io/projected/a0655389-4b2f-4973-b572-88df7ee28344-kube-api-access-rq47t\") pod \"a0655389-4b2f-4973-b572-88df7ee28344\" (UID: \"a0655389-4b2f-4973-b572-88df7ee28344\") " Oct 05 21:13:51 crc kubenswrapper[4754]: I1005 21:13:51.384223 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a0655389-4b2f-4973-b572-88df7ee28344-logs" (OuterVolumeSpecName: "logs") pod "a0655389-4b2f-4973-b572-88df7ee28344" (UID: "a0655389-4b2f-4973-b572-88df7ee28344"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 21:13:51 crc kubenswrapper[4754]: I1005 21:13:51.394376 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0655389-4b2f-4973-b572-88df7ee28344-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "a0655389-4b2f-4973-b572-88df7ee28344" (UID: "a0655389-4b2f-4973-b572-88df7ee28344"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:13:51 crc kubenswrapper[4754]: I1005 21:13:51.404458 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0655389-4b2f-4973-b572-88df7ee28344-kube-api-access-rq47t" (OuterVolumeSpecName: "kube-api-access-rq47t") pod "a0655389-4b2f-4973-b572-88df7ee28344" (UID: "a0655389-4b2f-4973-b572-88df7ee28344"). InnerVolumeSpecName "kube-api-access-rq47t". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:13:51 crc kubenswrapper[4754]: I1005 21:13:51.425221 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0655389-4b2f-4973-b572-88df7ee28344-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a0655389-4b2f-4973-b572-88df7ee28344" (UID: "a0655389-4b2f-4973-b572-88df7ee28344"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:13:51 crc kubenswrapper[4754]: I1005 21:13:51.445643 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0655389-4b2f-4973-b572-88df7ee28344-config-data" (OuterVolumeSpecName: "config-data") pod "a0655389-4b2f-4973-b572-88df7ee28344" (UID: "a0655389-4b2f-4973-b572-88df7ee28344"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:13:51 crc kubenswrapper[4754]: I1005 21:13:51.483408 4754 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0655389-4b2f-4973-b572-88df7ee28344-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 21:13:51 crc kubenswrapper[4754]: I1005 21:13:51.483436 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rq47t\" (UniqueName: \"kubernetes.io/projected/a0655389-4b2f-4973-b572-88df7ee28344-kube-api-access-rq47t\") on node \"crc\" DevicePath \"\"" Oct 05 21:13:51 crc kubenswrapper[4754]: I1005 21:13:51.483448 4754 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a0655389-4b2f-4973-b572-88df7ee28344-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 05 21:13:51 crc kubenswrapper[4754]: I1005 21:13:51.483481 4754 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0655389-4b2f-4973-b572-88df7ee28344-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 21:13:51 crc kubenswrapper[4754]: I1005 21:13:51.483540 4754 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a0655389-4b2f-4973-b572-88df7ee28344-logs\") on node \"crc\" DevicePath \"\"" Oct 05 21:13:51 crc kubenswrapper[4754]: I1005 21:13:51.583750 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 05 21:13:51 crc kubenswrapper[4754]: I1005 21:13:51.652911 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 05 21:13:51 crc kubenswrapper[4754]: I1005 21:13:51.674247 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-69c986f6d7-fksbs" Oct 05 21:13:51 crc kubenswrapper[4754]: I1005 21:13:51.690721 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e757f47-326c-4174-9c28-c577a314f70b-config-data\") pod \"5e757f47-326c-4174-9c28-c577a314f70b\" (UID: \"5e757f47-326c-4174-9c28-c577a314f70b\") " Oct 05 21:13:51 crc kubenswrapper[4754]: I1005 21:13:51.690800 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9fbf0155-8666-4b0c-a321-8f6c788b529d-log-httpd\") pod \"9fbf0155-8666-4b0c-a321-8f6c788b529d\" (UID: \"9fbf0155-8666-4b0c-a321-8f6c788b529d\") " Oct 05 21:13:51 crc kubenswrapper[4754]: I1005 21:13:51.690839 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9fbf0155-8666-4b0c-a321-8f6c788b529d-config-data\") pod \"9fbf0155-8666-4b0c-a321-8f6c788b529d\" (UID: \"9fbf0155-8666-4b0c-a321-8f6c788b529d\") " Oct 05 21:13:51 crc kubenswrapper[4754]: I1005 21:13:51.690867 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9fbf0155-8666-4b0c-a321-8f6c788b529d-combined-ca-bundle\") pod \"9fbf0155-8666-4b0c-a321-8f6c788b529d\" (UID: \"9fbf0155-8666-4b0c-a321-8f6c788b529d\") " Oct 05 21:13:51 crc kubenswrapper[4754]: I1005 21:13:51.690906 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bmvlt\" (UniqueName: \"kubernetes.io/projected/5e757f47-326c-4174-9c28-c577a314f70b-kube-api-access-bmvlt\") pod \"5e757f47-326c-4174-9c28-c577a314f70b\" (UID: \"5e757f47-326c-4174-9c28-c577a314f70b\") " Oct 05 21:13:51 crc kubenswrapper[4754]: I1005 21:13:51.690932 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9fbf0155-8666-4b0c-a321-8f6c788b529d-run-httpd\") pod \"9fbf0155-8666-4b0c-a321-8f6c788b529d\" (UID: \"9fbf0155-8666-4b0c-a321-8f6c788b529d\") " Oct 05 21:13:51 crc kubenswrapper[4754]: I1005 21:13:51.690962 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5e757f47-326c-4174-9c28-c577a314f70b-config-data-custom\") pod \"5e757f47-326c-4174-9c28-c577a314f70b\" (UID: \"5e757f47-326c-4174-9c28-c577a314f70b\") " Oct 05 21:13:51 crc kubenswrapper[4754]: I1005 21:13:51.690986 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9fbf0155-8666-4b0c-a321-8f6c788b529d-sg-core-conf-yaml\") pod \"9fbf0155-8666-4b0c-a321-8f6c788b529d\" (UID: \"9fbf0155-8666-4b0c-a321-8f6c788b529d\") " Oct 05 21:13:51 crc kubenswrapper[4754]: I1005 21:13:51.691083 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e757f47-326c-4174-9c28-c577a314f70b-combined-ca-bundle\") pod \"5e757f47-326c-4174-9c28-c577a314f70b\" (UID: \"5e757f47-326c-4174-9c28-c577a314f70b\") " Oct 05 21:13:51 crc kubenswrapper[4754]: I1005 21:13:51.691121 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5e757f47-326c-4174-9c28-c577a314f70b-scripts\") pod \"5e757f47-326c-4174-9c28-c577a314f70b\" (UID: \"5e757f47-326c-4174-9c28-c577a314f70b\") " Oct 05 21:13:51 crc kubenswrapper[4754]: I1005 21:13:51.691184 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wpr2t\" (UniqueName: \"kubernetes.io/projected/9fbf0155-8666-4b0c-a321-8f6c788b529d-kube-api-access-wpr2t\") pod \"9fbf0155-8666-4b0c-a321-8f6c788b529d\" (UID: \"9fbf0155-8666-4b0c-a321-8f6c788b529d\") " Oct 05 21:13:51 crc kubenswrapper[4754]: I1005 21:13:51.691202 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5e757f47-326c-4174-9c28-c577a314f70b-etc-machine-id\") pod \"5e757f47-326c-4174-9c28-c577a314f70b\" (UID: \"5e757f47-326c-4174-9c28-c577a314f70b\") " Oct 05 21:13:51 crc kubenswrapper[4754]: I1005 21:13:51.691219 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9fbf0155-8666-4b0c-a321-8f6c788b529d-scripts\") pod \"9fbf0155-8666-4b0c-a321-8f6c788b529d\" (UID: \"9fbf0155-8666-4b0c-a321-8f6c788b529d\") " Oct 05 21:13:51 crc kubenswrapper[4754]: I1005 21:13:51.691809 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9fbf0155-8666-4b0c-a321-8f6c788b529d-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "9fbf0155-8666-4b0c-a321-8f6c788b529d" (UID: "9fbf0155-8666-4b0c-a321-8f6c788b529d"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 21:13:51 crc kubenswrapper[4754]: I1005 21:13:51.693864 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9fbf0155-8666-4b0c-a321-8f6c788b529d-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "9fbf0155-8666-4b0c-a321-8f6c788b529d" (UID: "9fbf0155-8666-4b0c-a321-8f6c788b529d"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 21:13:51 crc kubenswrapper[4754]: I1005 21:13:51.710454 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5e757f47-326c-4174-9c28-c577a314f70b-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "5e757f47-326c-4174-9c28-c577a314f70b" (UID: "5e757f47-326c-4174-9c28-c577a314f70b"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 05 21:13:51 crc kubenswrapper[4754]: I1005 21:13:51.739677 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9fbf0155-8666-4b0c-a321-8f6c788b529d-scripts" (OuterVolumeSpecName: "scripts") pod "9fbf0155-8666-4b0c-a321-8f6c788b529d" (UID: "9fbf0155-8666-4b0c-a321-8f6c788b529d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:13:51 crc kubenswrapper[4754]: I1005 21:13:51.744723 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5e757f47-326c-4174-9c28-c577a314f70b-kube-api-access-bmvlt" (OuterVolumeSpecName: "kube-api-access-bmvlt") pod "5e757f47-326c-4174-9c28-c577a314f70b" (UID: "5e757f47-326c-4174-9c28-c577a314f70b"). InnerVolumeSpecName "kube-api-access-bmvlt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:13:51 crc kubenswrapper[4754]: I1005 21:13:51.752064 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e757f47-326c-4174-9c28-c577a314f70b-scripts" (OuterVolumeSpecName: "scripts") pod "5e757f47-326c-4174-9c28-c577a314f70b" (UID: "5e757f47-326c-4174-9c28-c577a314f70b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:13:51 crc kubenswrapper[4754]: I1005 21:13:51.757739 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9fbf0155-8666-4b0c-a321-8f6c788b529d-kube-api-access-wpr2t" (OuterVolumeSpecName: "kube-api-access-wpr2t") pod "9fbf0155-8666-4b0c-a321-8f6c788b529d" (UID: "9fbf0155-8666-4b0c-a321-8f6c788b529d"). InnerVolumeSpecName "kube-api-access-wpr2t". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:13:51 crc kubenswrapper[4754]: I1005 21:13:51.771708 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e757f47-326c-4174-9c28-c577a314f70b-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "5e757f47-326c-4174-9c28-c577a314f70b" (UID: "5e757f47-326c-4174-9c28-c577a314f70b"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:13:51 crc kubenswrapper[4754]: I1005 21:13:51.805261 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c79a3ebb-fbdd-4680-aff0-27fbd7d8a7eb-config\") pod \"c79a3ebb-fbdd-4680-aff0-27fbd7d8a7eb\" (UID: \"c79a3ebb-fbdd-4680-aff0-27fbd7d8a7eb\") " Oct 05 21:13:51 crc kubenswrapper[4754]: I1005 21:13:51.805302 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c79a3ebb-fbdd-4680-aff0-27fbd7d8a7eb-ovsdbserver-nb\") pod \"c79a3ebb-fbdd-4680-aff0-27fbd7d8a7eb\" (UID: \"c79a3ebb-fbdd-4680-aff0-27fbd7d8a7eb\") " Oct 05 21:13:51 crc kubenswrapper[4754]: I1005 21:13:51.805351 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c79a3ebb-fbdd-4680-aff0-27fbd7d8a7eb-dns-swift-storage-0\") pod \"c79a3ebb-fbdd-4680-aff0-27fbd7d8a7eb\" (UID: \"c79a3ebb-fbdd-4680-aff0-27fbd7d8a7eb\") " Oct 05 21:13:51 crc kubenswrapper[4754]: I1005 21:13:51.805533 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nkvdp\" (UniqueName: \"kubernetes.io/projected/c79a3ebb-fbdd-4680-aff0-27fbd7d8a7eb-kube-api-access-nkvdp\") pod \"c79a3ebb-fbdd-4680-aff0-27fbd7d8a7eb\" (UID: \"c79a3ebb-fbdd-4680-aff0-27fbd7d8a7eb\") " Oct 05 21:13:51 crc kubenswrapper[4754]: I1005 21:13:51.805561 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c79a3ebb-fbdd-4680-aff0-27fbd7d8a7eb-ovsdbserver-sb\") pod \"c79a3ebb-fbdd-4680-aff0-27fbd7d8a7eb\" (UID: \"c79a3ebb-fbdd-4680-aff0-27fbd7d8a7eb\") " Oct 05 21:13:51 crc kubenswrapper[4754]: I1005 21:13:51.805625 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c79a3ebb-fbdd-4680-aff0-27fbd7d8a7eb-dns-svc\") pod \"c79a3ebb-fbdd-4680-aff0-27fbd7d8a7eb\" (UID: \"c79a3ebb-fbdd-4680-aff0-27fbd7d8a7eb\") " Oct 05 21:13:51 crc kubenswrapper[4754]: I1005 21:13:51.806027 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bmvlt\" (UniqueName: \"kubernetes.io/projected/5e757f47-326c-4174-9c28-c577a314f70b-kube-api-access-bmvlt\") on node \"crc\" DevicePath \"\"" Oct 05 21:13:51 crc kubenswrapper[4754]: I1005 21:13:51.806044 4754 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9fbf0155-8666-4b0c-a321-8f6c788b529d-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 05 21:13:51 crc kubenswrapper[4754]: I1005 21:13:51.806063 4754 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5e757f47-326c-4174-9c28-c577a314f70b-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 05 21:13:51 crc kubenswrapper[4754]: I1005 21:13:51.806072 4754 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5e757f47-326c-4174-9c28-c577a314f70b-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 21:13:51 crc kubenswrapper[4754]: I1005 21:13:51.806081 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wpr2t\" (UniqueName: \"kubernetes.io/projected/9fbf0155-8666-4b0c-a321-8f6c788b529d-kube-api-access-wpr2t\") on node \"crc\" DevicePath \"\"" Oct 05 21:13:51 crc kubenswrapper[4754]: I1005 21:13:51.806090 4754 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5e757f47-326c-4174-9c28-c577a314f70b-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 05 21:13:51 crc kubenswrapper[4754]: I1005 21:13:51.806099 4754 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9fbf0155-8666-4b0c-a321-8f6c788b529d-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 21:13:51 crc kubenswrapper[4754]: I1005 21:13:51.806107 4754 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9fbf0155-8666-4b0c-a321-8f6c788b529d-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 05 21:13:51 crc kubenswrapper[4754]: I1005 21:13:51.830962 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c79a3ebb-fbdd-4680-aff0-27fbd7d8a7eb-kube-api-access-nkvdp" (OuterVolumeSpecName: "kube-api-access-nkvdp") pod "c79a3ebb-fbdd-4680-aff0-27fbd7d8a7eb" (UID: "c79a3ebb-fbdd-4680-aff0-27fbd7d8a7eb"). InnerVolumeSpecName "kube-api-access-nkvdp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:13:51 crc kubenswrapper[4754]: I1005 21:13:51.909895 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nkvdp\" (UniqueName: \"kubernetes.io/projected/c79a3ebb-fbdd-4680-aff0-27fbd7d8a7eb-kube-api-access-nkvdp\") on node \"crc\" DevicePath \"\"" Oct 05 21:13:51 crc kubenswrapper[4754]: I1005 21:13:51.995722 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9fbf0155-8666-4b0c-a321-8f6c788b529d","Type":"ContainerDied","Data":"383df62c656a37781ded70118f6132b7393e16d8c2c33715b6ebe1528d5715dc"} Oct 05 21:13:51 crc kubenswrapper[4754]: I1005 21:13:51.995777 4754 scope.go:117] "RemoveContainer" containerID="80d1fd901cea8412d079b32fc3b0d132e3b3755b31b3438ab736600c319c8d61" Oct 05 21:13:51 crc kubenswrapper[4754]: I1005 21:13:51.995871 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 05 21:13:52 crc kubenswrapper[4754]: I1005 21:13:52.004944 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5fb78978fb-2mr8x" event={"ID":"a0655389-4b2f-4973-b572-88df7ee28344","Type":"ContainerDied","Data":"7905d5d472685f664cbeb567f15d32bbde1d4a02eb2b026d1f81786c2145c5bc"} Oct 05 21:13:52 crc kubenswrapper[4754]: I1005 21:13:52.005109 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5fb78978fb-2mr8x" Oct 05 21:13:52 crc kubenswrapper[4754]: I1005 21:13:52.009394 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-69c986f6d7-fksbs" Oct 05 21:13:52 crc kubenswrapper[4754]: I1005 21:13:52.009267 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-69c986f6d7-fksbs" event={"ID":"c79a3ebb-fbdd-4680-aff0-27fbd7d8a7eb","Type":"ContainerDied","Data":"ce0e640bbc5d27b424afee479a96ad561119313962435a85170dfa605e5931a7"} Oct 05 21:13:52 crc kubenswrapper[4754]: I1005 21:13:52.016377 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-68c6b67864-9msm5" event={"ID":"27cd6be5-f5eb-4158-bdd1-37f4df3d530d","Type":"ContainerStarted","Data":"8cbf678dfb968e1a5f68bc5b282346cef6ff3c4725e9f27e572e9658e525228e"} Oct 05 21:13:52 crc kubenswrapper[4754]: I1005 21:13:52.019264 4754 scope.go:117] "RemoveContainer" containerID="2e3e7b10d0983827e68056feff99b72e1a1bb822b837d9c738e1730f1e6b70e4" Oct 05 21:13:52 crc kubenswrapper[4754]: I1005 21:13:52.023763 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5d6f8484b5-vstqw" event={"ID":"cb41212f-892e-4fe5-ae36-7ba898943277","Type":"ContainerStarted","Data":"f9af415a07d94b384308165b032015e1432ebb58966b6d58a7ecae3c51ee9fcf"} Oct 05 21:13:52 crc kubenswrapper[4754]: I1005 21:13:52.034542 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"5e757f47-326c-4174-9c28-c577a314f70b","Type":"ContainerDied","Data":"d2230dbff5c9e90a0a5d763f1160225660a5953534e2327fdd30590657c1a469"} Oct 05 21:13:52 crc kubenswrapper[4754]: I1005 21:13:52.034682 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 05 21:13:52 crc kubenswrapper[4754]: I1005 21:13:52.053653 4754 scope.go:117] "RemoveContainer" containerID="5cf22c8730a91a7041f82212493156f36557b87917c64440ab525727adffaab2" Oct 05 21:13:52 crc kubenswrapper[4754]: I1005 21:13:52.091658 4754 scope.go:117] "RemoveContainer" containerID="9cba681c08c510789981ad324899d5474bd4c967035d7a4f8ea21e4f81a5b44f" Oct 05 21:13:52 crc kubenswrapper[4754]: I1005 21:13:52.097594 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-5fb78978fb-2mr8x"] Oct 05 21:13:52 crc kubenswrapper[4754]: I1005 21:13:52.109269 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-5fb78978fb-2mr8x"] Oct 05 21:13:52 crc kubenswrapper[4754]: I1005 21:13:52.160316 4754 scope.go:117] "RemoveContainer" containerID="e09d95f4d0fe4bccba2be305d0ec39651d39f0c07e8fbb497b7f9bad6ff21f9b" Oct 05 21:13:52 crc kubenswrapper[4754]: I1005 21:13:52.210800 4754 scope.go:117] "RemoveContainer" containerID="7e2f47e1eba20b443bc6c1598258c83ac78c624aa918294f7f18a7ff0e5b1770" Oct 05 21:13:52 crc kubenswrapper[4754]: I1005 21:13:52.250673 4754 scope.go:117] "RemoveContainer" containerID="7bc4934b991c3e3b09d48b6d180324d91500beee9d23471553fa9e0b511ea99d" Oct 05 21:13:52 crc kubenswrapper[4754]: I1005 21:13:52.274857 4754 scope.go:117] "RemoveContainer" containerID="21ba4cc396f6c99a169e2dfa92b344ae6aa2976e2c3fc8e4fd96e75d9f6ca6c0" Oct 05 21:13:52 crc kubenswrapper[4754]: I1005 21:13:52.311990 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c79a3ebb-fbdd-4680-aff0-27fbd7d8a7eb-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "c79a3ebb-fbdd-4680-aff0-27fbd7d8a7eb" (UID: "c79a3ebb-fbdd-4680-aff0-27fbd7d8a7eb"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 21:13:52 crc kubenswrapper[4754]: I1005 21:13:52.319208 4754 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c79a3ebb-fbdd-4680-aff0-27fbd7d8a7eb-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 05 21:13:52 crc kubenswrapper[4754]: I1005 21:13:52.321656 4754 scope.go:117] "RemoveContainer" containerID="a9fcf577241addd321a4099ecf2e6d85990ed2142aebe79888bfa7f0a64aeb46" Oct 05 21:13:52 crc kubenswrapper[4754]: I1005 21:13:52.484719 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9fbf0155-8666-4b0c-a321-8f6c788b529d-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "9fbf0155-8666-4b0c-a321-8f6c788b529d" (UID: "9fbf0155-8666-4b0c-a321-8f6c788b529d"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:13:52 crc kubenswrapper[4754]: I1005 21:13:52.493571 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c79a3ebb-fbdd-4680-aff0-27fbd7d8a7eb-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "c79a3ebb-fbdd-4680-aff0-27fbd7d8a7eb" (UID: "c79a3ebb-fbdd-4680-aff0-27fbd7d8a7eb"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 21:13:52 crc kubenswrapper[4754]: I1005 21:13:52.499998 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c79a3ebb-fbdd-4680-aff0-27fbd7d8a7eb-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "c79a3ebb-fbdd-4680-aff0-27fbd7d8a7eb" (UID: "c79a3ebb-fbdd-4680-aff0-27fbd7d8a7eb"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 21:13:52 crc kubenswrapper[4754]: I1005 21:13:52.507699 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c79a3ebb-fbdd-4680-aff0-27fbd7d8a7eb-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "c79a3ebb-fbdd-4680-aff0-27fbd7d8a7eb" (UID: "c79a3ebb-fbdd-4680-aff0-27fbd7d8a7eb"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 21:13:52 crc kubenswrapper[4754]: I1005 21:13:52.534895 4754 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c79a3ebb-fbdd-4680-aff0-27fbd7d8a7eb-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 05 21:13:52 crc kubenswrapper[4754]: I1005 21:13:52.534934 4754 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c79a3ebb-fbdd-4680-aff0-27fbd7d8a7eb-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 05 21:13:52 crc kubenswrapper[4754]: I1005 21:13:52.534945 4754 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c79a3ebb-fbdd-4680-aff0-27fbd7d8a7eb-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 05 21:13:52 crc kubenswrapper[4754]: I1005 21:13:52.534956 4754 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9fbf0155-8666-4b0c-a321-8f6c788b529d-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 05 21:13:52 crc kubenswrapper[4754]: I1005 21:13:52.569106 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c79a3ebb-fbdd-4680-aff0-27fbd7d8a7eb-config" (OuterVolumeSpecName: "config") pod "c79a3ebb-fbdd-4680-aff0-27fbd7d8a7eb" (UID: "c79a3ebb-fbdd-4680-aff0-27fbd7d8a7eb"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 21:13:52 crc kubenswrapper[4754]: I1005 21:13:52.577555 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e757f47-326c-4174-9c28-c577a314f70b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5e757f47-326c-4174-9c28-c577a314f70b" (UID: "5e757f47-326c-4174-9c28-c577a314f70b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:13:52 crc kubenswrapper[4754]: I1005 21:13:52.602738 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e757f47-326c-4174-9c28-c577a314f70b-config-data" (OuterVolumeSpecName: "config-data") pod "5e757f47-326c-4174-9c28-c577a314f70b" (UID: "5e757f47-326c-4174-9c28-c577a314f70b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:13:52 crc kubenswrapper[4754]: I1005 21:13:52.614349 4754 scope.go:117] "RemoveContainer" containerID="5f27bec07a559ab193b04f03e1f03973140ba82e9195d2e4952585a2e52e40e7" Oct 05 21:13:52 crc kubenswrapper[4754]: I1005 21:13:52.636362 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9fbf0155-8666-4b0c-a321-8f6c788b529d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9fbf0155-8666-4b0c-a321-8f6c788b529d" (UID: "9fbf0155-8666-4b0c-a321-8f6c788b529d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:13:52 crc kubenswrapper[4754]: I1005 21:13:52.637215 4754 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e757f47-326c-4174-9c28-c577a314f70b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 21:13:52 crc kubenswrapper[4754]: I1005 21:13:52.637273 4754 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c79a3ebb-fbdd-4680-aff0-27fbd7d8a7eb-config\") on node \"crc\" DevicePath \"\"" Oct 05 21:13:52 crc kubenswrapper[4754]: I1005 21:13:52.637286 4754 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e757f47-326c-4174-9c28-c577a314f70b-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 21:13:52 crc kubenswrapper[4754]: I1005 21:13:52.637295 4754 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9fbf0155-8666-4b0c-a321-8f6c788b529d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 21:13:52 crc kubenswrapper[4754]: I1005 21:13:52.692625 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-69c986f6d7-fksbs"] Oct 05 21:13:52 crc kubenswrapper[4754]: I1005 21:13:52.712924 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-69c986f6d7-fksbs"] Oct 05 21:13:52 crc kubenswrapper[4754]: I1005 21:13:52.713810 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9fbf0155-8666-4b0c-a321-8f6c788b529d-config-data" (OuterVolumeSpecName: "config-data") pod "9fbf0155-8666-4b0c-a321-8f6c788b529d" (UID: "9fbf0155-8666-4b0c-a321-8f6c788b529d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:13:52 crc kubenswrapper[4754]: I1005 21:13:52.732579 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 05 21:13:52 crc kubenswrapper[4754]: I1005 21:13:52.738887 4754 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9fbf0155-8666-4b0c-a321-8f6c788b529d-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 21:13:52 crc kubenswrapper[4754]: I1005 21:13:52.747849 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 05 21:13:52 crc kubenswrapper[4754]: I1005 21:13:52.760581 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Oct 05 21:13:52 crc kubenswrapper[4754]: E1005 21:13:52.761103 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9fbf0155-8666-4b0c-a321-8f6c788b529d" containerName="sg-core" Oct 05 21:13:52 crc kubenswrapper[4754]: I1005 21:13:52.761127 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="9fbf0155-8666-4b0c-a321-8f6c788b529d" containerName="sg-core" Oct 05 21:13:52 crc kubenswrapper[4754]: E1005 21:13:52.761140 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c79a3ebb-fbdd-4680-aff0-27fbd7d8a7eb" containerName="init" Oct 05 21:13:52 crc kubenswrapper[4754]: I1005 21:13:52.761148 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="c79a3ebb-fbdd-4680-aff0-27fbd7d8a7eb" containerName="init" Oct 05 21:13:52 crc kubenswrapper[4754]: E1005 21:13:52.761164 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e757f47-326c-4174-9c28-c577a314f70b" containerName="cinder-scheduler" Oct 05 21:13:52 crc kubenswrapper[4754]: I1005 21:13:52.761172 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e757f47-326c-4174-9c28-c577a314f70b" containerName="cinder-scheduler" Oct 05 21:13:52 crc kubenswrapper[4754]: E1005 21:13:52.761190 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9fbf0155-8666-4b0c-a321-8f6c788b529d" containerName="ceilometer-notification-agent" Oct 05 21:13:52 crc kubenswrapper[4754]: I1005 21:13:52.761196 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="9fbf0155-8666-4b0c-a321-8f6c788b529d" containerName="ceilometer-notification-agent" Oct 05 21:13:52 crc kubenswrapper[4754]: E1005 21:13:52.761210 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0655389-4b2f-4973-b572-88df7ee28344" containerName="barbican-api-log" Oct 05 21:13:52 crc kubenswrapper[4754]: I1005 21:13:52.761217 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0655389-4b2f-4973-b572-88df7ee28344" containerName="barbican-api-log" Oct 05 21:13:52 crc kubenswrapper[4754]: E1005 21:13:52.761230 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c79a3ebb-fbdd-4680-aff0-27fbd7d8a7eb" containerName="dnsmasq-dns" Oct 05 21:13:52 crc kubenswrapper[4754]: I1005 21:13:52.761236 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="c79a3ebb-fbdd-4680-aff0-27fbd7d8a7eb" containerName="dnsmasq-dns" Oct 05 21:13:52 crc kubenswrapper[4754]: E1005 21:13:52.761245 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0655389-4b2f-4973-b572-88df7ee28344" containerName="barbican-api" Oct 05 21:13:52 crc kubenswrapper[4754]: I1005 21:13:52.761250 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0655389-4b2f-4973-b572-88df7ee28344" containerName="barbican-api" Oct 05 21:13:52 crc kubenswrapper[4754]: E1005 21:13:52.761265 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9fbf0155-8666-4b0c-a321-8f6c788b529d" containerName="proxy-httpd" Oct 05 21:13:52 crc kubenswrapper[4754]: I1005 21:13:52.761271 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="9fbf0155-8666-4b0c-a321-8f6c788b529d" containerName="proxy-httpd" Oct 05 21:13:52 crc kubenswrapper[4754]: E1005 21:13:52.761292 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e757f47-326c-4174-9c28-c577a314f70b" containerName="probe" Oct 05 21:13:52 crc kubenswrapper[4754]: I1005 21:13:52.761300 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e757f47-326c-4174-9c28-c577a314f70b" containerName="probe" Oct 05 21:13:52 crc kubenswrapper[4754]: E1005 21:13:52.761326 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9fbf0155-8666-4b0c-a321-8f6c788b529d" containerName="ceilometer-central-agent" Oct 05 21:13:52 crc kubenswrapper[4754]: I1005 21:13:52.761333 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="9fbf0155-8666-4b0c-a321-8f6c788b529d" containerName="ceilometer-central-agent" Oct 05 21:13:52 crc kubenswrapper[4754]: I1005 21:13:52.761535 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="a0655389-4b2f-4973-b572-88df7ee28344" containerName="barbican-api" Oct 05 21:13:52 crc kubenswrapper[4754]: I1005 21:13:52.761552 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e757f47-326c-4174-9c28-c577a314f70b" containerName="cinder-scheduler" Oct 05 21:13:52 crc kubenswrapper[4754]: I1005 21:13:52.761559 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="a0655389-4b2f-4973-b572-88df7ee28344" containerName="barbican-api-log" Oct 05 21:13:52 crc kubenswrapper[4754]: I1005 21:13:52.761568 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e757f47-326c-4174-9c28-c577a314f70b" containerName="probe" Oct 05 21:13:52 crc kubenswrapper[4754]: I1005 21:13:52.761579 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="9fbf0155-8666-4b0c-a321-8f6c788b529d" containerName="proxy-httpd" Oct 05 21:13:52 crc kubenswrapper[4754]: I1005 21:13:52.761595 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="c79a3ebb-fbdd-4680-aff0-27fbd7d8a7eb" containerName="dnsmasq-dns" Oct 05 21:13:52 crc kubenswrapper[4754]: I1005 21:13:52.761604 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="9fbf0155-8666-4b0c-a321-8f6c788b529d" containerName="ceilometer-notification-agent" Oct 05 21:13:52 crc kubenswrapper[4754]: I1005 21:13:52.761616 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="9fbf0155-8666-4b0c-a321-8f6c788b529d" containerName="sg-core" Oct 05 21:13:52 crc kubenswrapper[4754]: I1005 21:13:52.761627 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="9fbf0155-8666-4b0c-a321-8f6c788b529d" containerName="ceilometer-central-agent" Oct 05 21:13:52 crc kubenswrapper[4754]: I1005 21:13:52.762820 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 05 21:13:52 crc kubenswrapper[4754]: I1005 21:13:52.766136 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Oct 05 21:13:52 crc kubenswrapper[4754]: I1005 21:13:52.776067 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 05 21:13:52 crc kubenswrapper[4754]: I1005 21:13:52.840801 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8da021ab-a4b1-4a42-bf58-2973a0578d8e-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"8da021ab-a4b1-4a42-bf58-2973a0578d8e\") " pod="openstack/cinder-scheduler-0" Oct 05 21:13:52 crc kubenswrapper[4754]: I1005 21:13:52.840870 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8da021ab-a4b1-4a42-bf58-2973a0578d8e-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"8da021ab-a4b1-4a42-bf58-2973a0578d8e\") " pod="openstack/cinder-scheduler-0" Oct 05 21:13:52 crc kubenswrapper[4754]: I1005 21:13:52.840931 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8da021ab-a4b1-4a42-bf58-2973a0578d8e-config-data\") pod \"cinder-scheduler-0\" (UID: \"8da021ab-a4b1-4a42-bf58-2973a0578d8e\") " pod="openstack/cinder-scheduler-0" Oct 05 21:13:52 crc kubenswrapper[4754]: I1005 21:13:52.840999 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8da021ab-a4b1-4a42-bf58-2973a0578d8e-scripts\") pod \"cinder-scheduler-0\" (UID: \"8da021ab-a4b1-4a42-bf58-2973a0578d8e\") " pod="openstack/cinder-scheduler-0" Oct 05 21:13:52 crc kubenswrapper[4754]: I1005 21:13:52.841048 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5bmrx\" (UniqueName: \"kubernetes.io/projected/8da021ab-a4b1-4a42-bf58-2973a0578d8e-kube-api-access-5bmrx\") pod \"cinder-scheduler-0\" (UID: \"8da021ab-a4b1-4a42-bf58-2973a0578d8e\") " pod="openstack/cinder-scheduler-0" Oct 05 21:13:52 crc kubenswrapper[4754]: I1005 21:13:52.841104 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8da021ab-a4b1-4a42-bf58-2973a0578d8e-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"8da021ab-a4b1-4a42-bf58-2973a0578d8e\") " pod="openstack/cinder-scheduler-0" Oct 05 21:13:52 crc kubenswrapper[4754]: I1005 21:13:52.868551 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5e757f47-326c-4174-9c28-c577a314f70b" path="/var/lib/kubelet/pods/5e757f47-326c-4174-9c28-c577a314f70b/volumes" Oct 05 21:13:52 crc kubenswrapper[4754]: I1005 21:13:52.869203 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0655389-4b2f-4973-b572-88df7ee28344" path="/var/lib/kubelet/pods/a0655389-4b2f-4973-b572-88df7ee28344/volumes" Oct 05 21:13:52 crc kubenswrapper[4754]: I1005 21:13:52.869795 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c79a3ebb-fbdd-4680-aff0-27fbd7d8a7eb" path="/var/lib/kubelet/pods/c79a3ebb-fbdd-4680-aff0-27fbd7d8a7eb/volumes" Oct 05 21:13:52 crc kubenswrapper[4754]: I1005 21:13:52.943745 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8da021ab-a4b1-4a42-bf58-2973a0578d8e-scripts\") pod \"cinder-scheduler-0\" (UID: \"8da021ab-a4b1-4a42-bf58-2973a0578d8e\") " pod="openstack/cinder-scheduler-0" Oct 05 21:13:52 crc kubenswrapper[4754]: I1005 21:13:52.943831 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5bmrx\" (UniqueName: \"kubernetes.io/projected/8da021ab-a4b1-4a42-bf58-2973a0578d8e-kube-api-access-5bmrx\") pod \"cinder-scheduler-0\" (UID: \"8da021ab-a4b1-4a42-bf58-2973a0578d8e\") " pod="openstack/cinder-scheduler-0" Oct 05 21:13:52 crc kubenswrapper[4754]: I1005 21:13:52.943900 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8da021ab-a4b1-4a42-bf58-2973a0578d8e-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"8da021ab-a4b1-4a42-bf58-2973a0578d8e\") " pod="openstack/cinder-scheduler-0" Oct 05 21:13:52 crc kubenswrapper[4754]: I1005 21:13:52.943967 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8da021ab-a4b1-4a42-bf58-2973a0578d8e-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"8da021ab-a4b1-4a42-bf58-2973a0578d8e\") " pod="openstack/cinder-scheduler-0" Oct 05 21:13:52 crc kubenswrapper[4754]: I1005 21:13:52.943999 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8da021ab-a4b1-4a42-bf58-2973a0578d8e-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"8da021ab-a4b1-4a42-bf58-2973a0578d8e\") " pod="openstack/cinder-scheduler-0" Oct 05 21:13:52 crc kubenswrapper[4754]: I1005 21:13:52.944035 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8da021ab-a4b1-4a42-bf58-2973a0578d8e-config-data\") pod \"cinder-scheduler-0\" (UID: \"8da021ab-a4b1-4a42-bf58-2973a0578d8e\") " pod="openstack/cinder-scheduler-0" Oct 05 21:13:52 crc kubenswrapper[4754]: I1005 21:13:52.945780 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 05 21:13:52 crc kubenswrapper[4754]: I1005 21:13:52.947516 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8da021ab-a4b1-4a42-bf58-2973a0578d8e-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"8da021ab-a4b1-4a42-bf58-2973a0578d8e\") " pod="openstack/cinder-scheduler-0" Oct 05 21:13:52 crc kubenswrapper[4754]: I1005 21:13:52.956857 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 05 21:13:52 crc kubenswrapper[4754]: I1005 21:13:52.975131 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8da021ab-a4b1-4a42-bf58-2973a0578d8e-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"8da021ab-a4b1-4a42-bf58-2973a0578d8e\") " pod="openstack/cinder-scheduler-0" Oct 05 21:13:52 crc kubenswrapper[4754]: I1005 21:13:52.975611 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8da021ab-a4b1-4a42-bf58-2973a0578d8e-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"8da021ab-a4b1-4a42-bf58-2973a0578d8e\") " pod="openstack/cinder-scheduler-0" Oct 05 21:13:52 crc kubenswrapper[4754]: I1005 21:13:52.977960 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8da021ab-a4b1-4a42-bf58-2973a0578d8e-config-data\") pod \"cinder-scheduler-0\" (UID: \"8da021ab-a4b1-4a42-bf58-2973a0578d8e\") " pod="openstack/cinder-scheduler-0" Oct 05 21:13:52 crc kubenswrapper[4754]: I1005 21:13:52.979183 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5bmrx\" (UniqueName: \"kubernetes.io/projected/8da021ab-a4b1-4a42-bf58-2973a0578d8e-kube-api-access-5bmrx\") pod \"cinder-scheduler-0\" (UID: \"8da021ab-a4b1-4a42-bf58-2973a0578d8e\") " pod="openstack/cinder-scheduler-0" Oct 05 21:13:52 crc kubenswrapper[4754]: I1005 21:13:52.979739 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8da021ab-a4b1-4a42-bf58-2973a0578d8e-scripts\") pod \"cinder-scheduler-0\" (UID: \"8da021ab-a4b1-4a42-bf58-2973a0578d8e\") " pod="openstack/cinder-scheduler-0" Oct 05 21:13:53 crc kubenswrapper[4754]: I1005 21:13:53.003729 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 05 21:13:53 crc kubenswrapper[4754]: I1005 21:13:53.006397 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 05 21:13:53 crc kubenswrapper[4754]: I1005 21:13:53.013435 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 05 21:13:53 crc kubenswrapper[4754]: I1005 21:13:53.014656 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 05 21:13:53 crc kubenswrapper[4754]: I1005 21:13:53.015714 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 05 21:13:53 crc kubenswrapper[4754]: I1005 21:13:53.046348 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c7f03a15-953e-4de8-b7b7-711cc35c8323-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c7f03a15-953e-4de8-b7b7-711cc35c8323\") " pod="openstack/ceilometer-0" Oct 05 21:13:53 crc kubenswrapper[4754]: I1005 21:13:53.046393 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c7f03a15-953e-4de8-b7b7-711cc35c8323-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c7f03a15-953e-4de8-b7b7-711cc35c8323\") " pod="openstack/ceilometer-0" Oct 05 21:13:53 crc kubenswrapper[4754]: I1005 21:13:53.046448 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-74zgd\" (UniqueName: \"kubernetes.io/projected/c7f03a15-953e-4de8-b7b7-711cc35c8323-kube-api-access-74zgd\") pod \"ceilometer-0\" (UID: \"c7f03a15-953e-4de8-b7b7-711cc35c8323\") " pod="openstack/ceilometer-0" Oct 05 21:13:53 crc kubenswrapper[4754]: I1005 21:13:53.046473 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c7f03a15-953e-4de8-b7b7-711cc35c8323-scripts\") pod \"ceilometer-0\" (UID: \"c7f03a15-953e-4de8-b7b7-711cc35c8323\") " pod="openstack/ceilometer-0" Oct 05 21:13:53 crc kubenswrapper[4754]: I1005 21:13:53.046532 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c7f03a15-953e-4de8-b7b7-711cc35c8323-run-httpd\") pod \"ceilometer-0\" (UID: \"c7f03a15-953e-4de8-b7b7-711cc35c8323\") " pod="openstack/ceilometer-0" Oct 05 21:13:53 crc kubenswrapper[4754]: I1005 21:13:53.046575 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c7f03a15-953e-4de8-b7b7-711cc35c8323-log-httpd\") pod \"ceilometer-0\" (UID: \"c7f03a15-953e-4de8-b7b7-711cc35c8323\") " pod="openstack/ceilometer-0" Oct 05 21:13:53 crc kubenswrapper[4754]: I1005 21:13:53.046602 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c7f03a15-953e-4de8-b7b7-711cc35c8323-config-data\") pod \"ceilometer-0\" (UID: \"c7f03a15-953e-4de8-b7b7-711cc35c8323\") " pod="openstack/ceilometer-0" Oct 05 21:13:53 crc kubenswrapper[4754]: I1005 21:13:53.054899 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"95ef109d-c48b-40ea-b491-3e730d4d651b","Type":"ContainerStarted","Data":"4a984ea8b959ebebc99ca3f73b6fd7776bd410518c4a309a19f265fac62a8195"} Oct 05 21:13:53 crc kubenswrapper[4754]: I1005 21:13:53.067239 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-574454d6cb-vbnk6" event={"ID":"39d1aac3-14a1-45bf-9a9d-ba4cdb0afcb1","Type":"ContainerStarted","Data":"704413f5b17f48e4e3d7d104afa7e9d7beadfdcc2c562b18589f4f0a201b7281"} Oct 05 21:13:53 crc kubenswrapper[4754]: I1005 21:13:53.072524 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-646d45cbfc-6sjtl" event={"ID":"e4c2921a-fff3-4c97-98df-206a94054d24","Type":"ContainerStarted","Data":"4ae82dd4c745d99a31579c6ce4a317e8af4ff9de8d5a88641daf0ee268880ccf"} Oct 05 21:13:53 crc kubenswrapper[4754]: I1005 21:13:53.072597 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-646d45cbfc-6sjtl" event={"ID":"e4c2921a-fff3-4c97-98df-206a94054d24","Type":"ContainerStarted","Data":"dd68e9937b511b6cc7f06fe2ad82332c010191c3ccbe42f50a8a14037abb9006"} Oct 05 21:13:53 crc kubenswrapper[4754]: I1005 21:13:53.073645 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-646d45cbfc-6sjtl" Oct 05 21:13:53 crc kubenswrapper[4754]: I1005 21:13:53.073683 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-646d45cbfc-6sjtl" Oct 05 21:13:53 crc kubenswrapper[4754]: I1005 21:13:53.076639 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5d6f8484b5-vstqw" event={"ID":"cb41212f-892e-4fe5-ae36-7ba898943277","Type":"ContainerStarted","Data":"18806f1a9ff5527e9bf88dca4b41ac7c4a878da5e6a8a20f1fbd1215817a99d9"} Oct 05 21:13:53 crc kubenswrapper[4754]: I1005 21:13:53.076712 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-5d6f8484b5-vstqw" Oct 05 21:13:53 crc kubenswrapper[4754]: I1005 21:13:53.086277 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.771487489 podStartE2EDuration="29.086255715s" podCreationTimestamp="2025-10-05 21:13:24 +0000 UTC" firstStartedPulling="2025-10-05 21:13:25.191149125 +0000 UTC m=+1129.095267835" lastFinishedPulling="2025-10-05 21:13:51.505917351 +0000 UTC m=+1155.410036061" observedRunningTime="2025-10-05 21:13:53.076054677 +0000 UTC m=+1156.980173397" watchObservedRunningTime="2025-10-05 21:13:53.086255715 +0000 UTC m=+1156.990374425" Oct 05 21:13:53 crc kubenswrapper[4754]: I1005 21:13:53.103190 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 05 21:13:53 crc kubenswrapper[4754]: I1005 21:13:53.134343 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-646d45cbfc-6sjtl" podStartSLOduration=13.13431651 podStartE2EDuration="13.13431651s" podCreationTimestamp="2025-10-05 21:13:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 21:13:53.121234636 +0000 UTC m=+1157.025353346" watchObservedRunningTime="2025-10-05 21:13:53.13431651 +0000 UTC m=+1157.038435220" Oct 05 21:13:53 crc kubenswrapper[4754]: I1005 21:13:53.150204 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c7f03a15-953e-4de8-b7b7-711cc35c8323-run-httpd\") pod \"ceilometer-0\" (UID: \"c7f03a15-953e-4de8-b7b7-711cc35c8323\") " pod="openstack/ceilometer-0" Oct 05 21:13:53 crc kubenswrapper[4754]: I1005 21:13:53.150363 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c7f03a15-953e-4de8-b7b7-711cc35c8323-log-httpd\") pod \"ceilometer-0\" (UID: \"c7f03a15-953e-4de8-b7b7-711cc35c8323\") " pod="openstack/ceilometer-0" Oct 05 21:13:53 crc kubenswrapper[4754]: I1005 21:13:53.150466 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c7f03a15-953e-4de8-b7b7-711cc35c8323-config-data\") pod \"ceilometer-0\" (UID: \"c7f03a15-953e-4de8-b7b7-711cc35c8323\") " pod="openstack/ceilometer-0" Oct 05 21:13:53 crc kubenswrapper[4754]: I1005 21:13:53.150590 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c7f03a15-953e-4de8-b7b7-711cc35c8323-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c7f03a15-953e-4de8-b7b7-711cc35c8323\") " pod="openstack/ceilometer-0" Oct 05 21:13:53 crc kubenswrapper[4754]: I1005 21:13:53.150618 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c7f03a15-953e-4de8-b7b7-711cc35c8323-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c7f03a15-953e-4de8-b7b7-711cc35c8323\") " pod="openstack/ceilometer-0" Oct 05 21:13:53 crc kubenswrapper[4754]: I1005 21:13:53.150756 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-74zgd\" (UniqueName: \"kubernetes.io/projected/c7f03a15-953e-4de8-b7b7-711cc35c8323-kube-api-access-74zgd\") pod \"ceilometer-0\" (UID: \"c7f03a15-953e-4de8-b7b7-711cc35c8323\") " pod="openstack/ceilometer-0" Oct 05 21:13:53 crc kubenswrapper[4754]: I1005 21:13:53.150787 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c7f03a15-953e-4de8-b7b7-711cc35c8323-scripts\") pod \"ceilometer-0\" (UID: \"c7f03a15-953e-4de8-b7b7-711cc35c8323\") " pod="openstack/ceilometer-0" Oct 05 21:13:53 crc kubenswrapper[4754]: I1005 21:13:53.155849 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c7f03a15-953e-4de8-b7b7-711cc35c8323-run-httpd\") pod \"ceilometer-0\" (UID: \"c7f03a15-953e-4de8-b7b7-711cc35c8323\") " pod="openstack/ceilometer-0" Oct 05 21:13:53 crc kubenswrapper[4754]: I1005 21:13:53.162279 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c7f03a15-953e-4de8-b7b7-711cc35c8323-log-httpd\") pod \"ceilometer-0\" (UID: \"c7f03a15-953e-4de8-b7b7-711cc35c8323\") " pod="openstack/ceilometer-0" Oct 05 21:13:53 crc kubenswrapper[4754]: I1005 21:13:53.175084 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c7f03a15-953e-4de8-b7b7-711cc35c8323-config-data\") pod \"ceilometer-0\" (UID: \"c7f03a15-953e-4de8-b7b7-711cc35c8323\") " pod="openstack/ceilometer-0" Oct 05 21:13:53 crc kubenswrapper[4754]: I1005 21:13:53.177422 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c7f03a15-953e-4de8-b7b7-711cc35c8323-scripts\") pod \"ceilometer-0\" (UID: \"c7f03a15-953e-4de8-b7b7-711cc35c8323\") " pod="openstack/ceilometer-0" Oct 05 21:13:53 crc kubenswrapper[4754]: I1005 21:13:53.193752 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c7f03a15-953e-4de8-b7b7-711cc35c8323-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c7f03a15-953e-4de8-b7b7-711cc35c8323\") " pod="openstack/ceilometer-0" Oct 05 21:13:53 crc kubenswrapper[4754]: I1005 21:13:53.197042 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c7f03a15-953e-4de8-b7b7-711cc35c8323-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c7f03a15-953e-4de8-b7b7-711cc35c8323\") " pod="openstack/ceilometer-0" Oct 05 21:13:53 crc kubenswrapper[4754]: I1005 21:13:53.202898 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-5d6f8484b5-vstqw" podStartSLOduration=13.202883364 podStartE2EDuration="13.202883364s" podCreationTimestamp="2025-10-05 21:13:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 21:13:53.161619488 +0000 UTC m=+1157.065738198" watchObservedRunningTime="2025-10-05 21:13:53.202883364 +0000 UTC m=+1157.107002064" Oct 05 21:13:53 crc kubenswrapper[4754]: I1005 21:13:53.214297 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-74zgd\" (UniqueName: \"kubernetes.io/projected/c7f03a15-953e-4de8-b7b7-711cc35c8323-kube-api-access-74zgd\") pod \"ceilometer-0\" (UID: \"c7f03a15-953e-4de8-b7b7-711cc35c8323\") " pod="openstack/ceilometer-0" Oct 05 21:13:53 crc kubenswrapper[4754]: I1005 21:13:53.353935 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 05 21:13:53 crc kubenswrapper[4754]: I1005 21:13:53.922281 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 05 21:13:54 crc kubenswrapper[4754]: I1005 21:13:54.090252 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 05 21:13:54 crc kubenswrapper[4754]: I1005 21:13:54.112465 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"8da021ab-a4b1-4a42-bf58-2973a0578d8e","Type":"ContainerStarted","Data":"f8ede6745562ef46b004634250f9af684c2a50bf398335a0f52e06d13a07fddb"} Oct 05 21:13:54 crc kubenswrapper[4754]: I1005 21:13:54.929799 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9fbf0155-8666-4b0c-a321-8f6c788b529d" path="/var/lib/kubelet/pods/9fbf0155-8666-4b0c-a321-8f6c788b529d/volumes" Oct 05 21:13:55 crc kubenswrapper[4754]: I1005 21:13:55.122991 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c7f03a15-953e-4de8-b7b7-711cc35c8323","Type":"ContainerStarted","Data":"e720f41e8cb279f7b8b91bb1549ff3dfe463fbbe226f20aeff8c426f0e8aa76e"} Oct 05 21:13:55 crc kubenswrapper[4754]: I1005 21:13:55.130514 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"8da021ab-a4b1-4a42-bf58-2973a0578d8e","Type":"ContainerStarted","Data":"89fee3caf33896549b1f6608e244a127430401b387d4520e0038d60959fb7383"} Oct 05 21:13:55 crc kubenswrapper[4754]: I1005 21:13:55.603214 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-68c6b67864-9msm5" Oct 05 21:13:55 crc kubenswrapper[4754]: I1005 21:13:55.605184 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-68c6b67864-9msm5" Oct 05 21:13:55 crc kubenswrapper[4754]: I1005 21:13:55.721488 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-574454d6cb-vbnk6" Oct 05 21:13:55 crc kubenswrapper[4754]: I1005 21:13:55.721567 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-574454d6cb-vbnk6" Oct 05 21:13:56 crc kubenswrapper[4754]: I1005 21:13:56.140710 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c7f03a15-953e-4de8-b7b7-711cc35c8323","Type":"ContainerStarted","Data":"c81e9c627545c8ae7fec11f2768aba757f9b5f96a53410f2bbb3e6958d352382"} Oct 05 21:13:56 crc kubenswrapper[4754]: I1005 21:13:56.144361 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"8da021ab-a4b1-4a42-bf58-2973a0578d8e","Type":"ContainerStarted","Data":"92f2a2b431f12b9c164be4b7c936ad6a8da731c6f7eb5a17a71cecd7e25d4cdf"} Oct 05 21:13:56 crc kubenswrapper[4754]: I1005 21:13:56.421065 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=4.421042515 podStartE2EDuration="4.421042515s" podCreationTimestamp="2025-10-05 21:13:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 21:13:56.172762882 +0000 UTC m=+1160.076881592" watchObservedRunningTime="2025-10-05 21:13:56.421042515 +0000 UTC m=+1160.325161225" Oct 05 21:13:56 crc kubenswrapper[4754]: I1005 21:13:56.433115 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 05 21:13:56 crc kubenswrapper[4754]: I1005 21:13:56.433344 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="280f4e00-063d-486e-bd58-f4b246f54814" containerName="kube-state-metrics" containerID="cri-o://a59a9eb09e684bcff101063ecb5b4c6e28d60f5edeefe13b2f5db1d3257258be" gracePeriod=30 Oct 05 21:13:57 crc kubenswrapper[4754]: I1005 21:13:57.155780 4754 generic.go:334] "Generic (PLEG): container finished" podID="280f4e00-063d-486e-bd58-f4b246f54814" containerID="a59a9eb09e684bcff101063ecb5b4c6e28d60f5edeefe13b2f5db1d3257258be" exitCode=2 Oct 05 21:13:57 crc kubenswrapper[4754]: I1005 21:13:57.155917 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"280f4e00-063d-486e-bd58-f4b246f54814","Type":"ContainerDied","Data":"a59a9eb09e684bcff101063ecb5b4c6e28d60f5edeefe13b2f5db1d3257258be"} Oct 05 21:13:58 crc kubenswrapper[4754]: I1005 21:13:58.035688 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 05 21:13:58 crc kubenswrapper[4754]: I1005 21:13:58.103849 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Oct 05 21:13:58 crc kubenswrapper[4754]: I1005 21:13:58.167510 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c7f03a15-953e-4de8-b7b7-711cc35c8323","Type":"ContainerStarted","Data":"41c9b85f7408628829308d18d3aeee6968858f7de70656a6fe724f64f802b04f"} Oct 05 21:13:58 crc kubenswrapper[4754]: I1005 21:13:58.172703 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 05 21:13:58 crc kubenswrapper[4754]: I1005 21:13:58.173550 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"280f4e00-063d-486e-bd58-f4b246f54814","Type":"ContainerDied","Data":"9120a0bee39fd16a94bd3da4f4382072be682ad13213dc94566d9b585afb4750"} Oct 05 21:13:58 crc kubenswrapper[4754]: I1005 21:13:58.173580 4754 scope.go:117] "RemoveContainer" containerID="a59a9eb09e684bcff101063ecb5b4c6e28d60f5edeefe13b2f5db1d3257258be" Oct 05 21:13:58 crc kubenswrapper[4754]: I1005 21:13:58.179242 4754 generic.go:334] "Generic (PLEG): container finished" podID="0e4c84d5-ce7d-4b40-9b7c-980bc6b16c90" containerID="7d8e02519c00ab8dab67118f5aed2e8b655977781ccf0af4ff4939c450a39aef" exitCode=137 Oct 05 21:13:58 crc kubenswrapper[4754]: I1005 21:13:58.179405 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"0e4c84d5-ce7d-4b40-9b7c-980bc6b16c90","Type":"ContainerDied","Data":"7d8e02519c00ab8dab67118f5aed2e8b655977781ccf0af4ff4939c450a39aef"} Oct 05 21:13:58 crc kubenswrapper[4754]: I1005 21:13:58.197346 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qh7bz\" (UniqueName: \"kubernetes.io/projected/280f4e00-063d-486e-bd58-f4b246f54814-kube-api-access-qh7bz\") pod \"280f4e00-063d-486e-bd58-f4b246f54814\" (UID: \"280f4e00-063d-486e-bd58-f4b246f54814\") " Oct 05 21:13:58 crc kubenswrapper[4754]: I1005 21:13:58.205453 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/280f4e00-063d-486e-bd58-f4b246f54814-kube-api-access-qh7bz" (OuterVolumeSpecName: "kube-api-access-qh7bz") pod "280f4e00-063d-486e-bd58-f4b246f54814" (UID: "280f4e00-063d-486e-bd58-f4b246f54814"). InnerVolumeSpecName "kube-api-access-qh7bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:13:58 crc kubenswrapper[4754]: I1005 21:13:58.300082 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qh7bz\" (UniqueName: \"kubernetes.io/projected/280f4e00-063d-486e-bd58-f4b246f54814-kube-api-access-qh7bz\") on node \"crc\" DevicePath \"\"" Oct 05 21:13:58 crc kubenswrapper[4754]: I1005 21:13:58.402992 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 05 21:13:58 crc kubenswrapper[4754]: I1005 21:13:58.522926 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e4c84d5-ce7d-4b40-9b7c-980bc6b16c90-config-data\") pod \"0e4c84d5-ce7d-4b40-9b7c-980bc6b16c90\" (UID: \"0e4c84d5-ce7d-4b40-9b7c-980bc6b16c90\") " Oct 05 21:13:58 crc kubenswrapper[4754]: I1005 21:13:58.523006 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e4c84d5-ce7d-4b40-9b7c-980bc6b16c90-combined-ca-bundle\") pod \"0e4c84d5-ce7d-4b40-9b7c-980bc6b16c90\" (UID: \"0e4c84d5-ce7d-4b40-9b7c-980bc6b16c90\") " Oct 05 21:13:58 crc kubenswrapper[4754]: I1005 21:13:58.523239 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0e4c84d5-ce7d-4b40-9b7c-980bc6b16c90-config-data-custom\") pod \"0e4c84d5-ce7d-4b40-9b7c-980bc6b16c90\" (UID: \"0e4c84d5-ce7d-4b40-9b7c-980bc6b16c90\") " Oct 05 21:13:58 crc kubenswrapper[4754]: I1005 21:13:58.523284 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0e4c84d5-ce7d-4b40-9b7c-980bc6b16c90-scripts\") pod \"0e4c84d5-ce7d-4b40-9b7c-980bc6b16c90\" (UID: \"0e4c84d5-ce7d-4b40-9b7c-980bc6b16c90\") " Oct 05 21:13:58 crc kubenswrapper[4754]: I1005 21:13:58.523392 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0e4c84d5-ce7d-4b40-9b7c-980bc6b16c90-etc-machine-id\") pod \"0e4c84d5-ce7d-4b40-9b7c-980bc6b16c90\" (UID: \"0e4c84d5-ce7d-4b40-9b7c-980bc6b16c90\") " Oct 05 21:13:58 crc kubenswrapper[4754]: I1005 21:13:58.523444 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0e4c84d5-ce7d-4b40-9b7c-980bc6b16c90-logs\") pod \"0e4c84d5-ce7d-4b40-9b7c-980bc6b16c90\" (UID: \"0e4c84d5-ce7d-4b40-9b7c-980bc6b16c90\") " Oct 05 21:13:58 crc kubenswrapper[4754]: I1005 21:13:58.523608 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gnckd\" (UniqueName: \"kubernetes.io/projected/0e4c84d5-ce7d-4b40-9b7c-980bc6b16c90-kube-api-access-gnckd\") pod \"0e4c84d5-ce7d-4b40-9b7c-980bc6b16c90\" (UID: \"0e4c84d5-ce7d-4b40-9b7c-980bc6b16c90\") " Oct 05 21:13:58 crc kubenswrapper[4754]: I1005 21:13:58.528898 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0e4c84d5-ce7d-4b40-9b7c-980bc6b16c90-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "0e4c84d5-ce7d-4b40-9b7c-980bc6b16c90" (UID: "0e4c84d5-ce7d-4b40-9b7c-980bc6b16c90"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 05 21:13:58 crc kubenswrapper[4754]: I1005 21:13:58.529712 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0e4c84d5-ce7d-4b40-9b7c-980bc6b16c90-logs" (OuterVolumeSpecName: "logs") pod "0e4c84d5-ce7d-4b40-9b7c-980bc6b16c90" (UID: "0e4c84d5-ce7d-4b40-9b7c-980bc6b16c90"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 21:13:58 crc kubenswrapper[4754]: I1005 21:13:58.535213 4754 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0e4c84d5-ce7d-4b40-9b7c-980bc6b16c90-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 05 21:13:58 crc kubenswrapper[4754]: I1005 21:13:58.535246 4754 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0e4c84d5-ce7d-4b40-9b7c-980bc6b16c90-logs\") on node \"crc\" DevicePath \"\"" Oct 05 21:13:58 crc kubenswrapper[4754]: I1005 21:13:58.545771 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0e4c84d5-ce7d-4b40-9b7c-980bc6b16c90-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "0e4c84d5-ce7d-4b40-9b7c-980bc6b16c90" (UID: "0e4c84d5-ce7d-4b40-9b7c-980bc6b16c90"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:13:58 crc kubenswrapper[4754]: I1005 21:13:58.554081 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0e4c84d5-ce7d-4b40-9b7c-980bc6b16c90-kube-api-access-gnckd" (OuterVolumeSpecName: "kube-api-access-gnckd") pod "0e4c84d5-ce7d-4b40-9b7c-980bc6b16c90" (UID: "0e4c84d5-ce7d-4b40-9b7c-980bc6b16c90"). InnerVolumeSpecName "kube-api-access-gnckd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:13:58 crc kubenswrapper[4754]: I1005 21:13:58.554277 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0e4c84d5-ce7d-4b40-9b7c-980bc6b16c90-scripts" (OuterVolumeSpecName: "scripts") pod "0e4c84d5-ce7d-4b40-9b7c-980bc6b16c90" (UID: "0e4c84d5-ce7d-4b40-9b7c-980bc6b16c90"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:13:58 crc kubenswrapper[4754]: I1005 21:13:58.559681 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 05 21:13:58 crc kubenswrapper[4754]: I1005 21:13:58.603670 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 05 21:13:58 crc kubenswrapper[4754]: I1005 21:13:58.606807 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0e4c84d5-ce7d-4b40-9b7c-980bc6b16c90-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0e4c84d5-ce7d-4b40-9b7c-980bc6b16c90" (UID: "0e4c84d5-ce7d-4b40-9b7c-980bc6b16c90"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:13:58 crc kubenswrapper[4754]: I1005 21:13:58.638509 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gnckd\" (UniqueName: \"kubernetes.io/projected/0e4c84d5-ce7d-4b40-9b7c-980bc6b16c90-kube-api-access-gnckd\") on node \"crc\" DevicePath \"\"" Oct 05 21:13:58 crc kubenswrapper[4754]: I1005 21:13:58.638542 4754 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e4c84d5-ce7d-4b40-9b7c-980bc6b16c90-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 21:13:58 crc kubenswrapper[4754]: I1005 21:13:58.638552 4754 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0e4c84d5-ce7d-4b40-9b7c-980bc6b16c90-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 05 21:13:58 crc kubenswrapper[4754]: I1005 21:13:58.638657 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Oct 05 21:13:58 crc kubenswrapper[4754]: E1005 21:13:58.639253 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="280f4e00-063d-486e-bd58-f4b246f54814" containerName="kube-state-metrics" Oct 05 21:13:58 crc kubenswrapper[4754]: I1005 21:13:58.639266 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="280f4e00-063d-486e-bd58-f4b246f54814" containerName="kube-state-metrics" Oct 05 21:13:58 crc kubenswrapper[4754]: E1005 21:13:58.639285 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e4c84d5-ce7d-4b40-9b7c-980bc6b16c90" containerName="cinder-api" Oct 05 21:13:58 crc kubenswrapper[4754]: I1005 21:13:58.639293 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e4c84d5-ce7d-4b40-9b7c-980bc6b16c90" containerName="cinder-api" Oct 05 21:13:58 crc kubenswrapper[4754]: E1005 21:13:58.639316 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e4c84d5-ce7d-4b40-9b7c-980bc6b16c90" containerName="cinder-api-log" Oct 05 21:13:58 crc kubenswrapper[4754]: I1005 21:13:58.639324 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e4c84d5-ce7d-4b40-9b7c-980bc6b16c90" containerName="cinder-api-log" Oct 05 21:13:58 crc kubenswrapper[4754]: I1005 21:13:58.639383 4754 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0e4c84d5-ce7d-4b40-9b7c-980bc6b16c90-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 21:13:58 crc kubenswrapper[4754]: I1005 21:13:58.648778 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="280f4e00-063d-486e-bd58-f4b246f54814" containerName="kube-state-metrics" Oct 05 21:13:58 crc kubenswrapper[4754]: I1005 21:13:58.648813 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e4c84d5-ce7d-4b40-9b7c-980bc6b16c90" containerName="cinder-api-log" Oct 05 21:13:58 crc kubenswrapper[4754]: I1005 21:13:58.648824 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e4c84d5-ce7d-4b40-9b7c-980bc6b16c90" containerName="cinder-api" Oct 05 21:13:58 crc kubenswrapper[4754]: I1005 21:13:58.649718 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 05 21:13:58 crc kubenswrapper[4754]: I1005 21:13:58.659689 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 05 21:13:58 crc kubenswrapper[4754]: I1005 21:13:58.690420 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Oct 05 21:13:58 crc kubenswrapper[4754]: I1005 21:13:58.697767 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Oct 05 21:13:58 crc kubenswrapper[4754]: I1005 21:13:58.766336 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/b54ee63c-a198-4855-a04c-fc753939dbc0-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"b54ee63c-a198-4855-a04c-fc753939dbc0\") " pod="openstack/kube-state-metrics-0" Oct 05 21:13:58 crc kubenswrapper[4754]: I1005 21:13:58.766458 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b54ee63c-a198-4855-a04c-fc753939dbc0-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"b54ee63c-a198-4855-a04c-fc753939dbc0\") " pod="openstack/kube-state-metrics-0" Oct 05 21:13:58 crc kubenswrapper[4754]: I1005 21:13:58.766622 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6rqjl\" (UniqueName: \"kubernetes.io/projected/b54ee63c-a198-4855-a04c-fc753939dbc0-kube-api-access-6rqjl\") pod \"kube-state-metrics-0\" (UID: \"b54ee63c-a198-4855-a04c-fc753939dbc0\") " pod="openstack/kube-state-metrics-0" Oct 05 21:13:58 crc kubenswrapper[4754]: I1005 21:13:58.766645 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/b54ee63c-a198-4855-a04c-fc753939dbc0-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"b54ee63c-a198-4855-a04c-fc753939dbc0\") " pod="openstack/kube-state-metrics-0" Oct 05 21:13:58 crc kubenswrapper[4754]: I1005 21:13:58.814027 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0e4c84d5-ce7d-4b40-9b7c-980bc6b16c90-config-data" (OuterVolumeSpecName: "config-data") pod "0e4c84d5-ce7d-4b40-9b7c-980bc6b16c90" (UID: "0e4c84d5-ce7d-4b40-9b7c-980bc6b16c90"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:13:58 crc kubenswrapper[4754]: I1005 21:13:58.871362 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/b54ee63c-a198-4855-a04c-fc753939dbc0-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"b54ee63c-a198-4855-a04c-fc753939dbc0\") " pod="openstack/kube-state-metrics-0" Oct 05 21:13:58 crc kubenswrapper[4754]: I1005 21:13:58.871932 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b54ee63c-a198-4855-a04c-fc753939dbc0-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"b54ee63c-a198-4855-a04c-fc753939dbc0\") " pod="openstack/kube-state-metrics-0" Oct 05 21:13:58 crc kubenswrapper[4754]: I1005 21:13:58.872018 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6rqjl\" (UniqueName: \"kubernetes.io/projected/b54ee63c-a198-4855-a04c-fc753939dbc0-kube-api-access-6rqjl\") pod \"kube-state-metrics-0\" (UID: \"b54ee63c-a198-4855-a04c-fc753939dbc0\") " pod="openstack/kube-state-metrics-0" Oct 05 21:13:58 crc kubenswrapper[4754]: I1005 21:13:58.872041 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/b54ee63c-a198-4855-a04c-fc753939dbc0-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"b54ee63c-a198-4855-a04c-fc753939dbc0\") " pod="openstack/kube-state-metrics-0" Oct 05 21:13:58 crc kubenswrapper[4754]: I1005 21:13:58.872145 4754 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e4c84d5-ce7d-4b40-9b7c-980bc6b16c90-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 21:13:58 crc kubenswrapper[4754]: I1005 21:13:58.879288 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b54ee63c-a198-4855-a04c-fc753939dbc0-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"b54ee63c-a198-4855-a04c-fc753939dbc0\") " pod="openstack/kube-state-metrics-0" Oct 05 21:13:58 crc kubenswrapper[4754]: I1005 21:13:58.890947 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/b54ee63c-a198-4855-a04c-fc753939dbc0-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"b54ee63c-a198-4855-a04c-fc753939dbc0\") " pod="openstack/kube-state-metrics-0" Oct 05 21:13:58 crc kubenswrapper[4754]: I1005 21:13:58.895646 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="280f4e00-063d-486e-bd58-f4b246f54814" path="/var/lib/kubelet/pods/280f4e00-063d-486e-bd58-f4b246f54814/volumes" Oct 05 21:13:58 crc kubenswrapper[4754]: I1005 21:13:58.895972 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/b54ee63c-a198-4855-a04c-fc753939dbc0-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"b54ee63c-a198-4855-a04c-fc753939dbc0\") " pod="openstack/kube-state-metrics-0" Oct 05 21:13:58 crc kubenswrapper[4754]: I1005 21:13:58.909634 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6rqjl\" (UniqueName: \"kubernetes.io/projected/b54ee63c-a198-4855-a04c-fc753939dbc0-kube-api-access-6rqjl\") pod \"kube-state-metrics-0\" (UID: \"b54ee63c-a198-4855-a04c-fc753939dbc0\") " pod="openstack/kube-state-metrics-0" Oct 05 21:13:58 crc kubenswrapper[4754]: I1005 21:13:58.976468 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 05 21:13:59 crc kubenswrapper[4754]: I1005 21:13:59.218338 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c7f03a15-953e-4de8-b7b7-711cc35c8323","Type":"ContainerStarted","Data":"67ec590285159d5dd7ffbf3414c3f3154b09a08128f4538272dd9316d970ffd3"} Oct 05 21:13:59 crc kubenswrapper[4754]: I1005 21:13:59.256854 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"0e4c84d5-ce7d-4b40-9b7c-980bc6b16c90","Type":"ContainerDied","Data":"81e325c0a77688d8937756abb6992b384f46301927253d461f8c7925e40ce31e"} Oct 05 21:13:59 crc kubenswrapper[4754]: I1005 21:13:59.256926 4754 scope.go:117] "RemoveContainer" containerID="7d8e02519c00ab8dab67118f5aed2e8b655977781ccf0af4ff4939c450a39aef" Oct 05 21:13:59 crc kubenswrapper[4754]: I1005 21:13:59.256944 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 05 21:13:59 crc kubenswrapper[4754]: I1005 21:13:59.339787 4754 scope.go:117] "RemoveContainer" containerID="f0c9444516b4ed2de97be88fcc993a6d95e861794de12525343dfdf8876fd1fb" Oct 05 21:13:59 crc kubenswrapper[4754]: I1005 21:13:59.372538 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 05 21:13:59 crc kubenswrapper[4754]: I1005 21:13:59.395846 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Oct 05 21:13:59 crc kubenswrapper[4754]: I1005 21:13:59.412567 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Oct 05 21:13:59 crc kubenswrapper[4754]: I1005 21:13:59.414575 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 05 21:13:59 crc kubenswrapper[4754]: I1005 21:13:59.422433 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 05 21:13:59 crc kubenswrapper[4754]: I1005 21:13:59.423347 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Oct 05 21:13:59 crc kubenswrapper[4754]: I1005 21:13:59.423538 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Oct 05 21:13:59 crc kubenswrapper[4754]: I1005 21:13:59.423698 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Oct 05 21:13:59 crc kubenswrapper[4754]: I1005 21:13:59.493935 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/90277df3-c655-423b-b6a6-8f3142ab046e-config-data-custom\") pod \"cinder-api-0\" (UID: \"90277df3-c655-423b-b6a6-8f3142ab046e\") " pod="openstack/cinder-api-0" Oct 05 21:13:59 crc kubenswrapper[4754]: I1005 21:13:59.493973 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/90277df3-c655-423b-b6a6-8f3142ab046e-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"90277df3-c655-423b-b6a6-8f3142ab046e\") " pod="openstack/cinder-api-0" Oct 05 21:13:59 crc kubenswrapper[4754]: I1005 21:13:59.493999 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kbldw\" (UniqueName: \"kubernetes.io/projected/90277df3-c655-423b-b6a6-8f3142ab046e-kube-api-access-kbldw\") pod \"cinder-api-0\" (UID: \"90277df3-c655-423b-b6a6-8f3142ab046e\") " pod="openstack/cinder-api-0" Oct 05 21:13:59 crc kubenswrapper[4754]: I1005 21:13:59.494016 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/90277df3-c655-423b-b6a6-8f3142ab046e-etc-machine-id\") pod \"cinder-api-0\" (UID: \"90277df3-c655-423b-b6a6-8f3142ab046e\") " pod="openstack/cinder-api-0" Oct 05 21:13:59 crc kubenswrapper[4754]: I1005 21:13:59.494155 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/90277df3-c655-423b-b6a6-8f3142ab046e-public-tls-certs\") pod \"cinder-api-0\" (UID: \"90277df3-c655-423b-b6a6-8f3142ab046e\") " pod="openstack/cinder-api-0" Oct 05 21:13:59 crc kubenswrapper[4754]: I1005 21:13:59.494184 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/90277df3-c655-423b-b6a6-8f3142ab046e-logs\") pod \"cinder-api-0\" (UID: \"90277df3-c655-423b-b6a6-8f3142ab046e\") " pod="openstack/cinder-api-0" Oct 05 21:13:59 crc kubenswrapper[4754]: I1005 21:13:59.494419 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/90277df3-c655-423b-b6a6-8f3142ab046e-config-data\") pod \"cinder-api-0\" (UID: \"90277df3-c655-423b-b6a6-8f3142ab046e\") " pod="openstack/cinder-api-0" Oct 05 21:13:59 crc kubenswrapper[4754]: I1005 21:13:59.494453 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/90277df3-c655-423b-b6a6-8f3142ab046e-scripts\") pod \"cinder-api-0\" (UID: \"90277df3-c655-423b-b6a6-8f3142ab046e\") " pod="openstack/cinder-api-0" Oct 05 21:13:59 crc kubenswrapper[4754]: I1005 21:13:59.494473 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/90277df3-c655-423b-b6a6-8f3142ab046e-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"90277df3-c655-423b-b6a6-8f3142ab046e\") " pod="openstack/cinder-api-0" Oct 05 21:13:59 crc kubenswrapper[4754]: I1005 21:13:59.598818 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/90277df3-c655-423b-b6a6-8f3142ab046e-public-tls-certs\") pod \"cinder-api-0\" (UID: \"90277df3-c655-423b-b6a6-8f3142ab046e\") " pod="openstack/cinder-api-0" Oct 05 21:13:59 crc kubenswrapper[4754]: I1005 21:13:59.598878 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/90277df3-c655-423b-b6a6-8f3142ab046e-logs\") pod \"cinder-api-0\" (UID: \"90277df3-c655-423b-b6a6-8f3142ab046e\") " pod="openstack/cinder-api-0" Oct 05 21:13:59 crc kubenswrapper[4754]: I1005 21:13:59.598933 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/90277df3-c655-423b-b6a6-8f3142ab046e-config-data\") pod \"cinder-api-0\" (UID: \"90277df3-c655-423b-b6a6-8f3142ab046e\") " pod="openstack/cinder-api-0" Oct 05 21:13:59 crc kubenswrapper[4754]: I1005 21:13:59.598956 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/90277df3-c655-423b-b6a6-8f3142ab046e-scripts\") pod \"cinder-api-0\" (UID: \"90277df3-c655-423b-b6a6-8f3142ab046e\") " pod="openstack/cinder-api-0" Oct 05 21:13:59 crc kubenswrapper[4754]: I1005 21:13:59.598976 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/90277df3-c655-423b-b6a6-8f3142ab046e-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"90277df3-c655-423b-b6a6-8f3142ab046e\") " pod="openstack/cinder-api-0" Oct 05 21:13:59 crc kubenswrapper[4754]: I1005 21:13:59.599059 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/90277df3-c655-423b-b6a6-8f3142ab046e-config-data-custom\") pod \"cinder-api-0\" (UID: \"90277df3-c655-423b-b6a6-8f3142ab046e\") " pod="openstack/cinder-api-0" Oct 05 21:13:59 crc kubenswrapper[4754]: I1005 21:13:59.599080 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/90277df3-c655-423b-b6a6-8f3142ab046e-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"90277df3-c655-423b-b6a6-8f3142ab046e\") " pod="openstack/cinder-api-0" Oct 05 21:13:59 crc kubenswrapper[4754]: I1005 21:13:59.599103 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kbldw\" (UniqueName: \"kubernetes.io/projected/90277df3-c655-423b-b6a6-8f3142ab046e-kube-api-access-kbldw\") pod \"cinder-api-0\" (UID: \"90277df3-c655-423b-b6a6-8f3142ab046e\") " pod="openstack/cinder-api-0" Oct 05 21:13:59 crc kubenswrapper[4754]: I1005 21:13:59.599124 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/90277df3-c655-423b-b6a6-8f3142ab046e-etc-machine-id\") pod \"cinder-api-0\" (UID: \"90277df3-c655-423b-b6a6-8f3142ab046e\") " pod="openstack/cinder-api-0" Oct 05 21:13:59 crc kubenswrapper[4754]: I1005 21:13:59.599253 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/90277df3-c655-423b-b6a6-8f3142ab046e-etc-machine-id\") pod \"cinder-api-0\" (UID: \"90277df3-c655-423b-b6a6-8f3142ab046e\") " pod="openstack/cinder-api-0" Oct 05 21:13:59 crc kubenswrapper[4754]: I1005 21:13:59.642763 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/90277df3-c655-423b-b6a6-8f3142ab046e-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"90277df3-c655-423b-b6a6-8f3142ab046e\") " pod="openstack/cinder-api-0" Oct 05 21:13:59 crc kubenswrapper[4754]: I1005 21:13:59.646209 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/90277df3-c655-423b-b6a6-8f3142ab046e-public-tls-certs\") pod \"cinder-api-0\" (UID: \"90277df3-c655-423b-b6a6-8f3142ab046e\") " pod="openstack/cinder-api-0" Oct 05 21:13:59 crc kubenswrapper[4754]: I1005 21:13:59.646935 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/90277df3-c655-423b-b6a6-8f3142ab046e-scripts\") pod \"cinder-api-0\" (UID: \"90277df3-c655-423b-b6a6-8f3142ab046e\") " pod="openstack/cinder-api-0" Oct 05 21:13:59 crc kubenswrapper[4754]: I1005 21:13:59.652247 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/90277df3-c655-423b-b6a6-8f3142ab046e-logs\") pod \"cinder-api-0\" (UID: \"90277df3-c655-423b-b6a6-8f3142ab046e\") " pod="openstack/cinder-api-0" Oct 05 21:13:59 crc kubenswrapper[4754]: I1005 21:13:59.653120 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/90277df3-c655-423b-b6a6-8f3142ab046e-config-data-custom\") pod \"cinder-api-0\" (UID: \"90277df3-c655-423b-b6a6-8f3142ab046e\") " pod="openstack/cinder-api-0" Oct 05 21:13:59 crc kubenswrapper[4754]: I1005 21:13:59.657864 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/90277df3-c655-423b-b6a6-8f3142ab046e-config-data\") pod \"cinder-api-0\" (UID: \"90277df3-c655-423b-b6a6-8f3142ab046e\") " pod="openstack/cinder-api-0" Oct 05 21:13:59 crc kubenswrapper[4754]: I1005 21:13:59.663936 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kbldw\" (UniqueName: \"kubernetes.io/projected/90277df3-c655-423b-b6a6-8f3142ab046e-kube-api-access-kbldw\") pod \"cinder-api-0\" (UID: \"90277df3-c655-423b-b6a6-8f3142ab046e\") " pod="openstack/cinder-api-0" Oct 05 21:13:59 crc kubenswrapper[4754]: I1005 21:13:59.691094 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 05 21:13:59 crc kubenswrapper[4754]: I1005 21:13:59.698910 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/90277df3-c655-423b-b6a6-8f3142ab046e-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"90277df3-c655-423b-b6a6-8f3142ab046e\") " pod="openstack/cinder-api-0" Oct 05 21:13:59 crc kubenswrapper[4754]: I1005 21:13:59.781300 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 05 21:14:00 crc kubenswrapper[4754]: I1005 21:14:00.278023 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"b54ee63c-a198-4855-a04c-fc753939dbc0","Type":"ContainerStarted","Data":"dd27ca306e9cf0c4d525b339232a51ddd762cbf3aec4a0446ed0da13e2f36825"} Oct 05 21:14:00 crc kubenswrapper[4754]: I1005 21:14:00.381141 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 05 21:14:00 crc kubenswrapper[4754]: I1005 21:14:00.758399 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-646d45cbfc-6sjtl" Oct 05 21:14:00 crc kubenswrapper[4754]: I1005 21:14:00.768975 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-646d45cbfc-6sjtl" Oct 05 21:14:00 crc kubenswrapper[4754]: I1005 21:14:00.857552 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0e4c84d5-ce7d-4b40-9b7c-980bc6b16c90" path="/var/lib/kubelet/pods/0e4c84d5-ce7d-4b40-9b7c-980bc6b16c90/volumes" Oct 05 21:14:00 crc kubenswrapper[4754]: I1005 21:14:00.858434 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 05 21:14:00 crc kubenswrapper[4754]: I1005 21:14:00.858698 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="fb20a498-c584-4d04-b3bd-fd3517d48d4f" containerName="glance-log" containerID="cri-o://133bffa9d8418e6d9fd4f21872636494b0cc27e364eb272cbd0c43c78f09f737" gracePeriod=30 Oct 05 21:14:00 crc kubenswrapper[4754]: I1005 21:14:00.859155 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="fb20a498-c584-4d04-b3bd-fd3517d48d4f" containerName="glance-httpd" containerID="cri-o://58ebd0ada1f6bb5f01842c688aacdea75887cf282cfdb70b776799a10eb20507" gracePeriod=30 Oct 05 21:14:01 crc kubenswrapper[4754]: I1005 21:14:01.318574 4754 generic.go:334] "Generic (PLEG): container finished" podID="fb20a498-c584-4d04-b3bd-fd3517d48d4f" containerID="133bffa9d8418e6d9fd4f21872636494b0cc27e364eb272cbd0c43c78f09f737" exitCode=143 Oct 05 21:14:01 crc kubenswrapper[4754]: I1005 21:14:01.318704 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"fb20a498-c584-4d04-b3bd-fd3517d48d4f","Type":"ContainerDied","Data":"133bffa9d8418e6d9fd4f21872636494b0cc27e364eb272cbd0c43c78f09f737"} Oct 05 21:14:01 crc kubenswrapper[4754]: I1005 21:14:01.336218 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"90277df3-c655-423b-b6a6-8f3142ab046e","Type":"ContainerStarted","Data":"653a65a0967e1e472f21c7d1fccaf4574dc6d78d39c62bdd96ea0302097c6896"} Oct 05 21:14:01 crc kubenswrapper[4754]: I1005 21:14:01.348424 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"b54ee63c-a198-4855-a04c-fc753939dbc0","Type":"ContainerStarted","Data":"088b4ce276ddb01a5164a086440e904dca11e9fc1a6f24c3a7229fd3860e276c"} Oct 05 21:14:01 crc kubenswrapper[4754]: I1005 21:14:01.348514 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Oct 05 21:14:01 crc kubenswrapper[4754]: I1005 21:14:01.397826 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.828892176 podStartE2EDuration="3.397795266s" podCreationTimestamp="2025-10-05 21:13:58 +0000 UTC" firstStartedPulling="2025-10-05 21:13:59.699542264 +0000 UTC m=+1163.603660974" lastFinishedPulling="2025-10-05 21:14:00.268445354 +0000 UTC m=+1164.172564064" observedRunningTime="2025-10-05 21:14:01.371181041 +0000 UTC m=+1165.275299751" watchObservedRunningTime="2025-10-05 21:14:01.397795266 +0000 UTC m=+1165.301913966" Oct 05 21:14:01 crc kubenswrapper[4754]: I1005 21:14:01.755981 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-66d84555d4-xzsp4" Oct 05 21:14:01 crc kubenswrapper[4754]: I1005 21:14:01.794596 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-66d84555d4-xzsp4" Oct 05 21:14:02 crc kubenswrapper[4754]: I1005 21:14:02.357533 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"90277df3-c655-423b-b6a6-8f3142ab046e","Type":"ContainerStarted","Data":"dd066cb13e13f88046fe054acb220e786377beb883833f8c3fd8fc0b440adf41"} Oct 05 21:14:02 crc kubenswrapper[4754]: I1005 21:14:02.361601 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c7f03a15-953e-4de8-b7b7-711cc35c8323","Type":"ContainerStarted","Data":"e6ac7457d8a80fcb4fc42b1237603228691d5ff6f9672e3467bd4592dc1d0478"} Oct 05 21:14:02 crc kubenswrapper[4754]: I1005 21:14:02.362014 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 05 21:14:02 crc kubenswrapper[4754]: I1005 21:14:02.419688 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.447302135 podStartE2EDuration="10.419659852s" podCreationTimestamp="2025-10-05 21:13:52 +0000 UTC" firstStartedPulling="2025-10-05 21:13:54.104402136 +0000 UTC m=+1158.008520846" lastFinishedPulling="2025-10-05 21:14:01.076759853 +0000 UTC m=+1164.980878563" observedRunningTime="2025-10-05 21:14:02.408084701 +0000 UTC m=+1166.312203411" watchObservedRunningTime="2025-10-05 21:14:02.419659852 +0000 UTC m=+1166.323778562" Oct 05 21:14:03 crc kubenswrapper[4754]: I1005 21:14:03.374595 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"90277df3-c655-423b-b6a6-8f3142ab046e","Type":"ContainerStarted","Data":"2da3dd302fc8a75ef0c382fd207e35a41156a456ca17c66d45f36e5a38b17314"} Oct 05 21:14:03 crc kubenswrapper[4754]: I1005 21:14:03.374897 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Oct 05 21:14:03 crc kubenswrapper[4754]: I1005 21:14:03.391860 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=4.391843034 podStartE2EDuration="4.391843034s" podCreationTimestamp="2025-10-05 21:13:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 21:14:03.390267292 +0000 UTC m=+1167.294386002" watchObservedRunningTime="2025-10-05 21:14:03.391843034 +0000 UTC m=+1167.295961734" Oct 05 21:14:03 crc kubenswrapper[4754]: I1005 21:14:03.499777 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Oct 05 21:14:04 crc kubenswrapper[4754]: I1005 21:14:04.468082 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 05 21:14:04 crc kubenswrapper[4754]: I1005 21:14:04.922340 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-x2r84"] Oct 05 21:14:04 crc kubenswrapper[4754]: I1005 21:14:04.923703 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-x2r84" Oct 05 21:14:04 crc kubenswrapper[4754]: I1005 21:14:04.942416 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-x2r84"] Oct 05 21:14:04 crc kubenswrapper[4754]: I1005 21:14:04.995209 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zlxfq\" (UniqueName: \"kubernetes.io/projected/24624447-d035-4476-b248-b7776bb261cd-kube-api-access-zlxfq\") pod \"nova-api-db-create-x2r84\" (UID: \"24624447-d035-4476-b248-b7776bb261cd\") " pod="openstack/nova-api-db-create-x2r84" Oct 05 21:14:05 crc kubenswrapper[4754]: I1005 21:14:05.099631 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zlxfq\" (UniqueName: \"kubernetes.io/projected/24624447-d035-4476-b248-b7776bb261cd-kube-api-access-zlxfq\") pod \"nova-api-db-create-x2r84\" (UID: \"24624447-d035-4476-b248-b7776bb261cd\") " pod="openstack/nova-api-db-create-x2r84" Oct 05 21:14:05 crc kubenswrapper[4754]: I1005 21:14:05.126571 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zlxfq\" (UniqueName: \"kubernetes.io/projected/24624447-d035-4476-b248-b7776bb261cd-kube-api-access-zlxfq\") pod \"nova-api-db-create-x2r84\" (UID: \"24624447-d035-4476-b248-b7776bb261cd\") " pod="openstack/nova-api-db-create-x2r84" Oct 05 21:14:05 crc kubenswrapper[4754]: I1005 21:14:05.202237 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-tfbkl"] Oct 05 21:14:05 crc kubenswrapper[4754]: I1005 21:14:05.203472 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-tfbkl" Oct 05 21:14:05 crc kubenswrapper[4754]: I1005 21:14:05.212008 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-tfbkl"] Oct 05 21:14:05 crc kubenswrapper[4754]: I1005 21:14:05.217207 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 05 21:14:05 crc kubenswrapper[4754]: I1005 21:14:05.244759 4754 patch_prober.go:28] interesting pod/machine-config-daemon-b2h9k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 21:14:05 crc kubenswrapper[4754]: I1005 21:14:05.244825 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 21:14:05 crc kubenswrapper[4754]: I1005 21:14:05.244883 4754 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" Oct 05 21:14:05 crc kubenswrapper[4754]: I1005 21:14:05.245793 4754 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"03710aff934f922f0a7167007c25b57fd90897b7f134dd149dd0d1c75cb91da2"} pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 05 21:14:05 crc kubenswrapper[4754]: I1005 21:14:05.245865 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" containerName="machine-config-daemon" containerID="cri-o://03710aff934f922f0a7167007c25b57fd90897b7f134dd149dd0d1c75cb91da2" gracePeriod=600 Oct 05 21:14:05 crc kubenswrapper[4754]: I1005 21:14:05.300943 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-x2r84" Oct 05 21:14:05 crc kubenswrapper[4754]: I1005 21:14:05.311770 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"fb20a498-c584-4d04-b3bd-fd3517d48d4f\" (UID: \"fb20a498-c584-4d04-b3bd-fd3517d48d4f\") " Oct 05 21:14:05 crc kubenswrapper[4754]: I1005 21:14:05.311831 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s2dpj\" (UniqueName: \"kubernetes.io/projected/fb20a498-c584-4d04-b3bd-fd3517d48d4f-kube-api-access-s2dpj\") pod \"fb20a498-c584-4d04-b3bd-fd3517d48d4f\" (UID: \"fb20a498-c584-4d04-b3bd-fd3517d48d4f\") " Oct 05 21:14:05 crc kubenswrapper[4754]: I1005 21:14:05.311932 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb20a498-c584-4d04-b3bd-fd3517d48d4f-combined-ca-bundle\") pod \"fb20a498-c584-4d04-b3bd-fd3517d48d4f\" (UID: \"fb20a498-c584-4d04-b3bd-fd3517d48d4f\") " Oct 05 21:14:05 crc kubenswrapper[4754]: I1005 21:14:05.311990 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fb20a498-c584-4d04-b3bd-fd3517d48d4f-scripts\") pod \"fb20a498-c584-4d04-b3bd-fd3517d48d4f\" (UID: \"fb20a498-c584-4d04-b3bd-fd3517d48d4f\") " Oct 05 21:14:05 crc kubenswrapper[4754]: I1005 21:14:05.312017 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fb20a498-c584-4d04-b3bd-fd3517d48d4f-logs\") pod \"fb20a498-c584-4d04-b3bd-fd3517d48d4f\" (UID: \"fb20a498-c584-4d04-b3bd-fd3517d48d4f\") " Oct 05 21:14:05 crc kubenswrapper[4754]: I1005 21:14:05.312079 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/fb20a498-c584-4d04-b3bd-fd3517d48d4f-httpd-run\") pod \"fb20a498-c584-4d04-b3bd-fd3517d48d4f\" (UID: \"fb20a498-c584-4d04-b3bd-fd3517d48d4f\") " Oct 05 21:14:05 crc kubenswrapper[4754]: I1005 21:14:05.312118 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb20a498-c584-4d04-b3bd-fd3517d48d4f-config-data\") pod \"fb20a498-c584-4d04-b3bd-fd3517d48d4f\" (UID: \"fb20a498-c584-4d04-b3bd-fd3517d48d4f\") " Oct 05 21:14:05 crc kubenswrapper[4754]: I1005 21:14:05.312182 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb20a498-c584-4d04-b3bd-fd3517d48d4f-internal-tls-certs\") pod \"fb20a498-c584-4d04-b3bd-fd3517d48d4f\" (UID: \"fb20a498-c584-4d04-b3bd-fd3517d48d4f\") " Oct 05 21:14:05 crc kubenswrapper[4754]: I1005 21:14:05.312535 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bmvcz\" (UniqueName: \"kubernetes.io/projected/339a3c5a-8feb-4381-ae02-dcff29fc61fc-kube-api-access-bmvcz\") pod \"nova-cell0-db-create-tfbkl\" (UID: \"339a3c5a-8feb-4381-ae02-dcff29fc61fc\") " pod="openstack/nova-cell0-db-create-tfbkl" Oct 05 21:14:05 crc kubenswrapper[4754]: I1005 21:14:05.317754 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "glance") pod "fb20a498-c584-4d04-b3bd-fd3517d48d4f" (UID: "fb20a498-c584-4d04-b3bd-fd3517d48d4f"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 05 21:14:05 crc kubenswrapper[4754]: I1005 21:14:05.319095 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fb20a498-c584-4d04-b3bd-fd3517d48d4f-kube-api-access-s2dpj" (OuterVolumeSpecName: "kube-api-access-s2dpj") pod "fb20a498-c584-4d04-b3bd-fd3517d48d4f" (UID: "fb20a498-c584-4d04-b3bd-fd3517d48d4f"). InnerVolumeSpecName "kube-api-access-s2dpj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:14:05 crc kubenswrapper[4754]: I1005 21:14:05.319706 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fb20a498-c584-4d04-b3bd-fd3517d48d4f-logs" (OuterVolumeSpecName: "logs") pod "fb20a498-c584-4d04-b3bd-fd3517d48d4f" (UID: "fb20a498-c584-4d04-b3bd-fd3517d48d4f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 21:14:05 crc kubenswrapper[4754]: I1005 21:14:05.337798 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fb20a498-c584-4d04-b3bd-fd3517d48d4f-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "fb20a498-c584-4d04-b3bd-fd3517d48d4f" (UID: "fb20a498-c584-4d04-b3bd-fd3517d48d4f"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 21:14:05 crc kubenswrapper[4754]: I1005 21:14:05.344698 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb20a498-c584-4d04-b3bd-fd3517d48d4f-scripts" (OuterVolumeSpecName: "scripts") pod "fb20a498-c584-4d04-b3bd-fd3517d48d4f" (UID: "fb20a498-c584-4d04-b3bd-fd3517d48d4f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:14:05 crc kubenswrapper[4754]: I1005 21:14:05.345771 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-cwmnq"] Oct 05 21:14:05 crc kubenswrapper[4754]: E1005 21:14:05.346264 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb20a498-c584-4d04-b3bd-fd3517d48d4f" containerName="glance-log" Oct 05 21:14:05 crc kubenswrapper[4754]: I1005 21:14:05.346283 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb20a498-c584-4d04-b3bd-fd3517d48d4f" containerName="glance-log" Oct 05 21:14:05 crc kubenswrapper[4754]: E1005 21:14:05.346318 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb20a498-c584-4d04-b3bd-fd3517d48d4f" containerName="glance-httpd" Oct 05 21:14:05 crc kubenswrapper[4754]: I1005 21:14:05.346324 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb20a498-c584-4d04-b3bd-fd3517d48d4f" containerName="glance-httpd" Oct 05 21:14:05 crc kubenswrapper[4754]: I1005 21:14:05.346511 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb20a498-c584-4d04-b3bd-fd3517d48d4f" containerName="glance-log" Oct 05 21:14:05 crc kubenswrapper[4754]: I1005 21:14:05.346542 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb20a498-c584-4d04-b3bd-fd3517d48d4f" containerName="glance-httpd" Oct 05 21:14:05 crc kubenswrapper[4754]: I1005 21:14:05.370893 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-cwmnq" Oct 05 21:14:05 crc kubenswrapper[4754]: I1005 21:14:05.384190 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-cwmnq"] Oct 05 21:14:05 crc kubenswrapper[4754]: I1005 21:14:05.417798 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb20a498-c584-4d04-b3bd-fd3517d48d4f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fb20a498-c584-4d04-b3bd-fd3517d48d4f" (UID: "fb20a498-c584-4d04-b3bd-fd3517d48d4f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:14:05 crc kubenswrapper[4754]: I1005 21:14:05.431685 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2gj25\" (UniqueName: \"kubernetes.io/projected/5e45337f-b84d-473e-8530-942a17eb3c15-kube-api-access-2gj25\") pod \"nova-cell1-db-create-cwmnq\" (UID: \"5e45337f-b84d-473e-8530-942a17eb3c15\") " pod="openstack/nova-cell1-db-create-cwmnq" Oct 05 21:14:05 crc kubenswrapper[4754]: I1005 21:14:05.431859 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bmvcz\" (UniqueName: \"kubernetes.io/projected/339a3c5a-8feb-4381-ae02-dcff29fc61fc-kube-api-access-bmvcz\") pod \"nova-cell0-db-create-tfbkl\" (UID: \"339a3c5a-8feb-4381-ae02-dcff29fc61fc\") " pod="openstack/nova-cell0-db-create-tfbkl" Oct 05 21:14:05 crc kubenswrapper[4754]: I1005 21:14:05.432214 4754 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fb20a498-c584-4d04-b3bd-fd3517d48d4f-logs\") on node \"crc\" DevicePath \"\"" Oct 05 21:14:05 crc kubenswrapper[4754]: I1005 21:14:05.432236 4754 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/fb20a498-c584-4d04-b3bd-fd3517d48d4f-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 05 21:14:05 crc kubenswrapper[4754]: I1005 21:14:05.432259 4754 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Oct 05 21:14:05 crc kubenswrapper[4754]: I1005 21:14:05.432273 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s2dpj\" (UniqueName: \"kubernetes.io/projected/fb20a498-c584-4d04-b3bd-fd3517d48d4f-kube-api-access-s2dpj\") on node \"crc\" DevicePath \"\"" Oct 05 21:14:05 crc kubenswrapper[4754]: I1005 21:14:05.432283 4754 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb20a498-c584-4d04-b3bd-fd3517d48d4f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 21:14:05 crc kubenswrapper[4754]: I1005 21:14:05.432293 4754 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fb20a498-c584-4d04-b3bd-fd3517d48d4f-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 21:14:05 crc kubenswrapper[4754]: I1005 21:14:05.492771 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bmvcz\" (UniqueName: \"kubernetes.io/projected/339a3c5a-8feb-4381-ae02-dcff29fc61fc-kube-api-access-bmvcz\") pod \"nova-cell0-db-create-tfbkl\" (UID: \"339a3c5a-8feb-4381-ae02-dcff29fc61fc\") " pod="openstack/nova-cell0-db-create-tfbkl" Oct 05 21:14:05 crc kubenswrapper[4754]: I1005 21:14:05.493271 4754 generic.go:334] "Generic (PLEG): container finished" podID="fb20a498-c584-4d04-b3bd-fd3517d48d4f" containerID="58ebd0ada1f6bb5f01842c688aacdea75887cf282cfdb70b776799a10eb20507" exitCode=0 Oct 05 21:14:05 crc kubenswrapper[4754]: I1005 21:14:05.493453 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 05 21:14:05 crc kubenswrapper[4754]: I1005 21:14:05.493719 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"fb20a498-c584-4d04-b3bd-fd3517d48d4f","Type":"ContainerDied","Data":"58ebd0ada1f6bb5f01842c688aacdea75887cf282cfdb70b776799a10eb20507"} Oct 05 21:14:05 crc kubenswrapper[4754]: I1005 21:14:05.493779 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"fb20a498-c584-4d04-b3bd-fd3517d48d4f","Type":"ContainerDied","Data":"443c8126464536f5f38add9cc7d2e1e4ac37674aee47c1bcd6bfba501063bdd5"} Oct 05 21:14:05 crc kubenswrapper[4754]: I1005 21:14:05.493815 4754 scope.go:117] "RemoveContainer" containerID="58ebd0ada1f6bb5f01842c688aacdea75887cf282cfdb70b776799a10eb20507" Oct 05 21:14:05 crc kubenswrapper[4754]: I1005 21:14:05.513161 4754 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Oct 05 21:14:05 crc kubenswrapper[4754]: I1005 21:14:05.528259 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb20a498-c584-4d04-b3bd-fd3517d48d4f-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "fb20a498-c584-4d04-b3bd-fd3517d48d4f" (UID: "fb20a498-c584-4d04-b3bd-fd3517d48d4f"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:14:05 crc kubenswrapper[4754]: I1005 21:14:05.540104 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2gj25\" (UniqueName: \"kubernetes.io/projected/5e45337f-b84d-473e-8530-942a17eb3c15-kube-api-access-2gj25\") pod \"nova-cell1-db-create-cwmnq\" (UID: \"5e45337f-b84d-473e-8530-942a17eb3c15\") " pod="openstack/nova-cell1-db-create-cwmnq" Oct 05 21:14:05 crc kubenswrapper[4754]: I1005 21:14:05.540436 4754 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb20a498-c584-4d04-b3bd-fd3517d48d4f-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 05 21:14:05 crc kubenswrapper[4754]: I1005 21:14:05.540456 4754 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Oct 05 21:14:05 crc kubenswrapper[4754]: I1005 21:14:05.542054 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-tfbkl" Oct 05 21:14:05 crc kubenswrapper[4754]: I1005 21:14:05.564967 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2gj25\" (UniqueName: \"kubernetes.io/projected/5e45337f-b84d-473e-8530-942a17eb3c15-kube-api-access-2gj25\") pod \"nova-cell1-db-create-cwmnq\" (UID: \"5e45337f-b84d-473e-8530-942a17eb3c15\") " pod="openstack/nova-cell1-db-create-cwmnq" Oct 05 21:14:05 crc kubenswrapper[4754]: I1005 21:14:05.582050 4754 generic.go:334] "Generic (PLEG): container finished" podID="be95a413-6a59-45b4-84b7-b43cae694a26" containerID="03710aff934f922f0a7167007c25b57fd90897b7f134dd149dd0d1c75cb91da2" exitCode=0 Oct 05 21:14:05 crc kubenswrapper[4754]: I1005 21:14:05.582685 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c7f03a15-953e-4de8-b7b7-711cc35c8323" containerName="ceilometer-central-agent" containerID="cri-o://c81e9c627545c8ae7fec11f2768aba757f9b5f96a53410f2bbb3e6958d352382" gracePeriod=30 Oct 05 21:14:05 crc kubenswrapper[4754]: I1005 21:14:05.583480 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" event={"ID":"be95a413-6a59-45b4-84b7-b43cae694a26","Type":"ContainerDied","Data":"03710aff934f922f0a7167007c25b57fd90897b7f134dd149dd0d1c75cb91da2"} Oct 05 21:14:05 crc kubenswrapper[4754]: I1005 21:14:05.584440 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c7f03a15-953e-4de8-b7b7-711cc35c8323" containerName="sg-core" containerID="cri-o://67ec590285159d5dd7ffbf3414c3f3154b09a08128f4538272dd9316d970ffd3" gracePeriod=30 Oct 05 21:14:05 crc kubenswrapper[4754]: I1005 21:14:05.584796 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c7f03a15-953e-4de8-b7b7-711cc35c8323" containerName="proxy-httpd" containerID="cri-o://e6ac7457d8a80fcb4fc42b1237603228691d5ff6f9672e3467bd4592dc1d0478" gracePeriod=30 Oct 05 21:14:05 crc kubenswrapper[4754]: I1005 21:14:05.584983 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c7f03a15-953e-4de8-b7b7-711cc35c8323" containerName="ceilometer-notification-agent" containerID="cri-o://41c9b85f7408628829308d18d3aeee6968858f7de70656a6fe724f64f802b04f" gracePeriod=30 Oct 05 21:14:05 crc kubenswrapper[4754]: I1005 21:14:05.623516 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-68c6b67864-9msm5" podUID="27cd6be5-f5eb-4158-bdd1-37f4df3d530d" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.148:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.148:8443: connect: connection refused" Oct 05 21:14:05 crc kubenswrapper[4754]: I1005 21:14:05.713798 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb20a498-c584-4d04-b3bd-fd3517d48d4f-config-data" (OuterVolumeSpecName: "config-data") pod "fb20a498-c584-4d04-b3bd-fd3517d48d4f" (UID: "fb20a498-c584-4d04-b3bd-fd3517d48d4f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:14:05 crc kubenswrapper[4754]: I1005 21:14:05.738233 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-574454d6cb-vbnk6" podUID="39d1aac3-14a1-45bf-9a9d-ba4cdb0afcb1" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.149:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.149:8443: connect: connection refused" Oct 05 21:14:05 crc kubenswrapper[4754]: I1005 21:14:05.740100 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-cwmnq" Oct 05 21:14:05 crc kubenswrapper[4754]: I1005 21:14:05.750312 4754 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb20a498-c584-4d04-b3bd-fd3517d48d4f-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 21:14:05 crc kubenswrapper[4754]: I1005 21:14:05.757981 4754 scope.go:117] "RemoveContainer" containerID="133bffa9d8418e6d9fd4f21872636494b0cc27e364eb272cbd0c43c78f09f737" Oct 05 21:14:05 crc kubenswrapper[4754]: I1005 21:14:05.838256 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 05 21:14:05 crc kubenswrapper[4754]: I1005 21:14:05.851037 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 05 21:14:05 crc kubenswrapper[4754]: I1005 21:14:05.885059 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 05 21:14:05 crc kubenswrapper[4754]: I1005 21:14:05.888102 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 05 21:14:05 crc kubenswrapper[4754]: I1005 21:14:05.896670 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 05 21:14:05 crc kubenswrapper[4754]: I1005 21:14:05.900308 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Oct 05 21:14:05 crc kubenswrapper[4754]: I1005 21:14:05.925762 4754 scope.go:117] "RemoveContainer" containerID="58ebd0ada1f6bb5f01842c688aacdea75887cf282cfdb70b776799a10eb20507" Oct 05 21:14:05 crc kubenswrapper[4754]: E1005 21:14:05.932051 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"58ebd0ada1f6bb5f01842c688aacdea75887cf282cfdb70b776799a10eb20507\": container with ID starting with 58ebd0ada1f6bb5f01842c688aacdea75887cf282cfdb70b776799a10eb20507 not found: ID does not exist" containerID="58ebd0ada1f6bb5f01842c688aacdea75887cf282cfdb70b776799a10eb20507" Oct 05 21:14:05 crc kubenswrapper[4754]: I1005 21:14:05.932083 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"58ebd0ada1f6bb5f01842c688aacdea75887cf282cfdb70b776799a10eb20507"} err="failed to get container status \"58ebd0ada1f6bb5f01842c688aacdea75887cf282cfdb70b776799a10eb20507\": rpc error: code = NotFound desc = could not find container \"58ebd0ada1f6bb5f01842c688aacdea75887cf282cfdb70b776799a10eb20507\": container with ID starting with 58ebd0ada1f6bb5f01842c688aacdea75887cf282cfdb70b776799a10eb20507 not found: ID does not exist" Oct 05 21:14:05 crc kubenswrapper[4754]: I1005 21:14:05.932118 4754 scope.go:117] "RemoveContainer" containerID="133bffa9d8418e6d9fd4f21872636494b0cc27e364eb272cbd0c43c78f09f737" Oct 05 21:14:05 crc kubenswrapper[4754]: I1005 21:14:05.934259 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 05 21:14:05 crc kubenswrapper[4754]: E1005 21:14:05.935242 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"133bffa9d8418e6d9fd4f21872636494b0cc27e364eb272cbd0c43c78f09f737\": container with ID starting with 133bffa9d8418e6d9fd4f21872636494b0cc27e364eb272cbd0c43c78f09f737 not found: ID does not exist" containerID="133bffa9d8418e6d9fd4f21872636494b0cc27e364eb272cbd0c43c78f09f737" Oct 05 21:14:05 crc kubenswrapper[4754]: I1005 21:14:05.935270 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"133bffa9d8418e6d9fd4f21872636494b0cc27e364eb272cbd0c43c78f09f737"} err="failed to get container status \"133bffa9d8418e6d9fd4f21872636494b0cc27e364eb272cbd0c43c78f09f737\": rpc error: code = NotFound desc = could not find container \"133bffa9d8418e6d9fd4f21872636494b0cc27e364eb272cbd0c43c78f09f737\": container with ID starting with 133bffa9d8418e6d9fd4f21872636494b0cc27e364eb272cbd0c43c78f09f737 not found: ID does not exist" Oct 05 21:14:05 crc kubenswrapper[4754]: I1005 21:14:05.935292 4754 scope.go:117] "RemoveContainer" containerID="204a9d4b26fceb896f0714f752c6bee1970d2349381ac50fae1a7322b50bafde" Oct 05 21:14:05 crc kubenswrapper[4754]: I1005 21:14:05.953676 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b378ce06-065f-4d9e-80d2-d98d834a2556-config-data\") pod \"glance-default-internal-api-0\" (UID: \"b378ce06-065f-4d9e-80d2-d98d834a2556\") " pod="openstack/glance-default-internal-api-0" Oct 05 21:14:05 crc kubenswrapper[4754]: I1005 21:14:05.953754 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-52cxv\" (UniqueName: \"kubernetes.io/projected/b378ce06-065f-4d9e-80d2-d98d834a2556-kube-api-access-52cxv\") pod \"glance-default-internal-api-0\" (UID: \"b378ce06-065f-4d9e-80d2-d98d834a2556\") " pod="openstack/glance-default-internal-api-0" Oct 05 21:14:05 crc kubenswrapper[4754]: I1005 21:14:05.953920 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b378ce06-065f-4d9e-80d2-d98d834a2556-logs\") pod \"glance-default-internal-api-0\" (UID: \"b378ce06-065f-4d9e-80d2-d98d834a2556\") " pod="openstack/glance-default-internal-api-0" Oct 05 21:14:05 crc kubenswrapper[4754]: I1005 21:14:05.954004 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b378ce06-065f-4d9e-80d2-d98d834a2556-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"b378ce06-065f-4d9e-80d2-d98d834a2556\") " pod="openstack/glance-default-internal-api-0" Oct 05 21:14:05 crc kubenswrapper[4754]: I1005 21:14:05.954197 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b378ce06-065f-4d9e-80d2-d98d834a2556-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"b378ce06-065f-4d9e-80d2-d98d834a2556\") " pod="openstack/glance-default-internal-api-0" Oct 05 21:14:05 crc kubenswrapper[4754]: I1005 21:14:05.954257 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b378ce06-065f-4d9e-80d2-d98d834a2556-scripts\") pod \"glance-default-internal-api-0\" (UID: \"b378ce06-065f-4d9e-80d2-d98d834a2556\") " pod="openstack/glance-default-internal-api-0" Oct 05 21:14:05 crc kubenswrapper[4754]: I1005 21:14:05.954389 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-internal-api-0\" (UID: \"b378ce06-065f-4d9e-80d2-d98d834a2556\") " pod="openstack/glance-default-internal-api-0" Oct 05 21:14:05 crc kubenswrapper[4754]: I1005 21:14:05.954434 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b378ce06-065f-4d9e-80d2-d98d834a2556-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"b378ce06-065f-4d9e-80d2-d98d834a2556\") " pod="openstack/glance-default-internal-api-0" Oct 05 21:14:06 crc kubenswrapper[4754]: I1005 21:14:06.042060 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-x2r84"] Oct 05 21:14:06 crc kubenswrapper[4754]: I1005 21:14:06.066124 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-52cxv\" (UniqueName: \"kubernetes.io/projected/b378ce06-065f-4d9e-80d2-d98d834a2556-kube-api-access-52cxv\") pod \"glance-default-internal-api-0\" (UID: \"b378ce06-065f-4d9e-80d2-d98d834a2556\") " pod="openstack/glance-default-internal-api-0" Oct 05 21:14:06 crc kubenswrapper[4754]: I1005 21:14:06.066622 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b378ce06-065f-4d9e-80d2-d98d834a2556-logs\") pod \"glance-default-internal-api-0\" (UID: \"b378ce06-065f-4d9e-80d2-d98d834a2556\") " pod="openstack/glance-default-internal-api-0" Oct 05 21:14:06 crc kubenswrapper[4754]: I1005 21:14:06.066755 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b378ce06-065f-4d9e-80d2-d98d834a2556-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"b378ce06-065f-4d9e-80d2-d98d834a2556\") " pod="openstack/glance-default-internal-api-0" Oct 05 21:14:06 crc kubenswrapper[4754]: I1005 21:14:06.066921 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b378ce06-065f-4d9e-80d2-d98d834a2556-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"b378ce06-065f-4d9e-80d2-d98d834a2556\") " pod="openstack/glance-default-internal-api-0" Oct 05 21:14:06 crc kubenswrapper[4754]: I1005 21:14:06.067038 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b378ce06-065f-4d9e-80d2-d98d834a2556-scripts\") pod \"glance-default-internal-api-0\" (UID: \"b378ce06-065f-4d9e-80d2-d98d834a2556\") " pod="openstack/glance-default-internal-api-0" Oct 05 21:14:06 crc kubenswrapper[4754]: I1005 21:14:06.067201 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-internal-api-0\" (UID: \"b378ce06-065f-4d9e-80d2-d98d834a2556\") " pod="openstack/glance-default-internal-api-0" Oct 05 21:14:06 crc kubenswrapper[4754]: I1005 21:14:06.067335 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b378ce06-065f-4d9e-80d2-d98d834a2556-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"b378ce06-065f-4d9e-80d2-d98d834a2556\") " pod="openstack/glance-default-internal-api-0" Oct 05 21:14:06 crc kubenswrapper[4754]: I1005 21:14:06.067488 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b378ce06-065f-4d9e-80d2-d98d834a2556-config-data\") pod \"glance-default-internal-api-0\" (UID: \"b378ce06-065f-4d9e-80d2-d98d834a2556\") " pod="openstack/glance-default-internal-api-0" Oct 05 21:14:06 crc kubenswrapper[4754]: I1005 21:14:06.069093 4754 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-internal-api-0\" (UID: \"b378ce06-065f-4d9e-80d2-d98d834a2556\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/glance-default-internal-api-0" Oct 05 21:14:06 crc kubenswrapper[4754]: I1005 21:14:06.069603 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b378ce06-065f-4d9e-80d2-d98d834a2556-logs\") pod \"glance-default-internal-api-0\" (UID: \"b378ce06-065f-4d9e-80d2-d98d834a2556\") " pod="openstack/glance-default-internal-api-0" Oct 05 21:14:06 crc kubenswrapper[4754]: I1005 21:14:06.069885 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b378ce06-065f-4d9e-80d2-d98d834a2556-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"b378ce06-065f-4d9e-80d2-d98d834a2556\") " pod="openstack/glance-default-internal-api-0" Oct 05 21:14:06 crc kubenswrapper[4754]: I1005 21:14:06.088593 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b378ce06-065f-4d9e-80d2-d98d834a2556-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"b378ce06-065f-4d9e-80d2-d98d834a2556\") " pod="openstack/glance-default-internal-api-0" Oct 05 21:14:06 crc kubenswrapper[4754]: I1005 21:14:06.092286 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b378ce06-065f-4d9e-80d2-d98d834a2556-config-data\") pod \"glance-default-internal-api-0\" (UID: \"b378ce06-065f-4d9e-80d2-d98d834a2556\") " pod="openstack/glance-default-internal-api-0" Oct 05 21:14:06 crc kubenswrapper[4754]: I1005 21:14:06.099478 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b378ce06-065f-4d9e-80d2-d98d834a2556-scripts\") pod \"glance-default-internal-api-0\" (UID: \"b378ce06-065f-4d9e-80d2-d98d834a2556\") " pod="openstack/glance-default-internal-api-0" Oct 05 21:14:06 crc kubenswrapper[4754]: I1005 21:14:06.109260 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b378ce06-065f-4d9e-80d2-d98d834a2556-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"b378ce06-065f-4d9e-80d2-d98d834a2556\") " pod="openstack/glance-default-internal-api-0" Oct 05 21:14:06 crc kubenswrapper[4754]: I1005 21:14:06.113397 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-52cxv\" (UniqueName: \"kubernetes.io/projected/b378ce06-065f-4d9e-80d2-d98d834a2556-kube-api-access-52cxv\") pod \"glance-default-internal-api-0\" (UID: \"b378ce06-065f-4d9e-80d2-d98d834a2556\") " pod="openstack/glance-default-internal-api-0" Oct 05 21:14:06 crc kubenswrapper[4754]: I1005 21:14:06.239246 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-internal-api-0\" (UID: \"b378ce06-065f-4d9e-80d2-d98d834a2556\") " pod="openstack/glance-default-internal-api-0" Oct 05 21:14:06 crc kubenswrapper[4754]: I1005 21:14:06.365686 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-77ccdf5dfb-zv8nn" Oct 05 21:14:06 crc kubenswrapper[4754]: I1005 21:14:06.529024 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-tfbkl"] Oct 05 21:14:06 crc kubenswrapper[4754]: I1005 21:14:06.540442 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 05 21:14:06 crc kubenswrapper[4754]: W1005 21:14:06.558131 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod339a3c5a_8feb_4381_ae02_dcff29fc61fc.slice/crio-1dc9121e123f1ccb7a47a3db6e25592128f4471a224e6a778cea8cb7951c8f69 WatchSource:0}: Error finding container 1dc9121e123f1ccb7a47a3db6e25592128f4471a224e6a778cea8cb7951c8f69: Status 404 returned error can't find the container with id 1dc9121e123f1ccb7a47a3db6e25592128f4471a224e6a778cea8cb7951c8f69 Oct 05 21:14:06 crc kubenswrapper[4754]: I1005 21:14:06.623641 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-tfbkl" event={"ID":"339a3c5a-8feb-4381-ae02-dcff29fc61fc","Type":"ContainerStarted","Data":"1dc9121e123f1ccb7a47a3db6e25592128f4471a224e6a778cea8cb7951c8f69"} Oct 05 21:14:06 crc kubenswrapper[4754]: I1005 21:14:06.627250 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" event={"ID":"be95a413-6a59-45b4-84b7-b43cae694a26","Type":"ContainerStarted","Data":"a2229b13a2697fd212215ae5d196911669bc92f36ce70525da6f97d702c807b9"} Oct 05 21:14:06 crc kubenswrapper[4754]: I1005 21:14:06.675798 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c7f03a15-953e-4de8-b7b7-711cc35c8323","Type":"ContainerDied","Data":"e6ac7457d8a80fcb4fc42b1237603228691d5ff6f9672e3467bd4592dc1d0478"} Oct 05 21:14:06 crc kubenswrapper[4754]: I1005 21:14:06.675748 4754 generic.go:334] "Generic (PLEG): container finished" podID="c7f03a15-953e-4de8-b7b7-711cc35c8323" containerID="e6ac7457d8a80fcb4fc42b1237603228691d5ff6f9672e3467bd4592dc1d0478" exitCode=0 Oct 05 21:14:06 crc kubenswrapper[4754]: I1005 21:14:06.675860 4754 generic.go:334] "Generic (PLEG): container finished" podID="c7f03a15-953e-4de8-b7b7-711cc35c8323" containerID="67ec590285159d5dd7ffbf3414c3f3154b09a08128f4538272dd9316d970ffd3" exitCode=2 Oct 05 21:14:06 crc kubenswrapper[4754]: I1005 21:14:06.675873 4754 generic.go:334] "Generic (PLEG): container finished" podID="c7f03a15-953e-4de8-b7b7-711cc35c8323" containerID="41c9b85f7408628829308d18d3aeee6968858f7de70656a6fe724f64f802b04f" exitCode=0 Oct 05 21:14:06 crc kubenswrapper[4754]: I1005 21:14:06.675922 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c7f03a15-953e-4de8-b7b7-711cc35c8323","Type":"ContainerDied","Data":"67ec590285159d5dd7ffbf3414c3f3154b09a08128f4538272dd9316d970ffd3"} Oct 05 21:14:06 crc kubenswrapper[4754]: I1005 21:14:06.675935 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c7f03a15-953e-4de8-b7b7-711cc35c8323","Type":"ContainerDied","Data":"41c9b85f7408628829308d18d3aeee6968858f7de70656a6fe724f64f802b04f"} Oct 05 21:14:06 crc kubenswrapper[4754]: I1005 21:14:06.680675 4754 generic.go:334] "Generic (PLEG): container finished" podID="24624447-d035-4476-b248-b7776bb261cd" containerID="30976ef69a1385193880865c854658ecdb6d9fbdbbcf4197ad7bb575690f26d9" exitCode=0 Oct 05 21:14:06 crc kubenswrapper[4754]: I1005 21:14:06.680848 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-x2r84" event={"ID":"24624447-d035-4476-b248-b7776bb261cd","Type":"ContainerDied","Data":"30976ef69a1385193880865c854658ecdb6d9fbdbbcf4197ad7bb575690f26d9"} Oct 05 21:14:06 crc kubenswrapper[4754]: I1005 21:14:06.682268 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-x2r84" event={"ID":"24624447-d035-4476-b248-b7776bb261cd","Type":"ContainerStarted","Data":"5ea0c5eabffa1e22d842c482e673c987ffc0561850ef9017b830d657a66ab64b"} Oct 05 21:14:06 crc kubenswrapper[4754]: I1005 21:14:06.715590 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-cwmnq"] Oct 05 21:14:06 crc kubenswrapper[4754]: I1005 21:14:06.888186 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fb20a498-c584-4d04-b3bd-fd3517d48d4f" path="/var/lib/kubelet/pods/fb20a498-c584-4d04-b3bd-fd3517d48d4f/volumes" Oct 05 21:14:07 crc kubenswrapper[4754]: I1005 21:14:07.224359 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 05 21:14:07 crc kubenswrapper[4754]: W1005 21:14:07.260199 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb378ce06_065f_4d9e_80d2_d98d834a2556.slice/crio-1fefa13323df0ae08eed8441d5a4e310b9f57f96324d3cc127b935f905118292 WatchSource:0}: Error finding container 1fefa13323df0ae08eed8441d5a4e310b9f57f96324d3cc127b935f905118292: Status 404 returned error can't find the container with id 1fefa13323df0ae08eed8441d5a4e310b9f57f96324d3cc127b935f905118292 Oct 05 21:14:07 crc kubenswrapper[4754]: I1005 21:14:07.703652 4754 generic.go:334] "Generic (PLEG): container finished" podID="339a3c5a-8feb-4381-ae02-dcff29fc61fc" containerID="5d5ac350e26ebbf73c7b124d89e020d93d0c4cc3ee33bd5619260334c213e7d1" exitCode=0 Oct 05 21:14:07 crc kubenswrapper[4754]: I1005 21:14:07.703723 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-tfbkl" event={"ID":"339a3c5a-8feb-4381-ae02-dcff29fc61fc","Type":"ContainerDied","Data":"5d5ac350e26ebbf73c7b124d89e020d93d0c4cc3ee33bd5619260334c213e7d1"} Oct 05 21:14:07 crc kubenswrapper[4754]: I1005 21:14:07.708789 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"b378ce06-065f-4d9e-80d2-d98d834a2556","Type":"ContainerStarted","Data":"1fefa13323df0ae08eed8441d5a4e310b9f57f96324d3cc127b935f905118292"} Oct 05 21:14:07 crc kubenswrapper[4754]: I1005 21:14:07.715098 4754 generic.go:334] "Generic (PLEG): container finished" podID="5e45337f-b84d-473e-8530-942a17eb3c15" containerID="2e5929c9b37c6691cf5350c3d6078cf7fb5f65a0adafa8b9331f21166686756d" exitCode=0 Oct 05 21:14:07 crc kubenswrapper[4754]: I1005 21:14:07.715318 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-cwmnq" event={"ID":"5e45337f-b84d-473e-8530-942a17eb3c15","Type":"ContainerDied","Data":"2e5929c9b37c6691cf5350c3d6078cf7fb5f65a0adafa8b9331f21166686756d"} Oct 05 21:14:07 crc kubenswrapper[4754]: I1005 21:14:07.715367 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-cwmnq" event={"ID":"5e45337f-b84d-473e-8530-942a17eb3c15","Type":"ContainerStarted","Data":"94df4dda48452cd1d39d446a2d09340c7f1183ca0e19d036343598f9e38a7c4a"} Oct 05 21:14:08 crc kubenswrapper[4754]: I1005 21:14:08.263121 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-x2r84" Oct 05 21:14:08 crc kubenswrapper[4754]: I1005 21:14:08.362087 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zlxfq\" (UniqueName: \"kubernetes.io/projected/24624447-d035-4476-b248-b7776bb261cd-kube-api-access-zlxfq\") pod \"24624447-d035-4476-b248-b7776bb261cd\" (UID: \"24624447-d035-4476-b248-b7776bb261cd\") " Oct 05 21:14:08 crc kubenswrapper[4754]: I1005 21:14:08.370824 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/24624447-d035-4476-b248-b7776bb261cd-kube-api-access-zlxfq" (OuterVolumeSpecName: "kube-api-access-zlxfq") pod "24624447-d035-4476-b248-b7776bb261cd" (UID: "24624447-d035-4476-b248-b7776bb261cd"). InnerVolumeSpecName "kube-api-access-zlxfq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:14:08 crc kubenswrapper[4754]: I1005 21:14:08.466851 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zlxfq\" (UniqueName: \"kubernetes.io/projected/24624447-d035-4476-b248-b7776bb261cd-kube-api-access-zlxfq\") on node \"crc\" DevicePath \"\"" Oct 05 21:14:08 crc kubenswrapper[4754]: I1005 21:14:08.735371 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-x2r84" event={"ID":"24624447-d035-4476-b248-b7776bb261cd","Type":"ContainerDied","Data":"5ea0c5eabffa1e22d842c482e673c987ffc0561850ef9017b830d657a66ab64b"} Oct 05 21:14:08 crc kubenswrapper[4754]: I1005 21:14:08.736772 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5ea0c5eabffa1e22d842c482e673c987ffc0561850ef9017b830d657a66ab64b" Oct 05 21:14:08 crc kubenswrapper[4754]: I1005 21:14:08.735419 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-x2r84" Oct 05 21:14:08 crc kubenswrapper[4754]: I1005 21:14:08.737738 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"b378ce06-065f-4d9e-80d2-d98d834a2556","Type":"ContainerStarted","Data":"8d55e556b4e9bcbee5036c97289fd7f4fbbcdaee6ab955b735536fc37c5bdced"} Oct 05 21:14:09 crc kubenswrapper[4754]: I1005 21:14:09.091829 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Oct 05 21:14:09 crc kubenswrapper[4754]: I1005 21:14:09.406719 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-tfbkl" Oct 05 21:14:09 crc kubenswrapper[4754]: I1005 21:14:09.528643 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bmvcz\" (UniqueName: \"kubernetes.io/projected/339a3c5a-8feb-4381-ae02-dcff29fc61fc-kube-api-access-bmvcz\") pod \"339a3c5a-8feb-4381-ae02-dcff29fc61fc\" (UID: \"339a3c5a-8feb-4381-ae02-dcff29fc61fc\") " Oct 05 21:14:09 crc kubenswrapper[4754]: I1005 21:14:09.549762 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/339a3c5a-8feb-4381-ae02-dcff29fc61fc-kube-api-access-bmvcz" (OuterVolumeSpecName: "kube-api-access-bmvcz") pod "339a3c5a-8feb-4381-ae02-dcff29fc61fc" (UID: "339a3c5a-8feb-4381-ae02-dcff29fc61fc"). InnerVolumeSpecName "kube-api-access-bmvcz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:14:09 crc kubenswrapper[4754]: I1005 21:14:09.627393 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-cwmnq" Oct 05 21:14:09 crc kubenswrapper[4754]: I1005 21:14:09.630799 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2gj25\" (UniqueName: \"kubernetes.io/projected/5e45337f-b84d-473e-8530-942a17eb3c15-kube-api-access-2gj25\") pod \"5e45337f-b84d-473e-8530-942a17eb3c15\" (UID: \"5e45337f-b84d-473e-8530-942a17eb3c15\") " Oct 05 21:14:09 crc kubenswrapper[4754]: I1005 21:14:09.631326 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bmvcz\" (UniqueName: \"kubernetes.io/projected/339a3c5a-8feb-4381-ae02-dcff29fc61fc-kube-api-access-bmvcz\") on node \"crc\" DevicePath \"\"" Oct 05 21:14:09 crc kubenswrapper[4754]: I1005 21:14:09.634358 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5e45337f-b84d-473e-8530-942a17eb3c15-kube-api-access-2gj25" (OuterVolumeSpecName: "kube-api-access-2gj25") pod "5e45337f-b84d-473e-8530-942a17eb3c15" (UID: "5e45337f-b84d-473e-8530-942a17eb3c15"). InnerVolumeSpecName "kube-api-access-2gj25". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:14:09 crc kubenswrapper[4754]: I1005 21:14:09.733158 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2gj25\" (UniqueName: \"kubernetes.io/projected/5e45337f-b84d-473e-8530-942a17eb3c15-kube-api-access-2gj25\") on node \"crc\" DevicePath \"\"" Oct 05 21:14:09 crc kubenswrapper[4754]: I1005 21:14:09.751512 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-tfbkl" event={"ID":"339a3c5a-8feb-4381-ae02-dcff29fc61fc","Type":"ContainerDied","Data":"1dc9121e123f1ccb7a47a3db6e25592128f4471a224e6a778cea8cb7951c8f69"} Oct 05 21:14:09 crc kubenswrapper[4754]: I1005 21:14:09.751560 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1dc9121e123f1ccb7a47a3db6e25592128f4471a224e6a778cea8cb7951c8f69" Oct 05 21:14:09 crc kubenswrapper[4754]: I1005 21:14:09.751626 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-tfbkl" Oct 05 21:14:09 crc kubenswrapper[4754]: I1005 21:14:09.757983 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"b378ce06-065f-4d9e-80d2-d98d834a2556","Type":"ContainerStarted","Data":"af9de5496e0ab8b041ac3a58607333acda5db6e3a68cda03c000a45996343e81"} Oct 05 21:14:09 crc kubenswrapper[4754]: I1005 21:14:09.762909 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-cwmnq" event={"ID":"5e45337f-b84d-473e-8530-942a17eb3c15","Type":"ContainerDied","Data":"94df4dda48452cd1d39d446a2d09340c7f1183ca0e19d036343598f9e38a7c4a"} Oct 05 21:14:09 crc kubenswrapper[4754]: I1005 21:14:09.762944 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="94df4dda48452cd1d39d446a2d09340c7f1183ca0e19d036343598f9e38a7c4a" Oct 05 21:14:09 crc kubenswrapper[4754]: I1005 21:14:09.762996 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-cwmnq" Oct 05 21:14:09 crc kubenswrapper[4754]: I1005 21:14:09.799119 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=4.799100514 podStartE2EDuration="4.799100514s" podCreationTimestamp="2025-10-05 21:14:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 21:14:09.797561793 +0000 UTC m=+1173.701680503" watchObservedRunningTime="2025-10-05 21:14:09.799100514 +0000 UTC m=+1173.703219224" Oct 05 21:14:10 crc kubenswrapper[4754]: E1005 21:14:10.066234 4754 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5e45337f_b84d_473e_8530_942a17eb3c15.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5e45337f_b84d_473e_8530_942a17eb3c15.slice/crio-94df4dda48452cd1d39d446a2d09340c7f1183ca0e19d036343598f9e38a7c4a\": RecentStats: unable to find data in memory cache]" Oct 05 21:14:10 crc kubenswrapper[4754]: I1005 21:14:10.465549 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 05 21:14:10 crc kubenswrapper[4754]: I1005 21:14:10.465837 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="85cf3d49-cc86-43ad-ad97-0e4476de6ee6" containerName="glance-log" containerID="cri-o://6d5bd90dbdb3906057a2a30e1be4ae1d3f097b8e2659882a66083011d71ad64d" gracePeriod=30 Oct 05 21:14:10 crc kubenswrapper[4754]: I1005 21:14:10.465996 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="85cf3d49-cc86-43ad-ad97-0e4476de6ee6" containerName="glance-httpd" containerID="cri-o://61c6c3b71ff15d6f9ac21858b8162e6919168883eb60955b3b78ab5843bf6b52" gracePeriod=30 Oct 05 21:14:10 crc kubenswrapper[4754]: I1005 21:14:10.774306 4754 generic.go:334] "Generic (PLEG): container finished" podID="85cf3d49-cc86-43ad-ad97-0e4476de6ee6" containerID="6d5bd90dbdb3906057a2a30e1be4ae1d3f097b8e2659882a66083011d71ad64d" exitCode=143 Oct 05 21:14:10 crc kubenswrapper[4754]: I1005 21:14:10.775562 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"85cf3d49-cc86-43ad-ad97-0e4476de6ee6","Type":"ContainerDied","Data":"6d5bd90dbdb3906057a2a30e1be4ae1d3f097b8e2659882a66083011d71ad64d"} Oct 05 21:14:11 crc kubenswrapper[4754]: I1005 21:14:11.142673 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-5d6f8484b5-vstqw" Oct 05 21:14:11 crc kubenswrapper[4754]: I1005 21:14:11.233162 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-77ccdf5dfb-zv8nn"] Oct 05 21:14:11 crc kubenswrapper[4754]: I1005 21:14:11.233608 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-77ccdf5dfb-zv8nn" podUID="a42df2f2-2289-4a9e-9ab4-156d3c8d2145" containerName="neutron-httpd" containerID="cri-o://daa7d97295865858fb6daa372fbbc8824aebf09867c36045c8d5ba563ee53ed9" gracePeriod=30 Oct 05 21:14:11 crc kubenswrapper[4754]: I1005 21:14:11.233819 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-77ccdf5dfb-zv8nn" podUID="a42df2f2-2289-4a9e-9ab4-156d3c8d2145" containerName="neutron-api" containerID="cri-o://28d4be6a6eb29d90be12ae24e0bb51cb8fac12201eacdd04cc51c1bec5626a1d" gracePeriod=30 Oct 05 21:14:11 crc kubenswrapper[4754]: I1005 21:14:11.851817 4754 generic.go:334] "Generic (PLEG): container finished" podID="a42df2f2-2289-4a9e-9ab4-156d3c8d2145" containerID="daa7d97295865858fb6daa372fbbc8824aebf09867c36045c8d5ba563ee53ed9" exitCode=0 Oct 05 21:14:11 crc kubenswrapper[4754]: I1005 21:14:11.851871 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-77ccdf5dfb-zv8nn" event={"ID":"a42df2f2-2289-4a9e-9ab4-156d3c8d2145","Type":"ContainerDied","Data":"daa7d97295865858fb6daa372fbbc8824aebf09867c36045c8d5ba563ee53ed9"} Oct 05 21:14:11 crc kubenswrapper[4754]: I1005 21:14:11.887045 4754 generic.go:334] "Generic (PLEG): container finished" podID="c7f03a15-953e-4de8-b7b7-711cc35c8323" containerID="c81e9c627545c8ae7fec11f2768aba757f9b5f96a53410f2bbb3e6958d352382" exitCode=0 Oct 05 21:14:11 crc kubenswrapper[4754]: I1005 21:14:11.888074 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c7f03a15-953e-4de8-b7b7-711cc35c8323","Type":"ContainerDied","Data":"c81e9c627545c8ae7fec11f2768aba757f9b5f96a53410f2bbb3e6958d352382"} Oct 05 21:14:11 crc kubenswrapper[4754]: I1005 21:14:11.888132 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c7f03a15-953e-4de8-b7b7-711cc35c8323","Type":"ContainerDied","Data":"e720f41e8cb279f7b8b91bb1549ff3dfe463fbbe226f20aeff8c426f0e8aa76e"} Oct 05 21:14:11 crc kubenswrapper[4754]: I1005 21:14:11.888151 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e720f41e8cb279f7b8b91bb1549ff3dfe463fbbe226f20aeff8c426f0e8aa76e" Oct 05 21:14:11 crc kubenswrapper[4754]: I1005 21:14:11.891735 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 05 21:14:11 crc kubenswrapper[4754]: I1005 21:14:11.998474 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c7f03a15-953e-4de8-b7b7-711cc35c8323-scripts\") pod \"c7f03a15-953e-4de8-b7b7-711cc35c8323\" (UID: \"c7f03a15-953e-4de8-b7b7-711cc35c8323\") " Oct 05 21:14:11 crc kubenswrapper[4754]: I1005 21:14:11.998538 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c7f03a15-953e-4de8-b7b7-711cc35c8323-log-httpd\") pod \"c7f03a15-953e-4de8-b7b7-711cc35c8323\" (UID: \"c7f03a15-953e-4de8-b7b7-711cc35c8323\") " Oct 05 21:14:11 crc kubenswrapper[4754]: I1005 21:14:11.998568 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c7f03a15-953e-4de8-b7b7-711cc35c8323-combined-ca-bundle\") pod \"c7f03a15-953e-4de8-b7b7-711cc35c8323\" (UID: \"c7f03a15-953e-4de8-b7b7-711cc35c8323\") " Oct 05 21:14:11 crc kubenswrapper[4754]: I1005 21:14:11.998640 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c7f03a15-953e-4de8-b7b7-711cc35c8323-sg-core-conf-yaml\") pod \"c7f03a15-953e-4de8-b7b7-711cc35c8323\" (UID: \"c7f03a15-953e-4de8-b7b7-711cc35c8323\") " Oct 05 21:14:11 crc kubenswrapper[4754]: I1005 21:14:11.998663 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c7f03a15-953e-4de8-b7b7-711cc35c8323-config-data\") pod \"c7f03a15-953e-4de8-b7b7-711cc35c8323\" (UID: \"c7f03a15-953e-4de8-b7b7-711cc35c8323\") " Oct 05 21:14:11 crc kubenswrapper[4754]: I1005 21:14:11.998781 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c7f03a15-953e-4de8-b7b7-711cc35c8323-run-httpd\") pod \"c7f03a15-953e-4de8-b7b7-711cc35c8323\" (UID: \"c7f03a15-953e-4de8-b7b7-711cc35c8323\") " Oct 05 21:14:11 crc kubenswrapper[4754]: I1005 21:14:11.998839 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-74zgd\" (UniqueName: \"kubernetes.io/projected/c7f03a15-953e-4de8-b7b7-711cc35c8323-kube-api-access-74zgd\") pod \"c7f03a15-953e-4de8-b7b7-711cc35c8323\" (UID: \"c7f03a15-953e-4de8-b7b7-711cc35c8323\") " Oct 05 21:14:12 crc kubenswrapper[4754]: I1005 21:14:12.000206 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c7f03a15-953e-4de8-b7b7-711cc35c8323-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "c7f03a15-953e-4de8-b7b7-711cc35c8323" (UID: "c7f03a15-953e-4de8-b7b7-711cc35c8323"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 21:14:12 crc kubenswrapper[4754]: I1005 21:14:12.000544 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c7f03a15-953e-4de8-b7b7-711cc35c8323-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "c7f03a15-953e-4de8-b7b7-711cc35c8323" (UID: "c7f03a15-953e-4de8-b7b7-711cc35c8323"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 21:14:12 crc kubenswrapper[4754]: I1005 21:14:12.018718 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c7f03a15-953e-4de8-b7b7-711cc35c8323-scripts" (OuterVolumeSpecName: "scripts") pod "c7f03a15-953e-4de8-b7b7-711cc35c8323" (UID: "c7f03a15-953e-4de8-b7b7-711cc35c8323"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:14:12 crc kubenswrapper[4754]: I1005 21:14:12.044525 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c7f03a15-953e-4de8-b7b7-711cc35c8323-kube-api-access-74zgd" (OuterVolumeSpecName: "kube-api-access-74zgd") pod "c7f03a15-953e-4de8-b7b7-711cc35c8323" (UID: "c7f03a15-953e-4de8-b7b7-711cc35c8323"). InnerVolumeSpecName "kube-api-access-74zgd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:14:12 crc kubenswrapper[4754]: I1005 21:14:12.086642 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c7f03a15-953e-4de8-b7b7-711cc35c8323-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "c7f03a15-953e-4de8-b7b7-711cc35c8323" (UID: "c7f03a15-953e-4de8-b7b7-711cc35c8323"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:14:12 crc kubenswrapper[4754]: I1005 21:14:12.101828 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-74zgd\" (UniqueName: \"kubernetes.io/projected/c7f03a15-953e-4de8-b7b7-711cc35c8323-kube-api-access-74zgd\") on node \"crc\" DevicePath \"\"" Oct 05 21:14:12 crc kubenswrapper[4754]: I1005 21:14:12.101867 4754 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c7f03a15-953e-4de8-b7b7-711cc35c8323-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 21:14:12 crc kubenswrapper[4754]: I1005 21:14:12.101878 4754 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c7f03a15-953e-4de8-b7b7-711cc35c8323-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 05 21:14:12 crc kubenswrapper[4754]: I1005 21:14:12.101889 4754 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c7f03a15-953e-4de8-b7b7-711cc35c8323-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 05 21:14:12 crc kubenswrapper[4754]: I1005 21:14:12.101898 4754 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c7f03a15-953e-4de8-b7b7-711cc35c8323-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 05 21:14:12 crc kubenswrapper[4754]: I1005 21:14:12.194942 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c7f03a15-953e-4de8-b7b7-711cc35c8323-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c7f03a15-953e-4de8-b7b7-711cc35c8323" (UID: "c7f03a15-953e-4de8-b7b7-711cc35c8323"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:14:12 crc kubenswrapper[4754]: I1005 21:14:12.204514 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c7f03a15-953e-4de8-b7b7-711cc35c8323-config-data" (OuterVolumeSpecName: "config-data") pod "c7f03a15-953e-4de8-b7b7-711cc35c8323" (UID: "c7f03a15-953e-4de8-b7b7-711cc35c8323"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:14:12 crc kubenswrapper[4754]: I1005 21:14:12.205520 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c7f03a15-953e-4de8-b7b7-711cc35c8323-config-data\") pod \"c7f03a15-953e-4de8-b7b7-711cc35c8323\" (UID: \"c7f03a15-953e-4de8-b7b7-711cc35c8323\") " Oct 05 21:14:12 crc kubenswrapper[4754]: W1005 21:14:12.205695 4754 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/c7f03a15-953e-4de8-b7b7-711cc35c8323/volumes/kubernetes.io~secret/config-data Oct 05 21:14:12 crc kubenswrapper[4754]: I1005 21:14:12.205718 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c7f03a15-953e-4de8-b7b7-711cc35c8323-config-data" (OuterVolumeSpecName: "config-data") pod "c7f03a15-953e-4de8-b7b7-711cc35c8323" (UID: "c7f03a15-953e-4de8-b7b7-711cc35c8323"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:14:12 crc kubenswrapper[4754]: I1005 21:14:12.207150 4754 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c7f03a15-953e-4de8-b7b7-711cc35c8323-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 21:14:12 crc kubenswrapper[4754]: I1005 21:14:12.207175 4754 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c7f03a15-953e-4de8-b7b7-711cc35c8323-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 21:14:12 crc kubenswrapper[4754]: I1005 21:14:12.931872 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 05 21:14:13 crc kubenswrapper[4754]: I1005 21:14:13.001253 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 05 21:14:13 crc kubenswrapper[4754]: I1005 21:14:13.017077 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 05 21:14:13 crc kubenswrapper[4754]: I1005 21:14:13.041674 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 05 21:14:13 crc kubenswrapper[4754]: E1005 21:14:13.042201 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7f03a15-953e-4de8-b7b7-711cc35c8323" containerName="ceilometer-notification-agent" Oct 05 21:14:13 crc kubenswrapper[4754]: I1005 21:14:13.042223 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7f03a15-953e-4de8-b7b7-711cc35c8323" containerName="ceilometer-notification-agent" Oct 05 21:14:13 crc kubenswrapper[4754]: E1005 21:14:13.042253 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24624447-d035-4476-b248-b7776bb261cd" containerName="mariadb-database-create" Oct 05 21:14:13 crc kubenswrapper[4754]: I1005 21:14:13.042261 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="24624447-d035-4476-b248-b7776bb261cd" containerName="mariadb-database-create" Oct 05 21:14:13 crc kubenswrapper[4754]: E1005 21:14:13.042276 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7f03a15-953e-4de8-b7b7-711cc35c8323" containerName="proxy-httpd" Oct 05 21:14:13 crc kubenswrapper[4754]: I1005 21:14:13.042283 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7f03a15-953e-4de8-b7b7-711cc35c8323" containerName="proxy-httpd" Oct 05 21:14:13 crc kubenswrapper[4754]: E1005 21:14:13.042293 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e45337f-b84d-473e-8530-942a17eb3c15" containerName="mariadb-database-create" Oct 05 21:14:13 crc kubenswrapper[4754]: I1005 21:14:13.042300 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e45337f-b84d-473e-8530-942a17eb3c15" containerName="mariadb-database-create" Oct 05 21:14:13 crc kubenswrapper[4754]: E1005 21:14:13.042310 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="339a3c5a-8feb-4381-ae02-dcff29fc61fc" containerName="mariadb-database-create" Oct 05 21:14:13 crc kubenswrapper[4754]: I1005 21:14:13.042317 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="339a3c5a-8feb-4381-ae02-dcff29fc61fc" containerName="mariadb-database-create" Oct 05 21:14:13 crc kubenswrapper[4754]: E1005 21:14:13.042335 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7f03a15-953e-4de8-b7b7-711cc35c8323" containerName="ceilometer-central-agent" Oct 05 21:14:13 crc kubenswrapper[4754]: I1005 21:14:13.042342 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7f03a15-953e-4de8-b7b7-711cc35c8323" containerName="ceilometer-central-agent" Oct 05 21:14:13 crc kubenswrapper[4754]: E1005 21:14:13.042359 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7f03a15-953e-4de8-b7b7-711cc35c8323" containerName="sg-core" Oct 05 21:14:13 crc kubenswrapper[4754]: I1005 21:14:13.042367 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7f03a15-953e-4de8-b7b7-711cc35c8323" containerName="sg-core" Oct 05 21:14:13 crc kubenswrapper[4754]: I1005 21:14:13.042570 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="c7f03a15-953e-4de8-b7b7-711cc35c8323" containerName="ceilometer-notification-agent" Oct 05 21:14:13 crc kubenswrapper[4754]: I1005 21:14:13.042589 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="c7f03a15-953e-4de8-b7b7-711cc35c8323" containerName="ceilometer-central-agent" Oct 05 21:14:13 crc kubenswrapper[4754]: I1005 21:14:13.042604 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="339a3c5a-8feb-4381-ae02-dcff29fc61fc" containerName="mariadb-database-create" Oct 05 21:14:13 crc kubenswrapper[4754]: I1005 21:14:13.042615 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e45337f-b84d-473e-8530-942a17eb3c15" containerName="mariadb-database-create" Oct 05 21:14:13 crc kubenswrapper[4754]: I1005 21:14:13.042626 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="c7f03a15-953e-4de8-b7b7-711cc35c8323" containerName="proxy-httpd" Oct 05 21:14:13 crc kubenswrapper[4754]: I1005 21:14:13.042634 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="24624447-d035-4476-b248-b7776bb261cd" containerName="mariadb-database-create" Oct 05 21:14:13 crc kubenswrapper[4754]: I1005 21:14:13.042646 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="c7f03a15-953e-4de8-b7b7-711cc35c8323" containerName="sg-core" Oct 05 21:14:13 crc kubenswrapper[4754]: I1005 21:14:13.044519 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 05 21:14:13 crc kubenswrapper[4754]: I1005 21:14:13.050117 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 05 21:14:13 crc kubenswrapper[4754]: I1005 21:14:13.051841 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 05 21:14:13 crc kubenswrapper[4754]: I1005 21:14:13.054757 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 05 21:14:13 crc kubenswrapper[4754]: I1005 21:14:13.061274 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 05 21:14:13 crc kubenswrapper[4754]: I1005 21:14:13.135043 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b15616c9-4fdf-4a35-a852-6aa8673b977a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b15616c9-4fdf-4a35-a852-6aa8673b977a\") " pod="openstack/ceilometer-0" Oct 05 21:14:13 crc kubenswrapper[4754]: I1005 21:14:13.135155 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b15616c9-4fdf-4a35-a852-6aa8673b977a-config-data\") pod \"ceilometer-0\" (UID: \"b15616c9-4fdf-4a35-a852-6aa8673b977a\") " pod="openstack/ceilometer-0" Oct 05 21:14:13 crc kubenswrapper[4754]: I1005 21:14:13.135253 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b15616c9-4fdf-4a35-a852-6aa8673b977a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b15616c9-4fdf-4a35-a852-6aa8673b977a\") " pod="openstack/ceilometer-0" Oct 05 21:14:13 crc kubenswrapper[4754]: I1005 21:14:13.135274 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/b15616c9-4fdf-4a35-a852-6aa8673b977a-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"b15616c9-4fdf-4a35-a852-6aa8673b977a\") " pod="openstack/ceilometer-0" Oct 05 21:14:13 crc kubenswrapper[4754]: I1005 21:14:13.135325 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b15616c9-4fdf-4a35-a852-6aa8673b977a-scripts\") pod \"ceilometer-0\" (UID: \"b15616c9-4fdf-4a35-a852-6aa8673b977a\") " pod="openstack/ceilometer-0" Oct 05 21:14:13 crc kubenswrapper[4754]: I1005 21:14:13.135356 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8c7m8\" (UniqueName: \"kubernetes.io/projected/b15616c9-4fdf-4a35-a852-6aa8673b977a-kube-api-access-8c7m8\") pod \"ceilometer-0\" (UID: \"b15616c9-4fdf-4a35-a852-6aa8673b977a\") " pod="openstack/ceilometer-0" Oct 05 21:14:13 crc kubenswrapper[4754]: I1005 21:14:13.135408 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b15616c9-4fdf-4a35-a852-6aa8673b977a-log-httpd\") pod \"ceilometer-0\" (UID: \"b15616c9-4fdf-4a35-a852-6aa8673b977a\") " pod="openstack/ceilometer-0" Oct 05 21:14:13 crc kubenswrapper[4754]: I1005 21:14:13.135434 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b15616c9-4fdf-4a35-a852-6aa8673b977a-run-httpd\") pod \"ceilometer-0\" (UID: \"b15616c9-4fdf-4a35-a852-6aa8673b977a\") " pod="openstack/ceilometer-0" Oct 05 21:14:13 crc kubenswrapper[4754]: I1005 21:14:13.237302 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b15616c9-4fdf-4a35-a852-6aa8673b977a-log-httpd\") pod \"ceilometer-0\" (UID: \"b15616c9-4fdf-4a35-a852-6aa8673b977a\") " pod="openstack/ceilometer-0" Oct 05 21:14:13 crc kubenswrapper[4754]: I1005 21:14:13.237392 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b15616c9-4fdf-4a35-a852-6aa8673b977a-run-httpd\") pod \"ceilometer-0\" (UID: \"b15616c9-4fdf-4a35-a852-6aa8673b977a\") " pod="openstack/ceilometer-0" Oct 05 21:14:13 crc kubenswrapper[4754]: I1005 21:14:13.237428 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b15616c9-4fdf-4a35-a852-6aa8673b977a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b15616c9-4fdf-4a35-a852-6aa8673b977a\") " pod="openstack/ceilometer-0" Oct 05 21:14:13 crc kubenswrapper[4754]: I1005 21:14:13.237471 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b15616c9-4fdf-4a35-a852-6aa8673b977a-config-data\") pod \"ceilometer-0\" (UID: \"b15616c9-4fdf-4a35-a852-6aa8673b977a\") " pod="openstack/ceilometer-0" Oct 05 21:14:13 crc kubenswrapper[4754]: I1005 21:14:13.237778 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b15616c9-4fdf-4a35-a852-6aa8673b977a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b15616c9-4fdf-4a35-a852-6aa8673b977a\") " pod="openstack/ceilometer-0" Oct 05 21:14:13 crc kubenswrapper[4754]: I1005 21:14:13.237828 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/b15616c9-4fdf-4a35-a852-6aa8673b977a-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"b15616c9-4fdf-4a35-a852-6aa8673b977a\") " pod="openstack/ceilometer-0" Oct 05 21:14:13 crc kubenswrapper[4754]: I1005 21:14:13.237871 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b15616c9-4fdf-4a35-a852-6aa8673b977a-scripts\") pod \"ceilometer-0\" (UID: \"b15616c9-4fdf-4a35-a852-6aa8673b977a\") " pod="openstack/ceilometer-0" Oct 05 21:14:13 crc kubenswrapper[4754]: I1005 21:14:13.237920 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8c7m8\" (UniqueName: \"kubernetes.io/projected/b15616c9-4fdf-4a35-a852-6aa8673b977a-kube-api-access-8c7m8\") pod \"ceilometer-0\" (UID: \"b15616c9-4fdf-4a35-a852-6aa8673b977a\") " pod="openstack/ceilometer-0" Oct 05 21:14:13 crc kubenswrapper[4754]: I1005 21:14:13.239643 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b15616c9-4fdf-4a35-a852-6aa8673b977a-log-httpd\") pod \"ceilometer-0\" (UID: \"b15616c9-4fdf-4a35-a852-6aa8673b977a\") " pod="openstack/ceilometer-0" Oct 05 21:14:13 crc kubenswrapper[4754]: I1005 21:14:13.242692 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b15616c9-4fdf-4a35-a852-6aa8673b977a-run-httpd\") pod \"ceilometer-0\" (UID: \"b15616c9-4fdf-4a35-a852-6aa8673b977a\") " pod="openstack/ceilometer-0" Oct 05 21:14:13 crc kubenswrapper[4754]: I1005 21:14:13.243517 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b15616c9-4fdf-4a35-a852-6aa8673b977a-config-data\") pod \"ceilometer-0\" (UID: \"b15616c9-4fdf-4a35-a852-6aa8673b977a\") " pod="openstack/ceilometer-0" Oct 05 21:14:13 crc kubenswrapper[4754]: I1005 21:14:13.249372 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b15616c9-4fdf-4a35-a852-6aa8673b977a-scripts\") pod \"ceilometer-0\" (UID: \"b15616c9-4fdf-4a35-a852-6aa8673b977a\") " pod="openstack/ceilometer-0" Oct 05 21:14:13 crc kubenswrapper[4754]: I1005 21:14:13.250694 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/b15616c9-4fdf-4a35-a852-6aa8673b977a-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"b15616c9-4fdf-4a35-a852-6aa8673b977a\") " pod="openstack/ceilometer-0" Oct 05 21:14:13 crc kubenswrapper[4754]: I1005 21:14:13.251152 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b15616c9-4fdf-4a35-a852-6aa8673b977a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b15616c9-4fdf-4a35-a852-6aa8673b977a\") " pod="openstack/ceilometer-0" Oct 05 21:14:13 crc kubenswrapper[4754]: I1005 21:14:13.253885 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b15616c9-4fdf-4a35-a852-6aa8673b977a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b15616c9-4fdf-4a35-a852-6aa8673b977a\") " pod="openstack/ceilometer-0" Oct 05 21:14:13 crc kubenswrapper[4754]: I1005 21:14:13.260374 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8c7m8\" (UniqueName: \"kubernetes.io/projected/b15616c9-4fdf-4a35-a852-6aa8673b977a-kube-api-access-8c7m8\") pod \"ceilometer-0\" (UID: \"b15616c9-4fdf-4a35-a852-6aa8673b977a\") " pod="openstack/ceilometer-0" Oct 05 21:14:13 crc kubenswrapper[4754]: I1005 21:14:13.378455 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 05 21:14:13 crc kubenswrapper[4754]: I1005 21:14:13.825798 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/cinder-api-0" podUID="90277df3-c655-423b-b6a6-8f3142ab046e" containerName="cinder-api" probeResult="failure" output="Get \"https://10.217.0.172:8776/healthcheck\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 05 21:14:13 crc kubenswrapper[4754]: I1005 21:14:13.959159 4754 generic.go:334] "Generic (PLEG): container finished" podID="85cf3d49-cc86-43ad-ad97-0e4476de6ee6" containerID="61c6c3b71ff15d6f9ac21858b8162e6919168883eb60955b3b78ab5843bf6b52" exitCode=0 Oct 05 21:14:13 crc kubenswrapper[4754]: I1005 21:14:13.959223 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"85cf3d49-cc86-43ad-ad97-0e4476de6ee6","Type":"ContainerDied","Data":"61c6c3b71ff15d6f9ac21858b8162e6919168883eb60955b3b78ab5843bf6b52"} Oct 05 21:14:13 crc kubenswrapper[4754]: I1005 21:14:13.984213 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 05 21:14:14 crc kubenswrapper[4754]: I1005 21:14:14.355665 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 05 21:14:14 crc kubenswrapper[4754]: I1005 21:14:14.478998 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/85cf3d49-cc86-43ad-ad97-0e4476de6ee6-config-data\") pod \"85cf3d49-cc86-43ad-ad97-0e4476de6ee6\" (UID: \"85cf3d49-cc86-43ad-ad97-0e4476de6ee6\") " Oct 05 21:14:14 crc kubenswrapper[4754]: I1005 21:14:14.479063 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"85cf3d49-cc86-43ad-ad97-0e4476de6ee6\" (UID: \"85cf3d49-cc86-43ad-ad97-0e4476de6ee6\") " Oct 05 21:14:14 crc kubenswrapper[4754]: I1005 21:14:14.479107 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/85cf3d49-cc86-43ad-ad97-0e4476de6ee6-public-tls-certs\") pod \"85cf3d49-cc86-43ad-ad97-0e4476de6ee6\" (UID: \"85cf3d49-cc86-43ad-ad97-0e4476de6ee6\") " Oct 05 21:14:14 crc kubenswrapper[4754]: I1005 21:14:14.479183 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85cf3d49-cc86-43ad-ad97-0e4476de6ee6-combined-ca-bundle\") pod \"85cf3d49-cc86-43ad-ad97-0e4476de6ee6\" (UID: \"85cf3d49-cc86-43ad-ad97-0e4476de6ee6\") " Oct 05 21:14:14 crc kubenswrapper[4754]: I1005 21:14:14.479241 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6p5nx\" (UniqueName: \"kubernetes.io/projected/85cf3d49-cc86-43ad-ad97-0e4476de6ee6-kube-api-access-6p5nx\") pod \"85cf3d49-cc86-43ad-ad97-0e4476de6ee6\" (UID: \"85cf3d49-cc86-43ad-ad97-0e4476de6ee6\") " Oct 05 21:14:14 crc kubenswrapper[4754]: I1005 21:14:14.479315 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/85cf3d49-cc86-43ad-ad97-0e4476de6ee6-httpd-run\") pod \"85cf3d49-cc86-43ad-ad97-0e4476de6ee6\" (UID: \"85cf3d49-cc86-43ad-ad97-0e4476de6ee6\") " Oct 05 21:14:14 crc kubenswrapper[4754]: I1005 21:14:14.479395 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/85cf3d49-cc86-43ad-ad97-0e4476de6ee6-logs\") pod \"85cf3d49-cc86-43ad-ad97-0e4476de6ee6\" (UID: \"85cf3d49-cc86-43ad-ad97-0e4476de6ee6\") " Oct 05 21:14:14 crc kubenswrapper[4754]: I1005 21:14:14.479420 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/85cf3d49-cc86-43ad-ad97-0e4476de6ee6-scripts\") pod \"85cf3d49-cc86-43ad-ad97-0e4476de6ee6\" (UID: \"85cf3d49-cc86-43ad-ad97-0e4476de6ee6\") " Oct 05 21:14:14 crc kubenswrapper[4754]: I1005 21:14:14.480114 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/85cf3d49-cc86-43ad-ad97-0e4476de6ee6-logs" (OuterVolumeSpecName: "logs") pod "85cf3d49-cc86-43ad-ad97-0e4476de6ee6" (UID: "85cf3d49-cc86-43ad-ad97-0e4476de6ee6"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 21:14:14 crc kubenswrapper[4754]: I1005 21:14:14.480766 4754 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/85cf3d49-cc86-43ad-ad97-0e4476de6ee6-logs\") on node \"crc\" DevicePath \"\"" Oct 05 21:14:14 crc kubenswrapper[4754]: I1005 21:14:14.489997 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/85cf3d49-cc86-43ad-ad97-0e4476de6ee6-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "85cf3d49-cc86-43ad-ad97-0e4476de6ee6" (UID: "85cf3d49-cc86-43ad-ad97-0e4476de6ee6"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 21:14:14 crc kubenswrapper[4754]: I1005 21:14:14.495833 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "glance") pod "85cf3d49-cc86-43ad-ad97-0e4476de6ee6" (UID: "85cf3d49-cc86-43ad-ad97-0e4476de6ee6"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 05 21:14:14 crc kubenswrapper[4754]: I1005 21:14:14.505216 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/85cf3d49-cc86-43ad-ad97-0e4476de6ee6-scripts" (OuterVolumeSpecName: "scripts") pod "85cf3d49-cc86-43ad-ad97-0e4476de6ee6" (UID: "85cf3d49-cc86-43ad-ad97-0e4476de6ee6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:14:14 crc kubenswrapper[4754]: I1005 21:14:14.506580 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/85cf3d49-cc86-43ad-ad97-0e4476de6ee6-kube-api-access-6p5nx" (OuterVolumeSpecName: "kube-api-access-6p5nx") pod "85cf3d49-cc86-43ad-ad97-0e4476de6ee6" (UID: "85cf3d49-cc86-43ad-ad97-0e4476de6ee6"). InnerVolumeSpecName "kube-api-access-6p5nx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:14:14 crc kubenswrapper[4754]: I1005 21:14:14.534552 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/85cf3d49-cc86-43ad-ad97-0e4476de6ee6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "85cf3d49-cc86-43ad-ad97-0e4476de6ee6" (UID: "85cf3d49-cc86-43ad-ad97-0e4476de6ee6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:14:14 crc kubenswrapper[4754]: I1005 21:14:14.534907 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Oct 05 21:14:14 crc kubenswrapper[4754]: I1005 21:14:14.584313 4754 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85cf3d49-cc86-43ad-ad97-0e4476de6ee6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 21:14:14 crc kubenswrapper[4754]: I1005 21:14:14.584663 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6p5nx\" (UniqueName: \"kubernetes.io/projected/85cf3d49-cc86-43ad-ad97-0e4476de6ee6-kube-api-access-6p5nx\") on node \"crc\" DevicePath \"\"" Oct 05 21:14:14 crc kubenswrapper[4754]: I1005 21:14:14.584737 4754 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/85cf3d49-cc86-43ad-ad97-0e4476de6ee6-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 05 21:14:14 crc kubenswrapper[4754]: I1005 21:14:14.584793 4754 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/85cf3d49-cc86-43ad-ad97-0e4476de6ee6-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 21:14:14 crc kubenswrapper[4754]: I1005 21:14:14.584905 4754 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Oct 05 21:14:14 crc kubenswrapper[4754]: I1005 21:14:14.644243 4754 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Oct 05 21:14:14 crc kubenswrapper[4754]: I1005 21:14:14.652079 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/85cf3d49-cc86-43ad-ad97-0e4476de6ee6-config-data" (OuterVolumeSpecName: "config-data") pod "85cf3d49-cc86-43ad-ad97-0e4476de6ee6" (UID: "85cf3d49-cc86-43ad-ad97-0e4476de6ee6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:14:14 crc kubenswrapper[4754]: I1005 21:14:14.656460 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/85cf3d49-cc86-43ad-ad97-0e4476de6ee6-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "85cf3d49-cc86-43ad-ad97-0e4476de6ee6" (UID: "85cf3d49-cc86-43ad-ad97-0e4476de6ee6"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:14:14 crc kubenswrapper[4754]: I1005 21:14:14.686935 4754 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/85cf3d49-cc86-43ad-ad97-0e4476de6ee6-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 21:14:14 crc kubenswrapper[4754]: I1005 21:14:14.686965 4754 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Oct 05 21:14:14 crc kubenswrapper[4754]: I1005 21:14:14.686975 4754 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/85cf3d49-cc86-43ad-ad97-0e4476de6ee6-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 05 21:14:14 crc kubenswrapper[4754]: I1005 21:14:14.889800 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c7f03a15-953e-4de8-b7b7-711cc35c8323" path="/var/lib/kubelet/pods/c7f03a15-953e-4de8-b7b7-711cc35c8323/volumes" Oct 05 21:14:15 crc kubenswrapper[4754]: I1005 21:14:15.011237 4754 generic.go:334] "Generic (PLEG): container finished" podID="a42df2f2-2289-4a9e-9ab4-156d3c8d2145" containerID="28d4be6a6eb29d90be12ae24e0bb51cb8fac12201eacdd04cc51c1bec5626a1d" exitCode=0 Oct 05 21:14:15 crc kubenswrapper[4754]: I1005 21:14:15.011405 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-77ccdf5dfb-zv8nn" event={"ID":"a42df2f2-2289-4a9e-9ab4-156d3c8d2145","Type":"ContainerDied","Data":"28d4be6a6eb29d90be12ae24e0bb51cb8fac12201eacdd04cc51c1bec5626a1d"} Oct 05 21:14:15 crc kubenswrapper[4754]: I1005 21:14:15.026392 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"85cf3d49-cc86-43ad-ad97-0e4476de6ee6","Type":"ContainerDied","Data":"dfab69117ac0d0afa11a7554d470ac8b1555ae8cceee3021b3d6bce7fa4433e2"} Oct 05 21:14:15 crc kubenswrapper[4754]: I1005 21:14:15.026458 4754 scope.go:117] "RemoveContainer" containerID="61c6c3b71ff15d6f9ac21858b8162e6919168883eb60955b3b78ab5843bf6b52" Oct 05 21:14:15 crc kubenswrapper[4754]: I1005 21:14:15.026754 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 05 21:14:15 crc kubenswrapper[4754]: I1005 21:14:15.042040 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b15616c9-4fdf-4a35-a852-6aa8673b977a","Type":"ContainerStarted","Data":"7dec7105830944b78be8543663bcb16050e4734f652ded16797f7afc68d45c7e"} Oct 05 21:14:15 crc kubenswrapper[4754]: I1005 21:14:15.084004 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 05 21:14:15 crc kubenswrapper[4754]: I1005 21:14:15.098549 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 05 21:14:15 crc kubenswrapper[4754]: I1005 21:14:15.122195 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 05 21:14:15 crc kubenswrapper[4754]: E1005 21:14:15.122614 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85cf3d49-cc86-43ad-ad97-0e4476de6ee6" containerName="glance-httpd" Oct 05 21:14:15 crc kubenswrapper[4754]: I1005 21:14:15.122626 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="85cf3d49-cc86-43ad-ad97-0e4476de6ee6" containerName="glance-httpd" Oct 05 21:14:15 crc kubenswrapper[4754]: E1005 21:14:15.122674 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85cf3d49-cc86-43ad-ad97-0e4476de6ee6" containerName="glance-log" Oct 05 21:14:15 crc kubenswrapper[4754]: I1005 21:14:15.122680 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="85cf3d49-cc86-43ad-ad97-0e4476de6ee6" containerName="glance-log" Oct 05 21:14:15 crc kubenswrapper[4754]: I1005 21:14:15.122854 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="85cf3d49-cc86-43ad-ad97-0e4476de6ee6" containerName="glance-httpd" Oct 05 21:14:15 crc kubenswrapper[4754]: I1005 21:14:15.122873 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="85cf3d49-cc86-43ad-ad97-0e4476de6ee6" containerName="glance-log" Oct 05 21:14:15 crc kubenswrapper[4754]: I1005 21:14:15.141099 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 05 21:14:15 crc kubenswrapper[4754]: I1005 21:14:15.153565 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 05 21:14:15 crc kubenswrapper[4754]: I1005 21:14:15.154769 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Oct 05 21:14:15 crc kubenswrapper[4754]: I1005 21:14:15.161225 4754 scope.go:117] "RemoveContainer" containerID="6d5bd90dbdb3906057a2a30e1be4ae1d3f097b8e2659882a66083011d71ad64d" Oct 05 21:14:15 crc kubenswrapper[4754]: I1005 21:14:15.228668 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"49d68627-98f5-4eec-b544-1de274b6e786\") " pod="openstack/glance-default-external-api-0" Oct 05 21:14:15 crc kubenswrapper[4754]: I1005 21:14:15.228854 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ngfw2\" (UniqueName: \"kubernetes.io/projected/49d68627-98f5-4eec-b544-1de274b6e786-kube-api-access-ngfw2\") pod \"glance-default-external-api-0\" (UID: \"49d68627-98f5-4eec-b544-1de274b6e786\") " pod="openstack/glance-default-external-api-0" Oct 05 21:14:15 crc kubenswrapper[4754]: I1005 21:14:15.228888 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/49d68627-98f5-4eec-b544-1de274b6e786-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"49d68627-98f5-4eec-b544-1de274b6e786\") " pod="openstack/glance-default-external-api-0" Oct 05 21:14:15 crc kubenswrapper[4754]: I1005 21:14:15.228949 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/49d68627-98f5-4eec-b544-1de274b6e786-config-data\") pod \"glance-default-external-api-0\" (UID: \"49d68627-98f5-4eec-b544-1de274b6e786\") " pod="openstack/glance-default-external-api-0" Oct 05 21:14:15 crc kubenswrapper[4754]: I1005 21:14:15.228986 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49d68627-98f5-4eec-b544-1de274b6e786-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"49d68627-98f5-4eec-b544-1de274b6e786\") " pod="openstack/glance-default-external-api-0" Oct 05 21:14:15 crc kubenswrapper[4754]: I1005 21:14:15.229033 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/49d68627-98f5-4eec-b544-1de274b6e786-scripts\") pod \"glance-default-external-api-0\" (UID: \"49d68627-98f5-4eec-b544-1de274b6e786\") " pod="openstack/glance-default-external-api-0" Oct 05 21:14:15 crc kubenswrapper[4754]: I1005 21:14:15.229196 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/49d68627-98f5-4eec-b544-1de274b6e786-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"49d68627-98f5-4eec-b544-1de274b6e786\") " pod="openstack/glance-default-external-api-0" Oct 05 21:14:15 crc kubenswrapper[4754]: I1005 21:14:15.229301 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/49d68627-98f5-4eec-b544-1de274b6e786-logs\") pod \"glance-default-external-api-0\" (UID: \"49d68627-98f5-4eec-b544-1de274b6e786\") " pod="openstack/glance-default-external-api-0" Oct 05 21:14:15 crc kubenswrapper[4754]: I1005 21:14:15.294962 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 05 21:14:15 crc kubenswrapper[4754]: I1005 21:14:15.332338 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/49d68627-98f5-4eec-b544-1de274b6e786-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"49d68627-98f5-4eec-b544-1de274b6e786\") " pod="openstack/glance-default-external-api-0" Oct 05 21:14:15 crc kubenswrapper[4754]: I1005 21:14:15.332614 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/49d68627-98f5-4eec-b544-1de274b6e786-logs\") pod \"glance-default-external-api-0\" (UID: \"49d68627-98f5-4eec-b544-1de274b6e786\") " pod="openstack/glance-default-external-api-0" Oct 05 21:14:15 crc kubenswrapper[4754]: I1005 21:14:15.332756 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"49d68627-98f5-4eec-b544-1de274b6e786\") " pod="openstack/glance-default-external-api-0" Oct 05 21:14:15 crc kubenswrapper[4754]: I1005 21:14:15.332866 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ngfw2\" (UniqueName: \"kubernetes.io/projected/49d68627-98f5-4eec-b544-1de274b6e786-kube-api-access-ngfw2\") pod \"glance-default-external-api-0\" (UID: \"49d68627-98f5-4eec-b544-1de274b6e786\") " pod="openstack/glance-default-external-api-0" Oct 05 21:14:15 crc kubenswrapper[4754]: I1005 21:14:15.332941 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/49d68627-98f5-4eec-b544-1de274b6e786-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"49d68627-98f5-4eec-b544-1de274b6e786\") " pod="openstack/glance-default-external-api-0" Oct 05 21:14:15 crc kubenswrapper[4754]: I1005 21:14:15.333020 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/49d68627-98f5-4eec-b544-1de274b6e786-config-data\") pod \"glance-default-external-api-0\" (UID: \"49d68627-98f5-4eec-b544-1de274b6e786\") " pod="openstack/glance-default-external-api-0" Oct 05 21:14:15 crc kubenswrapper[4754]: I1005 21:14:15.333116 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49d68627-98f5-4eec-b544-1de274b6e786-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"49d68627-98f5-4eec-b544-1de274b6e786\") " pod="openstack/glance-default-external-api-0" Oct 05 21:14:15 crc kubenswrapper[4754]: I1005 21:14:15.333192 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/49d68627-98f5-4eec-b544-1de274b6e786-scripts\") pod \"glance-default-external-api-0\" (UID: \"49d68627-98f5-4eec-b544-1de274b6e786\") " pod="openstack/glance-default-external-api-0" Oct 05 21:14:15 crc kubenswrapper[4754]: I1005 21:14:15.333697 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-89d4-account-create-vxv7h"] Oct 05 21:14:15 crc kubenswrapper[4754]: I1005 21:14:15.338287 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-89d4-account-create-vxv7h" Oct 05 21:14:15 crc kubenswrapper[4754]: I1005 21:14:15.338894 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/49d68627-98f5-4eec-b544-1de274b6e786-logs\") pod \"glance-default-external-api-0\" (UID: \"49d68627-98f5-4eec-b544-1de274b6e786\") " pod="openstack/glance-default-external-api-0" Oct 05 21:14:15 crc kubenswrapper[4754]: I1005 21:14:15.339454 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/49d68627-98f5-4eec-b544-1de274b6e786-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"49d68627-98f5-4eec-b544-1de274b6e786\") " pod="openstack/glance-default-external-api-0" Oct 05 21:14:15 crc kubenswrapper[4754]: I1005 21:14:15.340515 4754 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"49d68627-98f5-4eec-b544-1de274b6e786\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/glance-default-external-api-0" Oct 05 21:14:15 crc kubenswrapper[4754]: I1005 21:14:15.344062 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/49d68627-98f5-4eec-b544-1de274b6e786-scripts\") pod \"glance-default-external-api-0\" (UID: \"49d68627-98f5-4eec-b544-1de274b6e786\") " pod="openstack/glance-default-external-api-0" Oct 05 21:14:15 crc kubenswrapper[4754]: I1005 21:14:15.345688 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/49d68627-98f5-4eec-b544-1de274b6e786-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"49d68627-98f5-4eec-b544-1de274b6e786\") " pod="openstack/glance-default-external-api-0" Oct 05 21:14:15 crc kubenswrapper[4754]: I1005 21:14:15.345978 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Oct 05 21:14:15 crc kubenswrapper[4754]: I1005 21:14:15.406355 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49d68627-98f5-4eec-b544-1de274b6e786-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"49d68627-98f5-4eec-b544-1de274b6e786\") " pod="openstack/glance-default-external-api-0" Oct 05 21:14:15 crc kubenswrapper[4754]: I1005 21:14:15.409146 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ngfw2\" (UniqueName: \"kubernetes.io/projected/49d68627-98f5-4eec-b544-1de274b6e786-kube-api-access-ngfw2\") pod \"glance-default-external-api-0\" (UID: \"49d68627-98f5-4eec-b544-1de274b6e786\") " pod="openstack/glance-default-external-api-0" Oct 05 21:14:15 crc kubenswrapper[4754]: I1005 21:14:15.412243 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/49d68627-98f5-4eec-b544-1de274b6e786-config-data\") pod \"glance-default-external-api-0\" (UID: \"49d68627-98f5-4eec-b544-1de274b6e786\") " pod="openstack/glance-default-external-api-0" Oct 05 21:14:15 crc kubenswrapper[4754]: I1005 21:14:15.438080 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-49nd9\" (UniqueName: \"kubernetes.io/projected/6f2f8cf7-43a7-4845-b9a5-d39680402e09-kube-api-access-49nd9\") pod \"nova-api-89d4-account-create-vxv7h\" (UID: \"6f2f8cf7-43a7-4845-b9a5-d39680402e09\") " pod="openstack/nova-api-89d4-account-create-vxv7h" Oct 05 21:14:15 crc kubenswrapper[4754]: I1005 21:14:15.438348 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-89d4-account-create-vxv7h"] Oct 05 21:14:15 crc kubenswrapper[4754]: I1005 21:14:15.486131 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"49d68627-98f5-4eec-b544-1de274b6e786\") " pod="openstack/glance-default-external-api-0" Oct 05 21:14:15 crc kubenswrapper[4754]: I1005 21:14:15.497646 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-77ccdf5dfb-zv8nn" Oct 05 21:14:15 crc kubenswrapper[4754]: I1005 21:14:15.515264 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 05 21:14:15 crc kubenswrapper[4754]: I1005 21:14:15.539241 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/a42df2f2-2289-4a9e-9ab4-156d3c8d2145-httpd-config\") pod \"a42df2f2-2289-4a9e-9ab4-156d3c8d2145\" (UID: \"a42df2f2-2289-4a9e-9ab4-156d3c8d2145\") " Oct 05 21:14:15 crc kubenswrapper[4754]: I1005 21:14:15.539479 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/a42df2f2-2289-4a9e-9ab4-156d3c8d2145-ovndb-tls-certs\") pod \"a42df2f2-2289-4a9e-9ab4-156d3c8d2145\" (UID: \"a42df2f2-2289-4a9e-9ab4-156d3c8d2145\") " Oct 05 21:14:15 crc kubenswrapper[4754]: I1005 21:14:15.539529 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a42df2f2-2289-4a9e-9ab4-156d3c8d2145-combined-ca-bundle\") pod \"a42df2f2-2289-4a9e-9ab4-156d3c8d2145\" (UID: \"a42df2f2-2289-4a9e-9ab4-156d3c8d2145\") " Oct 05 21:14:15 crc kubenswrapper[4754]: I1005 21:14:15.539561 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/a42df2f2-2289-4a9e-9ab4-156d3c8d2145-config\") pod \"a42df2f2-2289-4a9e-9ab4-156d3c8d2145\" (UID: \"a42df2f2-2289-4a9e-9ab4-156d3c8d2145\") " Oct 05 21:14:15 crc kubenswrapper[4754]: I1005 21:14:15.539653 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhxrj\" (UniqueName: \"kubernetes.io/projected/a42df2f2-2289-4a9e-9ab4-156d3c8d2145-kube-api-access-jhxrj\") pod \"a42df2f2-2289-4a9e-9ab4-156d3c8d2145\" (UID: \"a42df2f2-2289-4a9e-9ab4-156d3c8d2145\") " Oct 05 21:14:15 crc kubenswrapper[4754]: I1005 21:14:15.540046 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-49nd9\" (UniqueName: \"kubernetes.io/projected/6f2f8cf7-43a7-4845-b9a5-d39680402e09-kube-api-access-49nd9\") pod \"nova-api-89d4-account-create-vxv7h\" (UID: \"6f2f8cf7-43a7-4845-b9a5-d39680402e09\") " pod="openstack/nova-api-89d4-account-create-vxv7h" Oct 05 21:14:15 crc kubenswrapper[4754]: I1005 21:14:15.559410 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a42df2f2-2289-4a9e-9ab4-156d3c8d2145-kube-api-access-jhxrj" (OuterVolumeSpecName: "kube-api-access-jhxrj") pod "a42df2f2-2289-4a9e-9ab4-156d3c8d2145" (UID: "a42df2f2-2289-4a9e-9ab4-156d3c8d2145"). InnerVolumeSpecName "kube-api-access-jhxrj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:14:15 crc kubenswrapper[4754]: I1005 21:14:15.564685 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a42df2f2-2289-4a9e-9ab4-156d3c8d2145-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "a42df2f2-2289-4a9e-9ab4-156d3c8d2145" (UID: "a42df2f2-2289-4a9e-9ab4-156d3c8d2145"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:14:15 crc kubenswrapper[4754]: I1005 21:14:15.567797 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-49nd9\" (UniqueName: \"kubernetes.io/projected/6f2f8cf7-43a7-4845-b9a5-d39680402e09-kube-api-access-49nd9\") pod \"nova-api-89d4-account-create-vxv7h\" (UID: \"6f2f8cf7-43a7-4845-b9a5-d39680402e09\") " pod="openstack/nova-api-89d4-account-create-vxv7h" Oct 05 21:14:15 crc kubenswrapper[4754]: I1005 21:14:15.615219 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-68c6b67864-9msm5" podUID="27cd6be5-f5eb-4158-bdd1-37f4df3d530d" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.148:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.148:8443: connect: connection refused" Oct 05 21:14:15 crc kubenswrapper[4754]: I1005 21:14:15.643671 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhxrj\" (UniqueName: \"kubernetes.io/projected/a42df2f2-2289-4a9e-9ab4-156d3c8d2145-kube-api-access-jhxrj\") on node \"crc\" DevicePath \"\"" Oct 05 21:14:15 crc kubenswrapper[4754]: I1005 21:14:15.643704 4754 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/a42df2f2-2289-4a9e-9ab4-156d3c8d2145-httpd-config\") on node \"crc\" DevicePath \"\"" Oct 05 21:14:15 crc kubenswrapper[4754]: I1005 21:14:15.645356 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a42df2f2-2289-4a9e-9ab4-156d3c8d2145-config" (OuterVolumeSpecName: "config") pod "a42df2f2-2289-4a9e-9ab4-156d3c8d2145" (UID: "a42df2f2-2289-4a9e-9ab4-156d3c8d2145"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:14:15 crc kubenswrapper[4754]: I1005 21:14:15.649100 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a42df2f2-2289-4a9e-9ab4-156d3c8d2145-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a42df2f2-2289-4a9e-9ab4-156d3c8d2145" (UID: "a42df2f2-2289-4a9e-9ab4-156d3c8d2145"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:14:15 crc kubenswrapper[4754]: I1005 21:14:15.712512 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a42df2f2-2289-4a9e-9ab4-156d3c8d2145-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "a42df2f2-2289-4a9e-9ab4-156d3c8d2145" (UID: "a42df2f2-2289-4a9e-9ab4-156d3c8d2145"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:14:15 crc kubenswrapper[4754]: I1005 21:14:15.725051 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-574454d6cb-vbnk6" podUID="39d1aac3-14a1-45bf-9a9d-ba4cdb0afcb1" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.149:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.149:8443: connect: connection refused" Oct 05 21:14:15 crc kubenswrapper[4754]: I1005 21:14:15.745147 4754 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/a42df2f2-2289-4a9e-9ab4-156d3c8d2145-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 05 21:14:15 crc kubenswrapper[4754]: I1005 21:14:15.745184 4754 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a42df2f2-2289-4a9e-9ab4-156d3c8d2145-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 21:14:15 crc kubenswrapper[4754]: I1005 21:14:15.745198 4754 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/a42df2f2-2289-4a9e-9ab4-156d3c8d2145-config\") on node \"crc\" DevicePath \"\"" Oct 05 21:14:15 crc kubenswrapper[4754]: I1005 21:14:15.847715 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-89d4-account-create-vxv7h" Oct 05 21:14:16 crc kubenswrapper[4754]: I1005 21:14:16.082869 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b15616c9-4fdf-4a35-a852-6aa8673b977a","Type":"ContainerStarted","Data":"650df3e2f17985e98ecd8928c1e230b948813039ccfc6799727843635d02f07e"} Oct 05 21:14:16 crc kubenswrapper[4754]: I1005 21:14:16.103819 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-77ccdf5dfb-zv8nn" event={"ID":"a42df2f2-2289-4a9e-9ab4-156d3c8d2145","Type":"ContainerDied","Data":"dbaeab7aae3fc7fc5e91dc81679ec8b0c29f7333c686562bfb7b18703ef34eef"} Oct 05 21:14:16 crc kubenswrapper[4754]: I1005 21:14:16.103884 4754 scope.go:117] "RemoveContainer" containerID="daa7d97295865858fb6daa372fbbc8824aebf09867c36045c8d5ba563ee53ed9" Oct 05 21:14:16 crc kubenswrapper[4754]: I1005 21:14:16.104001 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-77ccdf5dfb-zv8nn" Oct 05 21:14:16 crc kubenswrapper[4754]: I1005 21:14:16.211082 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 05 21:14:16 crc kubenswrapper[4754]: I1005 21:14:16.229894 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-77ccdf5dfb-zv8nn"] Oct 05 21:14:16 crc kubenswrapper[4754]: W1005 21:14:16.233956 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod49d68627_98f5_4eec_b544_1de274b6e786.slice/crio-3a04cdc7ae36cdb45550cd966bbdc69920d6ec38d24f3a6c920e5b8c84dccb19 WatchSource:0}: Error finding container 3a04cdc7ae36cdb45550cd966bbdc69920d6ec38d24f3a6c920e5b8c84dccb19: Status 404 returned error can't find the container with id 3a04cdc7ae36cdb45550cd966bbdc69920d6ec38d24f3a6c920e5b8c84dccb19 Oct 05 21:14:16 crc kubenswrapper[4754]: I1005 21:14:16.242962 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-77ccdf5dfb-zv8nn"] Oct 05 21:14:16 crc kubenswrapper[4754]: I1005 21:14:16.285850 4754 scope.go:117] "RemoveContainer" containerID="28d4be6a6eb29d90be12ae24e0bb51cb8fac12201eacdd04cc51c1bec5626a1d" Oct 05 21:14:16 crc kubenswrapper[4754]: I1005 21:14:16.540996 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 05 21:14:16 crc kubenswrapper[4754]: I1005 21:14:16.541372 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 05 21:14:16 crc kubenswrapper[4754]: I1005 21:14:16.575900 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-89d4-account-create-vxv7h"] Oct 05 21:14:16 crc kubenswrapper[4754]: I1005 21:14:16.629267 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 05 21:14:16 crc kubenswrapper[4754]: I1005 21:14:16.702240 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 05 21:14:16 crc kubenswrapper[4754]: I1005 21:14:16.881595 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="85cf3d49-cc86-43ad-ad97-0e4476de6ee6" path="/var/lib/kubelet/pods/85cf3d49-cc86-43ad-ad97-0e4476de6ee6/volumes" Oct 05 21:14:16 crc kubenswrapper[4754]: I1005 21:14:16.882409 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a42df2f2-2289-4a9e-9ab4-156d3c8d2145" path="/var/lib/kubelet/pods/a42df2f2-2289-4a9e-9ab4-156d3c8d2145/volumes" Oct 05 21:14:17 crc kubenswrapper[4754]: I1005 21:14:17.215935 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-89d4-account-create-vxv7h" event={"ID":"6f2f8cf7-43a7-4845-b9a5-d39680402e09","Type":"ContainerStarted","Data":"e770d8047e6ea28e9c235c0bb4015aa91597f690343bb0169990301305cd4df6"} Oct 05 21:14:17 crc kubenswrapper[4754]: I1005 21:14:17.238462 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b15616c9-4fdf-4a35-a852-6aa8673b977a","Type":"ContainerStarted","Data":"8a4a5b78bf4a081dbf3dca51440ad48381da380a1fc7cdf875a12491e252c7ff"} Oct 05 21:14:17 crc kubenswrapper[4754]: I1005 21:14:17.259364 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"49d68627-98f5-4eec-b544-1de274b6e786","Type":"ContainerStarted","Data":"3a04cdc7ae36cdb45550cd966bbdc69920d6ec38d24f3a6c920e5b8c84dccb19"} Oct 05 21:14:17 crc kubenswrapper[4754]: I1005 21:14:17.259441 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 05 21:14:17 crc kubenswrapper[4754]: I1005 21:14:17.259454 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 05 21:14:18 crc kubenswrapper[4754]: I1005 21:14:18.279320 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"49d68627-98f5-4eec-b544-1de274b6e786","Type":"ContainerStarted","Data":"37970b9ee341ddd5fb45d165c48ee918de24e696ed8e41a96b17899c0dc4f174"} Oct 05 21:14:18 crc kubenswrapper[4754]: I1005 21:14:18.282017 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-89d4-account-create-vxv7h" event={"ID":"6f2f8cf7-43a7-4845-b9a5-d39680402e09","Type":"ContainerStarted","Data":"d4d26a2314f7524c5a56cc71b40492123a49edd7c648d3266e69d2b6bffc2a7f"} Oct 05 21:14:18 crc kubenswrapper[4754]: I1005 21:14:18.296583 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b15616c9-4fdf-4a35-a852-6aa8673b977a","Type":"ContainerStarted","Data":"ffdd06d32874aa55b407f0be95f91d5501e5a978d7fd31bb1e1bda4847180178"} Oct 05 21:14:18 crc kubenswrapper[4754]: I1005 21:14:18.306762 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-89d4-account-create-vxv7h" podStartSLOduration=3.306740758 podStartE2EDuration="3.306740758s" podCreationTimestamp="2025-10-05 21:14:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 21:14:18.304768095 +0000 UTC m=+1182.208886805" watchObservedRunningTime="2025-10-05 21:14:18.306740758 +0000 UTC m=+1182.210859468" Oct 05 21:14:18 crc kubenswrapper[4754]: I1005 21:14:18.831757 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/cinder-api-0" podUID="90277df3-c655-423b-b6a6-8f3142ab046e" containerName="cinder-api" probeResult="failure" output="Get \"https://10.217.0.172:8776/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 05 21:14:19 crc kubenswrapper[4754]: I1005 21:14:19.334839 4754 generic.go:334] "Generic (PLEG): container finished" podID="6f2f8cf7-43a7-4845-b9a5-d39680402e09" containerID="d4d26a2314f7524c5a56cc71b40492123a49edd7c648d3266e69d2b6bffc2a7f" exitCode=0 Oct 05 21:14:19 crc kubenswrapper[4754]: I1005 21:14:19.334943 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-89d4-account-create-vxv7h" event={"ID":"6f2f8cf7-43a7-4845-b9a5-d39680402e09","Type":"ContainerDied","Data":"d4d26a2314f7524c5a56cc71b40492123a49edd7c648d3266e69d2b6bffc2a7f"} Oct 05 21:14:19 crc kubenswrapper[4754]: I1005 21:14:19.345652 4754 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 05 21:14:19 crc kubenswrapper[4754]: I1005 21:14:19.345699 4754 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 05 21:14:19 crc kubenswrapper[4754]: I1005 21:14:19.347232 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"49d68627-98f5-4eec-b544-1de274b6e786","Type":"ContainerStarted","Data":"00416a7683137e945f40b8648ffc6f851b0b79904e984d12fa3aa31e81b3c4ad"} Oct 05 21:14:19 crc kubenswrapper[4754]: I1005 21:14:19.385212 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=4.385188383 podStartE2EDuration="4.385188383s" podCreationTimestamp="2025-10-05 21:14:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 21:14:19.382788559 +0000 UTC m=+1183.286907289" watchObservedRunningTime="2025-10-05 21:14:19.385188383 +0000 UTC m=+1183.289307093" Oct 05 21:14:20 crc kubenswrapper[4754]: I1005 21:14:20.408407 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b15616c9-4fdf-4a35-a852-6aa8673b977a","Type":"ContainerStarted","Data":"8ecc956bd51b18d859bbc521c05ced28fe4286e7455e20544b6bb79c5302ce28"} Oct 05 21:14:20 crc kubenswrapper[4754]: I1005 21:14:20.416527 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 05 21:14:20 crc kubenswrapper[4754]: I1005 21:14:20.443590 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.72807719 podStartE2EDuration="8.44356474s" podCreationTimestamp="2025-10-05 21:14:12 +0000 UTC" firstStartedPulling="2025-10-05 21:14:14.007688791 +0000 UTC m=+1177.911807501" lastFinishedPulling="2025-10-05 21:14:19.723176341 +0000 UTC m=+1183.627295051" observedRunningTime="2025-10-05 21:14:20.439250944 +0000 UTC m=+1184.343369654" watchObservedRunningTime="2025-10-05 21:14:20.44356474 +0000 UTC m=+1184.347683450" Oct 05 21:14:20 crc kubenswrapper[4754]: I1005 21:14:20.829825 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-89d4-account-create-vxv7h" Oct 05 21:14:20 crc kubenswrapper[4754]: I1005 21:14:20.889660 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-49nd9\" (UniqueName: \"kubernetes.io/projected/6f2f8cf7-43a7-4845-b9a5-d39680402e09-kube-api-access-49nd9\") pod \"6f2f8cf7-43a7-4845-b9a5-d39680402e09\" (UID: \"6f2f8cf7-43a7-4845-b9a5-d39680402e09\") " Oct 05 21:14:20 crc kubenswrapper[4754]: I1005 21:14:20.903062 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6f2f8cf7-43a7-4845-b9a5-d39680402e09-kube-api-access-49nd9" (OuterVolumeSpecName: "kube-api-access-49nd9") pod "6f2f8cf7-43a7-4845-b9a5-d39680402e09" (UID: "6f2f8cf7-43a7-4845-b9a5-d39680402e09"). InnerVolumeSpecName "kube-api-access-49nd9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:14:20 crc kubenswrapper[4754]: I1005 21:14:20.994265 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-49nd9\" (UniqueName: \"kubernetes.io/projected/6f2f8cf7-43a7-4845-b9a5-d39680402e09-kube-api-access-49nd9\") on node \"crc\" DevicePath \"\"" Oct 05 21:14:21 crc kubenswrapper[4754]: I1005 21:14:21.424355 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-89d4-account-create-vxv7h" Oct 05 21:14:21 crc kubenswrapper[4754]: I1005 21:14:21.424427 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-89d4-account-create-vxv7h" event={"ID":"6f2f8cf7-43a7-4845-b9a5-d39680402e09","Type":"ContainerDied","Data":"e770d8047e6ea28e9c235c0bb4015aa91597f690343bb0169990301305cd4df6"} Oct 05 21:14:21 crc kubenswrapper[4754]: I1005 21:14:21.424465 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e770d8047e6ea28e9c235c0bb4015aa91597f690343bb0169990301305cd4df6" Oct 05 21:14:21 crc kubenswrapper[4754]: I1005 21:14:21.612171 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 05 21:14:21 crc kubenswrapper[4754]: I1005 21:14:21.612321 4754 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 05 21:14:21 crc kubenswrapper[4754]: I1005 21:14:21.790303 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 05 21:14:25 crc kubenswrapper[4754]: I1005 21:14:25.435093 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-1d7a-account-create-dtspg"] Oct 05 21:14:25 crc kubenswrapper[4754]: E1005 21:14:25.436130 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a42df2f2-2289-4a9e-9ab4-156d3c8d2145" containerName="neutron-httpd" Oct 05 21:14:25 crc kubenswrapper[4754]: I1005 21:14:25.436142 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="a42df2f2-2289-4a9e-9ab4-156d3c8d2145" containerName="neutron-httpd" Oct 05 21:14:25 crc kubenswrapper[4754]: E1005 21:14:25.436159 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f2f8cf7-43a7-4845-b9a5-d39680402e09" containerName="mariadb-account-create" Oct 05 21:14:25 crc kubenswrapper[4754]: I1005 21:14:25.436165 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f2f8cf7-43a7-4845-b9a5-d39680402e09" containerName="mariadb-account-create" Oct 05 21:14:25 crc kubenswrapper[4754]: E1005 21:14:25.436177 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a42df2f2-2289-4a9e-9ab4-156d3c8d2145" containerName="neutron-api" Oct 05 21:14:25 crc kubenswrapper[4754]: I1005 21:14:25.436183 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="a42df2f2-2289-4a9e-9ab4-156d3c8d2145" containerName="neutron-api" Oct 05 21:14:25 crc kubenswrapper[4754]: I1005 21:14:25.436340 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="6f2f8cf7-43a7-4845-b9a5-d39680402e09" containerName="mariadb-account-create" Oct 05 21:14:25 crc kubenswrapper[4754]: I1005 21:14:25.436354 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="a42df2f2-2289-4a9e-9ab4-156d3c8d2145" containerName="neutron-api" Oct 05 21:14:25 crc kubenswrapper[4754]: I1005 21:14:25.436372 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="a42df2f2-2289-4a9e-9ab4-156d3c8d2145" containerName="neutron-httpd" Oct 05 21:14:25 crc kubenswrapper[4754]: I1005 21:14:25.437550 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-1d7a-account-create-dtspg" Oct 05 21:14:25 crc kubenswrapper[4754]: I1005 21:14:25.439295 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Oct 05 21:14:25 crc kubenswrapper[4754]: I1005 21:14:25.446361 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-1d7a-account-create-dtspg"] Oct 05 21:14:25 crc kubenswrapper[4754]: I1005 21:14:25.516382 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 05 21:14:25 crc kubenswrapper[4754]: I1005 21:14:25.516463 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 05 21:14:25 crc kubenswrapper[4754]: I1005 21:14:25.555020 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 05 21:14:25 crc kubenswrapper[4754]: I1005 21:14:25.566904 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 05 21:14:25 crc kubenswrapper[4754]: I1005 21:14:25.598079 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2lqd7\" (UniqueName: \"kubernetes.io/projected/48089a5e-fc86-40cd-96db-c522040598c4-kube-api-access-2lqd7\") pod \"nova-cell0-1d7a-account-create-dtspg\" (UID: \"48089a5e-fc86-40cd-96db-c522040598c4\") " pod="openstack/nova-cell0-1d7a-account-create-dtspg" Oct 05 21:14:25 crc kubenswrapper[4754]: I1005 21:14:25.603924 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-68c6b67864-9msm5" podUID="27cd6be5-f5eb-4158-bdd1-37f4df3d530d" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.148:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.148:8443: connect: connection refused" Oct 05 21:14:25 crc kubenswrapper[4754]: I1005 21:14:25.604038 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-68c6b67864-9msm5" Oct 05 21:14:25 crc kubenswrapper[4754]: I1005 21:14:25.605100 4754 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="horizon" containerStatusID={"Type":"cri-o","ID":"8cbf678dfb968e1a5f68bc5b282346cef6ff3c4725e9f27e572e9658e525228e"} pod="openstack/horizon-68c6b67864-9msm5" containerMessage="Container horizon failed startup probe, will be restarted" Oct 05 21:14:25 crc kubenswrapper[4754]: I1005 21:14:25.605148 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-68c6b67864-9msm5" podUID="27cd6be5-f5eb-4158-bdd1-37f4df3d530d" containerName="horizon" containerID="cri-o://8cbf678dfb968e1a5f68bc5b282346cef6ff3c4725e9f27e572e9658e525228e" gracePeriod=30 Oct 05 21:14:25 crc kubenswrapper[4754]: I1005 21:14:25.634204 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-e205-account-create-hh2rc"] Oct 05 21:14:25 crc kubenswrapper[4754]: I1005 21:14:25.635958 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-e205-account-create-hh2rc" Oct 05 21:14:25 crc kubenswrapper[4754]: I1005 21:14:25.645257 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-e205-account-create-hh2rc"] Oct 05 21:14:25 crc kubenswrapper[4754]: I1005 21:14:25.650922 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Oct 05 21:14:25 crc kubenswrapper[4754]: I1005 21:14:25.700976 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2lqd7\" (UniqueName: \"kubernetes.io/projected/48089a5e-fc86-40cd-96db-c522040598c4-kube-api-access-2lqd7\") pod \"nova-cell0-1d7a-account-create-dtspg\" (UID: \"48089a5e-fc86-40cd-96db-c522040598c4\") " pod="openstack/nova-cell0-1d7a-account-create-dtspg" Oct 05 21:14:25 crc kubenswrapper[4754]: I1005 21:14:25.721187 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2lqd7\" (UniqueName: \"kubernetes.io/projected/48089a5e-fc86-40cd-96db-c522040598c4-kube-api-access-2lqd7\") pod \"nova-cell0-1d7a-account-create-dtspg\" (UID: \"48089a5e-fc86-40cd-96db-c522040598c4\") " pod="openstack/nova-cell0-1d7a-account-create-dtspg" Oct 05 21:14:25 crc kubenswrapper[4754]: I1005 21:14:25.721216 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-574454d6cb-vbnk6" podUID="39d1aac3-14a1-45bf-9a9d-ba4cdb0afcb1" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.149:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.149:8443: connect: connection refused" Oct 05 21:14:25 crc kubenswrapper[4754]: I1005 21:14:25.721305 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-574454d6cb-vbnk6" Oct 05 21:14:25 crc kubenswrapper[4754]: I1005 21:14:25.722449 4754 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="horizon" containerStatusID={"Type":"cri-o","ID":"704413f5b17f48e4e3d7d104afa7e9d7beadfdcc2c562b18589f4f0a201b7281"} pod="openstack/horizon-574454d6cb-vbnk6" containerMessage="Container horizon failed startup probe, will be restarted" Oct 05 21:14:25 crc kubenswrapper[4754]: I1005 21:14:25.722506 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-574454d6cb-vbnk6" podUID="39d1aac3-14a1-45bf-9a9d-ba4cdb0afcb1" containerName="horizon" containerID="cri-o://704413f5b17f48e4e3d7d104afa7e9d7beadfdcc2c562b18589f4f0a201b7281" gracePeriod=30 Oct 05 21:14:25 crc kubenswrapper[4754]: I1005 21:14:25.764072 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-1d7a-account-create-dtspg" Oct 05 21:14:25 crc kubenswrapper[4754]: I1005 21:14:25.802912 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-29jjf\" (UniqueName: \"kubernetes.io/projected/fa9e9f3c-9ee0-426d-aafe-0c39c9980b53-kube-api-access-29jjf\") pod \"nova-cell1-e205-account-create-hh2rc\" (UID: \"fa9e9f3c-9ee0-426d-aafe-0c39c9980b53\") " pod="openstack/nova-cell1-e205-account-create-hh2rc" Oct 05 21:14:25 crc kubenswrapper[4754]: I1005 21:14:25.905321 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-29jjf\" (UniqueName: \"kubernetes.io/projected/fa9e9f3c-9ee0-426d-aafe-0c39c9980b53-kube-api-access-29jjf\") pod \"nova-cell1-e205-account-create-hh2rc\" (UID: \"fa9e9f3c-9ee0-426d-aafe-0c39c9980b53\") " pod="openstack/nova-cell1-e205-account-create-hh2rc" Oct 05 21:14:25 crc kubenswrapper[4754]: I1005 21:14:25.981127 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-29jjf\" (UniqueName: \"kubernetes.io/projected/fa9e9f3c-9ee0-426d-aafe-0c39c9980b53-kube-api-access-29jjf\") pod \"nova-cell1-e205-account-create-hh2rc\" (UID: \"fa9e9f3c-9ee0-426d-aafe-0c39c9980b53\") " pod="openstack/nova-cell1-e205-account-create-hh2rc" Oct 05 21:14:26 crc kubenswrapper[4754]: I1005 21:14:26.254626 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-e205-account-create-hh2rc" Oct 05 21:14:26 crc kubenswrapper[4754]: I1005 21:14:26.394106 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-1d7a-account-create-dtspg"] Oct 05 21:14:26 crc kubenswrapper[4754]: I1005 21:14:26.491152 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-1d7a-account-create-dtspg" event={"ID":"48089a5e-fc86-40cd-96db-c522040598c4","Type":"ContainerStarted","Data":"717fa57df84605863b0df2853225b4c9e05259b4bbaa011a491d8688c63c0b94"} Oct 05 21:14:26 crc kubenswrapper[4754]: I1005 21:14:26.491290 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 05 21:14:26 crc kubenswrapper[4754]: I1005 21:14:26.491311 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 05 21:14:26 crc kubenswrapper[4754]: I1005 21:14:26.519601 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-e205-account-create-hh2rc"] Oct 05 21:14:27 crc kubenswrapper[4754]: I1005 21:14:27.503180 4754 generic.go:334] "Generic (PLEG): container finished" podID="48089a5e-fc86-40cd-96db-c522040598c4" containerID="cc96f5fe37567878834101eddfd72e749514fc5b7abf47dabfa6c8684f47ac67" exitCode=0 Oct 05 21:14:27 crc kubenswrapper[4754]: I1005 21:14:27.503312 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-1d7a-account-create-dtspg" event={"ID":"48089a5e-fc86-40cd-96db-c522040598c4","Type":"ContainerDied","Data":"cc96f5fe37567878834101eddfd72e749514fc5b7abf47dabfa6c8684f47ac67"} Oct 05 21:14:27 crc kubenswrapper[4754]: I1005 21:14:27.507038 4754 generic.go:334] "Generic (PLEG): container finished" podID="fa9e9f3c-9ee0-426d-aafe-0c39c9980b53" containerID="14be147d93ae83ceeb0bc2bb4a7a02b0e0b47263ec54292f2da834c0a8aeea03" exitCode=0 Oct 05 21:14:27 crc kubenswrapper[4754]: I1005 21:14:27.507311 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-e205-account-create-hh2rc" event={"ID":"fa9e9f3c-9ee0-426d-aafe-0c39c9980b53","Type":"ContainerDied","Data":"14be147d93ae83ceeb0bc2bb4a7a02b0e0b47263ec54292f2da834c0a8aeea03"} Oct 05 21:14:27 crc kubenswrapper[4754]: I1005 21:14:27.507368 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-e205-account-create-hh2rc" event={"ID":"fa9e9f3c-9ee0-426d-aafe-0c39c9980b53","Type":"ContainerStarted","Data":"d539d3482763a7f62839582a76aaf165aa9556739148892d6bb91efe97d218cc"} Oct 05 21:14:28 crc kubenswrapper[4754]: I1005 21:14:28.514737 4754 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 05 21:14:28 crc kubenswrapper[4754]: I1005 21:14:28.515126 4754 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 05 21:14:28 crc kubenswrapper[4754]: I1005 21:14:28.987771 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-1d7a-account-create-dtspg" Oct 05 21:14:28 crc kubenswrapper[4754]: I1005 21:14:28.995182 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-e205-account-create-hh2rc" Oct 05 21:14:29 crc kubenswrapper[4754]: I1005 21:14:29.128788 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-29jjf\" (UniqueName: \"kubernetes.io/projected/fa9e9f3c-9ee0-426d-aafe-0c39c9980b53-kube-api-access-29jjf\") pod \"fa9e9f3c-9ee0-426d-aafe-0c39c9980b53\" (UID: \"fa9e9f3c-9ee0-426d-aafe-0c39c9980b53\") " Oct 05 21:14:29 crc kubenswrapper[4754]: I1005 21:14:29.128880 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2lqd7\" (UniqueName: \"kubernetes.io/projected/48089a5e-fc86-40cd-96db-c522040598c4-kube-api-access-2lqd7\") pod \"48089a5e-fc86-40cd-96db-c522040598c4\" (UID: \"48089a5e-fc86-40cd-96db-c522040598c4\") " Oct 05 21:14:29 crc kubenswrapper[4754]: I1005 21:14:29.135878 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fa9e9f3c-9ee0-426d-aafe-0c39c9980b53-kube-api-access-29jjf" (OuterVolumeSpecName: "kube-api-access-29jjf") pod "fa9e9f3c-9ee0-426d-aafe-0c39c9980b53" (UID: "fa9e9f3c-9ee0-426d-aafe-0c39c9980b53"). InnerVolumeSpecName "kube-api-access-29jjf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:14:29 crc kubenswrapper[4754]: I1005 21:14:29.136545 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/48089a5e-fc86-40cd-96db-c522040598c4-kube-api-access-2lqd7" (OuterVolumeSpecName: "kube-api-access-2lqd7") pod "48089a5e-fc86-40cd-96db-c522040598c4" (UID: "48089a5e-fc86-40cd-96db-c522040598c4"). InnerVolumeSpecName "kube-api-access-2lqd7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:14:29 crc kubenswrapper[4754]: I1005 21:14:29.232280 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-29jjf\" (UniqueName: \"kubernetes.io/projected/fa9e9f3c-9ee0-426d-aafe-0c39c9980b53-kube-api-access-29jjf\") on node \"crc\" DevicePath \"\"" Oct 05 21:14:29 crc kubenswrapper[4754]: I1005 21:14:29.232328 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2lqd7\" (UniqueName: \"kubernetes.io/projected/48089a5e-fc86-40cd-96db-c522040598c4-kube-api-access-2lqd7\") on node \"crc\" DevicePath \"\"" Oct 05 21:14:29 crc kubenswrapper[4754]: I1005 21:14:29.256248 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 05 21:14:29 crc kubenswrapper[4754]: I1005 21:14:29.259296 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 05 21:14:29 crc kubenswrapper[4754]: I1005 21:14:29.525880 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-1d7a-account-create-dtspg" event={"ID":"48089a5e-fc86-40cd-96db-c522040598c4","Type":"ContainerDied","Data":"717fa57df84605863b0df2853225b4c9e05259b4bbaa011a491d8688c63c0b94"} Oct 05 21:14:29 crc kubenswrapper[4754]: I1005 21:14:29.526255 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="717fa57df84605863b0df2853225b4c9e05259b4bbaa011a491d8688c63c0b94" Oct 05 21:14:29 crc kubenswrapper[4754]: I1005 21:14:29.525904 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-1d7a-account-create-dtspg" Oct 05 21:14:29 crc kubenswrapper[4754]: I1005 21:14:29.528695 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-e205-account-create-hh2rc" Oct 05 21:14:29 crc kubenswrapper[4754]: I1005 21:14:29.540110 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-e205-account-create-hh2rc" event={"ID":"fa9e9f3c-9ee0-426d-aafe-0c39c9980b53","Type":"ContainerDied","Data":"d539d3482763a7f62839582a76aaf165aa9556739148892d6bb91efe97d218cc"} Oct 05 21:14:29 crc kubenswrapper[4754]: I1005 21:14:29.540152 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d539d3482763a7f62839582a76aaf165aa9556739148892d6bb91efe97d218cc" Oct 05 21:14:30 crc kubenswrapper[4754]: I1005 21:14:30.643110 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-4p4pw"] Oct 05 21:14:30 crc kubenswrapper[4754]: E1005 21:14:30.644163 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48089a5e-fc86-40cd-96db-c522040598c4" containerName="mariadb-account-create" Oct 05 21:14:30 crc kubenswrapper[4754]: I1005 21:14:30.644184 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="48089a5e-fc86-40cd-96db-c522040598c4" containerName="mariadb-account-create" Oct 05 21:14:30 crc kubenswrapper[4754]: E1005 21:14:30.644207 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa9e9f3c-9ee0-426d-aafe-0c39c9980b53" containerName="mariadb-account-create" Oct 05 21:14:30 crc kubenswrapper[4754]: I1005 21:14:30.644214 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa9e9f3c-9ee0-426d-aafe-0c39c9980b53" containerName="mariadb-account-create" Oct 05 21:14:30 crc kubenswrapper[4754]: I1005 21:14:30.644434 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa9e9f3c-9ee0-426d-aafe-0c39c9980b53" containerName="mariadb-account-create" Oct 05 21:14:30 crc kubenswrapper[4754]: I1005 21:14:30.644449 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="48089a5e-fc86-40cd-96db-c522040598c4" containerName="mariadb-account-create" Oct 05 21:14:30 crc kubenswrapper[4754]: I1005 21:14:30.645258 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-4p4pw" Oct 05 21:14:30 crc kubenswrapper[4754]: I1005 21:14:30.647417 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Oct 05 21:14:30 crc kubenswrapper[4754]: I1005 21:14:30.647614 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-kg5j2" Oct 05 21:14:30 crc kubenswrapper[4754]: I1005 21:14:30.653266 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 05 21:14:30 crc kubenswrapper[4754]: I1005 21:14:30.668965 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-4p4pw"] Oct 05 21:14:30 crc kubenswrapper[4754]: I1005 21:14:30.775816 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/498207e0-f869-46d9-bdc9-d83f44d02385-scripts\") pod \"nova-cell0-conductor-db-sync-4p4pw\" (UID: \"498207e0-f869-46d9-bdc9-d83f44d02385\") " pod="openstack/nova-cell0-conductor-db-sync-4p4pw" Oct 05 21:14:30 crc kubenswrapper[4754]: I1005 21:14:30.776344 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/498207e0-f869-46d9-bdc9-d83f44d02385-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-4p4pw\" (UID: \"498207e0-f869-46d9-bdc9-d83f44d02385\") " pod="openstack/nova-cell0-conductor-db-sync-4p4pw" Oct 05 21:14:30 crc kubenswrapper[4754]: I1005 21:14:30.776457 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/498207e0-f869-46d9-bdc9-d83f44d02385-config-data\") pod \"nova-cell0-conductor-db-sync-4p4pw\" (UID: \"498207e0-f869-46d9-bdc9-d83f44d02385\") " pod="openstack/nova-cell0-conductor-db-sync-4p4pw" Oct 05 21:14:30 crc kubenswrapper[4754]: I1005 21:14:30.776503 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wsjjx\" (UniqueName: \"kubernetes.io/projected/498207e0-f869-46d9-bdc9-d83f44d02385-kube-api-access-wsjjx\") pod \"nova-cell0-conductor-db-sync-4p4pw\" (UID: \"498207e0-f869-46d9-bdc9-d83f44d02385\") " pod="openstack/nova-cell0-conductor-db-sync-4p4pw" Oct 05 21:14:30 crc kubenswrapper[4754]: I1005 21:14:30.878599 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/498207e0-f869-46d9-bdc9-d83f44d02385-scripts\") pod \"nova-cell0-conductor-db-sync-4p4pw\" (UID: \"498207e0-f869-46d9-bdc9-d83f44d02385\") " pod="openstack/nova-cell0-conductor-db-sync-4p4pw" Oct 05 21:14:30 crc kubenswrapper[4754]: I1005 21:14:30.878654 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/498207e0-f869-46d9-bdc9-d83f44d02385-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-4p4pw\" (UID: \"498207e0-f869-46d9-bdc9-d83f44d02385\") " pod="openstack/nova-cell0-conductor-db-sync-4p4pw" Oct 05 21:14:30 crc kubenswrapper[4754]: I1005 21:14:30.878744 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/498207e0-f869-46d9-bdc9-d83f44d02385-config-data\") pod \"nova-cell0-conductor-db-sync-4p4pw\" (UID: \"498207e0-f869-46d9-bdc9-d83f44d02385\") " pod="openstack/nova-cell0-conductor-db-sync-4p4pw" Oct 05 21:14:30 crc kubenswrapper[4754]: I1005 21:14:30.878770 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wsjjx\" (UniqueName: \"kubernetes.io/projected/498207e0-f869-46d9-bdc9-d83f44d02385-kube-api-access-wsjjx\") pod \"nova-cell0-conductor-db-sync-4p4pw\" (UID: \"498207e0-f869-46d9-bdc9-d83f44d02385\") " pod="openstack/nova-cell0-conductor-db-sync-4p4pw" Oct 05 21:14:30 crc kubenswrapper[4754]: I1005 21:14:30.893202 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/498207e0-f869-46d9-bdc9-d83f44d02385-scripts\") pod \"nova-cell0-conductor-db-sync-4p4pw\" (UID: \"498207e0-f869-46d9-bdc9-d83f44d02385\") " pod="openstack/nova-cell0-conductor-db-sync-4p4pw" Oct 05 21:14:30 crc kubenswrapper[4754]: I1005 21:14:30.894430 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/498207e0-f869-46d9-bdc9-d83f44d02385-config-data\") pod \"nova-cell0-conductor-db-sync-4p4pw\" (UID: \"498207e0-f869-46d9-bdc9-d83f44d02385\") " pod="openstack/nova-cell0-conductor-db-sync-4p4pw" Oct 05 21:14:30 crc kubenswrapper[4754]: I1005 21:14:30.900064 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/498207e0-f869-46d9-bdc9-d83f44d02385-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-4p4pw\" (UID: \"498207e0-f869-46d9-bdc9-d83f44d02385\") " pod="openstack/nova-cell0-conductor-db-sync-4p4pw" Oct 05 21:14:30 crc kubenswrapper[4754]: I1005 21:14:30.916050 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wsjjx\" (UniqueName: \"kubernetes.io/projected/498207e0-f869-46d9-bdc9-d83f44d02385-kube-api-access-wsjjx\") pod \"nova-cell0-conductor-db-sync-4p4pw\" (UID: \"498207e0-f869-46d9-bdc9-d83f44d02385\") " pod="openstack/nova-cell0-conductor-db-sync-4p4pw" Oct 05 21:14:30 crc kubenswrapper[4754]: I1005 21:14:30.971471 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-4p4pw" Oct 05 21:14:31 crc kubenswrapper[4754]: W1005 21:14:31.465725 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod498207e0_f869_46d9_bdc9_d83f44d02385.slice/crio-1a536c303aeabab78e83ed25bc2ffc966a59b83ecea5e576a53b49267d534869 WatchSource:0}: Error finding container 1a536c303aeabab78e83ed25bc2ffc966a59b83ecea5e576a53b49267d534869: Status 404 returned error can't find the container with id 1a536c303aeabab78e83ed25bc2ffc966a59b83ecea5e576a53b49267d534869 Oct 05 21:14:31 crc kubenswrapper[4754]: I1005 21:14:31.474426 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-4p4pw"] Oct 05 21:14:31 crc kubenswrapper[4754]: I1005 21:14:31.550770 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-4p4pw" event={"ID":"498207e0-f869-46d9-bdc9-d83f44d02385","Type":"ContainerStarted","Data":"1a536c303aeabab78e83ed25bc2ffc966a59b83ecea5e576a53b49267d534869"} Oct 05 21:14:33 crc kubenswrapper[4754]: I1005 21:14:33.001992 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 05 21:14:33 crc kubenswrapper[4754]: I1005 21:14:33.002727 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b15616c9-4fdf-4a35-a852-6aa8673b977a" containerName="sg-core" containerID="cri-o://ffdd06d32874aa55b407f0be95f91d5501e5a978d7fd31bb1e1bda4847180178" gracePeriod=30 Oct 05 21:14:33 crc kubenswrapper[4754]: I1005 21:14:33.002979 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b15616c9-4fdf-4a35-a852-6aa8673b977a" containerName="proxy-httpd" containerID="cri-o://8ecc956bd51b18d859bbc521c05ced28fe4286e7455e20544b6bb79c5302ce28" gracePeriod=30 Oct 05 21:14:33 crc kubenswrapper[4754]: I1005 21:14:33.003033 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b15616c9-4fdf-4a35-a852-6aa8673b977a" containerName="ceilometer-notification-agent" containerID="cri-o://8a4a5b78bf4a081dbf3dca51440ad48381da380a1fc7cdf875a12491e252c7ff" gracePeriod=30 Oct 05 21:14:33 crc kubenswrapper[4754]: I1005 21:14:33.002620 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b15616c9-4fdf-4a35-a852-6aa8673b977a" containerName="ceilometer-central-agent" containerID="cri-o://650df3e2f17985e98ecd8928c1e230b948813039ccfc6799727843635d02f07e" gracePeriod=30 Oct 05 21:14:33 crc kubenswrapper[4754]: I1005 21:14:33.016078 4754 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="b15616c9-4fdf-4a35-a852-6aa8673b977a" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.177:3000/\": read tcp 10.217.0.2:51652->10.217.0.177:3000: read: connection reset by peer" Oct 05 21:14:33 crc kubenswrapper[4754]: I1005 21:14:33.582573 4754 generic.go:334] "Generic (PLEG): container finished" podID="b15616c9-4fdf-4a35-a852-6aa8673b977a" containerID="8ecc956bd51b18d859bbc521c05ced28fe4286e7455e20544b6bb79c5302ce28" exitCode=0 Oct 05 21:14:33 crc kubenswrapper[4754]: I1005 21:14:33.582855 4754 generic.go:334] "Generic (PLEG): container finished" podID="b15616c9-4fdf-4a35-a852-6aa8673b977a" containerID="ffdd06d32874aa55b407f0be95f91d5501e5a978d7fd31bb1e1bda4847180178" exitCode=2 Oct 05 21:14:33 crc kubenswrapper[4754]: I1005 21:14:33.582864 4754 generic.go:334] "Generic (PLEG): container finished" podID="b15616c9-4fdf-4a35-a852-6aa8673b977a" containerID="650df3e2f17985e98ecd8928c1e230b948813039ccfc6799727843635d02f07e" exitCode=0 Oct 05 21:14:33 crc kubenswrapper[4754]: I1005 21:14:33.582883 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b15616c9-4fdf-4a35-a852-6aa8673b977a","Type":"ContainerDied","Data":"8ecc956bd51b18d859bbc521c05ced28fe4286e7455e20544b6bb79c5302ce28"} Oct 05 21:14:33 crc kubenswrapper[4754]: I1005 21:14:33.582928 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b15616c9-4fdf-4a35-a852-6aa8673b977a","Type":"ContainerDied","Data":"ffdd06d32874aa55b407f0be95f91d5501e5a978d7fd31bb1e1bda4847180178"} Oct 05 21:14:33 crc kubenswrapper[4754]: I1005 21:14:33.582940 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b15616c9-4fdf-4a35-a852-6aa8673b977a","Type":"ContainerDied","Data":"650df3e2f17985e98ecd8928c1e230b948813039ccfc6799727843635d02f07e"} Oct 05 21:14:37 crc kubenswrapper[4754]: I1005 21:14:37.629740 4754 generic.go:334] "Generic (PLEG): container finished" podID="b15616c9-4fdf-4a35-a852-6aa8673b977a" containerID="8a4a5b78bf4a081dbf3dca51440ad48381da380a1fc7cdf875a12491e252c7ff" exitCode=0 Oct 05 21:14:37 crc kubenswrapper[4754]: I1005 21:14:37.629778 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b15616c9-4fdf-4a35-a852-6aa8673b977a","Type":"ContainerDied","Data":"8a4a5b78bf4a081dbf3dca51440ad48381da380a1fc7cdf875a12491e252c7ff"} Oct 05 21:14:39 crc kubenswrapper[4754]: I1005 21:14:39.766277 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 05 21:14:39 crc kubenswrapper[4754]: I1005 21:14:39.893923 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8c7m8\" (UniqueName: \"kubernetes.io/projected/b15616c9-4fdf-4a35-a852-6aa8673b977a-kube-api-access-8c7m8\") pod \"b15616c9-4fdf-4a35-a852-6aa8673b977a\" (UID: \"b15616c9-4fdf-4a35-a852-6aa8673b977a\") " Oct 05 21:14:39 crc kubenswrapper[4754]: I1005 21:14:39.894848 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b15616c9-4fdf-4a35-a852-6aa8673b977a-run-httpd\") pod \"b15616c9-4fdf-4a35-a852-6aa8673b977a\" (UID: \"b15616c9-4fdf-4a35-a852-6aa8673b977a\") " Oct 05 21:14:39 crc kubenswrapper[4754]: I1005 21:14:39.895035 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b15616c9-4fdf-4a35-a852-6aa8673b977a-sg-core-conf-yaml\") pod \"b15616c9-4fdf-4a35-a852-6aa8673b977a\" (UID: \"b15616c9-4fdf-4a35-a852-6aa8673b977a\") " Oct 05 21:14:39 crc kubenswrapper[4754]: I1005 21:14:39.895135 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b15616c9-4fdf-4a35-a852-6aa8673b977a-config-data\") pod \"b15616c9-4fdf-4a35-a852-6aa8673b977a\" (UID: \"b15616c9-4fdf-4a35-a852-6aa8673b977a\") " Oct 05 21:14:39 crc kubenswrapper[4754]: I1005 21:14:39.895266 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b15616c9-4fdf-4a35-a852-6aa8673b977a-combined-ca-bundle\") pod \"b15616c9-4fdf-4a35-a852-6aa8673b977a\" (UID: \"b15616c9-4fdf-4a35-a852-6aa8673b977a\") " Oct 05 21:14:39 crc kubenswrapper[4754]: I1005 21:14:39.895440 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b15616c9-4fdf-4a35-a852-6aa8673b977a-log-httpd\") pod \"b15616c9-4fdf-4a35-a852-6aa8673b977a\" (UID: \"b15616c9-4fdf-4a35-a852-6aa8673b977a\") " Oct 05 21:14:39 crc kubenswrapper[4754]: I1005 21:14:39.895695 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/b15616c9-4fdf-4a35-a852-6aa8673b977a-ceilometer-tls-certs\") pod \"b15616c9-4fdf-4a35-a852-6aa8673b977a\" (UID: \"b15616c9-4fdf-4a35-a852-6aa8673b977a\") " Oct 05 21:14:39 crc kubenswrapper[4754]: I1005 21:14:39.896115 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b15616c9-4fdf-4a35-a852-6aa8673b977a-scripts\") pod \"b15616c9-4fdf-4a35-a852-6aa8673b977a\" (UID: \"b15616c9-4fdf-4a35-a852-6aa8673b977a\") " Oct 05 21:14:39 crc kubenswrapper[4754]: I1005 21:14:39.895202 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b15616c9-4fdf-4a35-a852-6aa8673b977a-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "b15616c9-4fdf-4a35-a852-6aa8673b977a" (UID: "b15616c9-4fdf-4a35-a852-6aa8673b977a"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 21:14:39 crc kubenswrapper[4754]: I1005 21:14:39.895990 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b15616c9-4fdf-4a35-a852-6aa8673b977a-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "b15616c9-4fdf-4a35-a852-6aa8673b977a" (UID: "b15616c9-4fdf-4a35-a852-6aa8673b977a"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 21:14:39 crc kubenswrapper[4754]: I1005 21:14:39.896991 4754 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b15616c9-4fdf-4a35-a852-6aa8673b977a-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 05 21:14:39 crc kubenswrapper[4754]: I1005 21:14:39.897098 4754 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b15616c9-4fdf-4a35-a852-6aa8673b977a-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 05 21:14:39 crc kubenswrapper[4754]: I1005 21:14:39.900397 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b15616c9-4fdf-4a35-a852-6aa8673b977a-kube-api-access-8c7m8" (OuterVolumeSpecName: "kube-api-access-8c7m8") pod "b15616c9-4fdf-4a35-a852-6aa8673b977a" (UID: "b15616c9-4fdf-4a35-a852-6aa8673b977a"). InnerVolumeSpecName "kube-api-access-8c7m8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:14:39 crc kubenswrapper[4754]: I1005 21:14:39.909739 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b15616c9-4fdf-4a35-a852-6aa8673b977a-scripts" (OuterVolumeSpecName: "scripts") pod "b15616c9-4fdf-4a35-a852-6aa8673b977a" (UID: "b15616c9-4fdf-4a35-a852-6aa8673b977a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:14:39 crc kubenswrapper[4754]: I1005 21:14:39.951506 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b15616c9-4fdf-4a35-a852-6aa8673b977a-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "b15616c9-4fdf-4a35-a852-6aa8673b977a" (UID: "b15616c9-4fdf-4a35-a852-6aa8673b977a"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:14:39 crc kubenswrapper[4754]: I1005 21:14:39.999482 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8c7m8\" (UniqueName: \"kubernetes.io/projected/b15616c9-4fdf-4a35-a852-6aa8673b977a-kube-api-access-8c7m8\") on node \"crc\" DevicePath \"\"" Oct 05 21:14:39 crc kubenswrapper[4754]: I1005 21:14:39.999530 4754 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b15616c9-4fdf-4a35-a852-6aa8673b977a-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 05 21:14:39 crc kubenswrapper[4754]: I1005 21:14:39.999540 4754 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b15616c9-4fdf-4a35-a852-6aa8673b977a-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 21:14:40 crc kubenswrapper[4754]: I1005 21:14:40.001093 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b15616c9-4fdf-4a35-a852-6aa8673b977a-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "b15616c9-4fdf-4a35-a852-6aa8673b977a" (UID: "b15616c9-4fdf-4a35-a852-6aa8673b977a"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:14:40 crc kubenswrapper[4754]: I1005 21:14:40.003679 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b15616c9-4fdf-4a35-a852-6aa8673b977a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b15616c9-4fdf-4a35-a852-6aa8673b977a" (UID: "b15616c9-4fdf-4a35-a852-6aa8673b977a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:14:40 crc kubenswrapper[4754]: I1005 21:14:40.084123 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b15616c9-4fdf-4a35-a852-6aa8673b977a-config-data" (OuterVolumeSpecName: "config-data") pod "b15616c9-4fdf-4a35-a852-6aa8673b977a" (UID: "b15616c9-4fdf-4a35-a852-6aa8673b977a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:14:40 crc kubenswrapper[4754]: I1005 21:14:40.101958 4754 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b15616c9-4fdf-4a35-a852-6aa8673b977a-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 21:14:40 crc kubenswrapper[4754]: I1005 21:14:40.101982 4754 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b15616c9-4fdf-4a35-a852-6aa8673b977a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 21:14:40 crc kubenswrapper[4754]: I1005 21:14:40.101993 4754 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/b15616c9-4fdf-4a35-a852-6aa8673b977a-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 05 21:14:40 crc kubenswrapper[4754]: I1005 21:14:40.656848 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b15616c9-4fdf-4a35-a852-6aa8673b977a","Type":"ContainerDied","Data":"7dec7105830944b78be8543663bcb16050e4734f652ded16797f7afc68d45c7e"} Oct 05 21:14:40 crc kubenswrapper[4754]: I1005 21:14:40.656901 4754 scope.go:117] "RemoveContainer" containerID="8ecc956bd51b18d859bbc521c05ced28fe4286e7455e20544b6bb79c5302ce28" Oct 05 21:14:40 crc kubenswrapper[4754]: I1005 21:14:40.657023 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 05 21:14:40 crc kubenswrapper[4754]: I1005 21:14:40.667733 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-4p4pw" event={"ID":"498207e0-f869-46d9-bdc9-d83f44d02385","Type":"ContainerStarted","Data":"5c0c3769cc7e996e252f750662bfc170350b169e66fd2eb2db56ef545fd81f57"} Oct 05 21:14:40 crc kubenswrapper[4754]: I1005 21:14:40.708749 4754 scope.go:117] "RemoveContainer" containerID="ffdd06d32874aa55b407f0be95f91d5501e5a978d7fd31bb1e1bda4847180178" Oct 05 21:14:40 crc kubenswrapper[4754]: I1005 21:14:40.735980 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-4p4pw" podStartSLOduration=2.33660109 podStartE2EDuration="10.735960276s" podCreationTimestamp="2025-10-05 21:14:30 +0000 UTC" firstStartedPulling="2025-10-05 21:14:31.468267798 +0000 UTC m=+1195.372386508" lastFinishedPulling="2025-10-05 21:14:39.867626974 +0000 UTC m=+1203.771745694" observedRunningTime="2025-10-05 21:14:40.717048498 +0000 UTC m=+1204.621167208" watchObservedRunningTime="2025-10-05 21:14:40.735960276 +0000 UTC m=+1204.640078986" Oct 05 21:14:40 crc kubenswrapper[4754]: I1005 21:14:40.741759 4754 scope.go:117] "RemoveContainer" containerID="8a4a5b78bf4a081dbf3dca51440ad48381da380a1fc7cdf875a12491e252c7ff" Oct 05 21:14:40 crc kubenswrapper[4754]: I1005 21:14:40.741909 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 05 21:14:40 crc kubenswrapper[4754]: I1005 21:14:40.791925 4754 scope.go:117] "RemoveContainer" containerID="650df3e2f17985e98ecd8928c1e230b948813039ccfc6799727843635d02f07e" Oct 05 21:14:40 crc kubenswrapper[4754]: I1005 21:14:40.818641 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 05 21:14:40 crc kubenswrapper[4754]: I1005 21:14:40.833199 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 05 21:14:40 crc kubenswrapper[4754]: E1005 21:14:40.834530 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b15616c9-4fdf-4a35-a852-6aa8673b977a" containerName="proxy-httpd" Oct 05 21:14:40 crc kubenswrapper[4754]: I1005 21:14:40.834552 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="b15616c9-4fdf-4a35-a852-6aa8673b977a" containerName="proxy-httpd" Oct 05 21:14:40 crc kubenswrapper[4754]: E1005 21:14:40.834601 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b15616c9-4fdf-4a35-a852-6aa8673b977a" containerName="sg-core" Oct 05 21:14:40 crc kubenswrapper[4754]: I1005 21:14:40.834608 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="b15616c9-4fdf-4a35-a852-6aa8673b977a" containerName="sg-core" Oct 05 21:14:40 crc kubenswrapper[4754]: E1005 21:14:40.834618 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b15616c9-4fdf-4a35-a852-6aa8673b977a" containerName="ceilometer-notification-agent" Oct 05 21:14:40 crc kubenswrapper[4754]: I1005 21:14:40.834624 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="b15616c9-4fdf-4a35-a852-6aa8673b977a" containerName="ceilometer-notification-agent" Oct 05 21:14:40 crc kubenswrapper[4754]: E1005 21:14:40.834634 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b15616c9-4fdf-4a35-a852-6aa8673b977a" containerName="ceilometer-central-agent" Oct 05 21:14:40 crc kubenswrapper[4754]: I1005 21:14:40.834640 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="b15616c9-4fdf-4a35-a852-6aa8673b977a" containerName="ceilometer-central-agent" Oct 05 21:14:40 crc kubenswrapper[4754]: I1005 21:14:40.834824 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="b15616c9-4fdf-4a35-a852-6aa8673b977a" containerName="ceilometer-notification-agent" Oct 05 21:14:40 crc kubenswrapper[4754]: I1005 21:14:40.834838 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="b15616c9-4fdf-4a35-a852-6aa8673b977a" containerName="sg-core" Oct 05 21:14:40 crc kubenswrapper[4754]: I1005 21:14:40.834855 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="b15616c9-4fdf-4a35-a852-6aa8673b977a" containerName="proxy-httpd" Oct 05 21:14:40 crc kubenswrapper[4754]: I1005 21:14:40.834870 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="b15616c9-4fdf-4a35-a852-6aa8673b977a" containerName="ceilometer-central-agent" Oct 05 21:14:40 crc kubenswrapper[4754]: I1005 21:14:40.840886 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 05 21:14:40 crc kubenswrapper[4754]: I1005 21:14:40.843826 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 05 21:14:40 crc kubenswrapper[4754]: I1005 21:14:40.844001 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 05 21:14:40 crc kubenswrapper[4754]: I1005 21:14:40.851473 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 05 21:14:40 crc kubenswrapper[4754]: I1005 21:14:40.866476 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b15616c9-4fdf-4a35-a852-6aa8673b977a" path="/var/lib/kubelet/pods/b15616c9-4fdf-4a35-a852-6aa8673b977a/volumes" Oct 05 21:14:40 crc kubenswrapper[4754]: I1005 21:14:40.867951 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 05 21:14:40 crc kubenswrapper[4754]: I1005 21:14:40.922575 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6s95q\" (UniqueName: \"kubernetes.io/projected/72a8e713-d71f-4339-a979-85ee309fa945-kube-api-access-6s95q\") pod \"ceilometer-0\" (UID: \"72a8e713-d71f-4339-a979-85ee309fa945\") " pod="openstack/ceilometer-0" Oct 05 21:14:40 crc kubenswrapper[4754]: I1005 21:14:40.922679 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/72a8e713-d71f-4339-a979-85ee309fa945-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"72a8e713-d71f-4339-a979-85ee309fa945\") " pod="openstack/ceilometer-0" Oct 05 21:14:40 crc kubenswrapper[4754]: I1005 21:14:40.922701 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72a8e713-d71f-4339-a979-85ee309fa945-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"72a8e713-d71f-4339-a979-85ee309fa945\") " pod="openstack/ceilometer-0" Oct 05 21:14:40 crc kubenswrapper[4754]: I1005 21:14:40.922720 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/72a8e713-d71f-4339-a979-85ee309fa945-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"72a8e713-d71f-4339-a979-85ee309fa945\") " pod="openstack/ceilometer-0" Oct 05 21:14:40 crc kubenswrapper[4754]: I1005 21:14:40.922768 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/72a8e713-d71f-4339-a979-85ee309fa945-scripts\") pod \"ceilometer-0\" (UID: \"72a8e713-d71f-4339-a979-85ee309fa945\") " pod="openstack/ceilometer-0" Oct 05 21:14:40 crc kubenswrapper[4754]: I1005 21:14:40.922792 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72a8e713-d71f-4339-a979-85ee309fa945-config-data\") pod \"ceilometer-0\" (UID: \"72a8e713-d71f-4339-a979-85ee309fa945\") " pod="openstack/ceilometer-0" Oct 05 21:14:40 crc kubenswrapper[4754]: I1005 21:14:40.922981 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/72a8e713-d71f-4339-a979-85ee309fa945-run-httpd\") pod \"ceilometer-0\" (UID: \"72a8e713-d71f-4339-a979-85ee309fa945\") " pod="openstack/ceilometer-0" Oct 05 21:14:40 crc kubenswrapper[4754]: I1005 21:14:40.923111 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/72a8e713-d71f-4339-a979-85ee309fa945-log-httpd\") pod \"ceilometer-0\" (UID: \"72a8e713-d71f-4339-a979-85ee309fa945\") " pod="openstack/ceilometer-0" Oct 05 21:14:40 crc kubenswrapper[4754]: E1005 21:14:40.926094 4754 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb15616c9_4fdf_4a35_a852_6aa8673b977a.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb15616c9_4fdf_4a35_a852_6aa8673b977a.slice/crio-7dec7105830944b78be8543663bcb16050e4734f652ded16797f7afc68d45c7e\": RecentStats: unable to find data in memory cache]" Oct 05 21:14:41 crc kubenswrapper[4754]: I1005 21:14:41.024509 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/72a8e713-d71f-4339-a979-85ee309fa945-scripts\") pod \"ceilometer-0\" (UID: \"72a8e713-d71f-4339-a979-85ee309fa945\") " pod="openstack/ceilometer-0" Oct 05 21:14:41 crc kubenswrapper[4754]: I1005 21:14:41.024809 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72a8e713-d71f-4339-a979-85ee309fa945-config-data\") pod \"ceilometer-0\" (UID: \"72a8e713-d71f-4339-a979-85ee309fa945\") " pod="openstack/ceilometer-0" Oct 05 21:14:41 crc kubenswrapper[4754]: I1005 21:14:41.024857 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/72a8e713-d71f-4339-a979-85ee309fa945-run-httpd\") pod \"ceilometer-0\" (UID: \"72a8e713-d71f-4339-a979-85ee309fa945\") " pod="openstack/ceilometer-0" Oct 05 21:14:41 crc kubenswrapper[4754]: I1005 21:14:41.024899 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/72a8e713-d71f-4339-a979-85ee309fa945-log-httpd\") pod \"ceilometer-0\" (UID: \"72a8e713-d71f-4339-a979-85ee309fa945\") " pod="openstack/ceilometer-0" Oct 05 21:14:41 crc kubenswrapper[4754]: I1005 21:14:41.024948 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6s95q\" (UniqueName: \"kubernetes.io/projected/72a8e713-d71f-4339-a979-85ee309fa945-kube-api-access-6s95q\") pod \"ceilometer-0\" (UID: \"72a8e713-d71f-4339-a979-85ee309fa945\") " pod="openstack/ceilometer-0" Oct 05 21:14:41 crc kubenswrapper[4754]: I1005 21:14:41.025020 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/72a8e713-d71f-4339-a979-85ee309fa945-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"72a8e713-d71f-4339-a979-85ee309fa945\") " pod="openstack/ceilometer-0" Oct 05 21:14:41 crc kubenswrapper[4754]: I1005 21:14:41.025040 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72a8e713-d71f-4339-a979-85ee309fa945-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"72a8e713-d71f-4339-a979-85ee309fa945\") " pod="openstack/ceilometer-0" Oct 05 21:14:41 crc kubenswrapper[4754]: I1005 21:14:41.025058 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/72a8e713-d71f-4339-a979-85ee309fa945-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"72a8e713-d71f-4339-a979-85ee309fa945\") " pod="openstack/ceilometer-0" Oct 05 21:14:41 crc kubenswrapper[4754]: I1005 21:14:41.025313 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/72a8e713-d71f-4339-a979-85ee309fa945-run-httpd\") pod \"ceilometer-0\" (UID: \"72a8e713-d71f-4339-a979-85ee309fa945\") " pod="openstack/ceilometer-0" Oct 05 21:14:41 crc kubenswrapper[4754]: I1005 21:14:41.025536 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/72a8e713-d71f-4339-a979-85ee309fa945-log-httpd\") pod \"ceilometer-0\" (UID: \"72a8e713-d71f-4339-a979-85ee309fa945\") " pod="openstack/ceilometer-0" Oct 05 21:14:41 crc kubenswrapper[4754]: I1005 21:14:41.030092 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/72a8e713-d71f-4339-a979-85ee309fa945-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"72a8e713-d71f-4339-a979-85ee309fa945\") " pod="openstack/ceilometer-0" Oct 05 21:14:41 crc kubenswrapper[4754]: I1005 21:14:41.030474 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/72a8e713-d71f-4339-a979-85ee309fa945-scripts\") pod \"ceilometer-0\" (UID: \"72a8e713-d71f-4339-a979-85ee309fa945\") " pod="openstack/ceilometer-0" Oct 05 21:14:41 crc kubenswrapper[4754]: I1005 21:14:41.030812 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72a8e713-d71f-4339-a979-85ee309fa945-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"72a8e713-d71f-4339-a979-85ee309fa945\") " pod="openstack/ceilometer-0" Oct 05 21:14:41 crc kubenswrapper[4754]: I1005 21:14:41.031517 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/72a8e713-d71f-4339-a979-85ee309fa945-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"72a8e713-d71f-4339-a979-85ee309fa945\") " pod="openstack/ceilometer-0" Oct 05 21:14:41 crc kubenswrapper[4754]: I1005 21:14:41.032472 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72a8e713-d71f-4339-a979-85ee309fa945-config-data\") pod \"ceilometer-0\" (UID: \"72a8e713-d71f-4339-a979-85ee309fa945\") " pod="openstack/ceilometer-0" Oct 05 21:14:41 crc kubenswrapper[4754]: I1005 21:14:41.050481 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6s95q\" (UniqueName: \"kubernetes.io/projected/72a8e713-d71f-4339-a979-85ee309fa945-kube-api-access-6s95q\") pod \"ceilometer-0\" (UID: \"72a8e713-d71f-4339-a979-85ee309fa945\") " pod="openstack/ceilometer-0" Oct 05 21:14:41 crc kubenswrapper[4754]: I1005 21:14:41.170543 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 05 21:14:41 crc kubenswrapper[4754]: I1005 21:14:41.604400 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 05 21:14:41 crc kubenswrapper[4754]: W1005 21:14:41.611224 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod72a8e713_d71f_4339_a979_85ee309fa945.slice/crio-51c78451914e10b01ed0cc474d684e93b47a0baacc67214692680472f7ccafd1 WatchSource:0}: Error finding container 51c78451914e10b01ed0cc474d684e93b47a0baacc67214692680472f7ccafd1: Status 404 returned error can't find the container with id 51c78451914e10b01ed0cc474d684e93b47a0baacc67214692680472f7ccafd1 Oct 05 21:14:41 crc kubenswrapper[4754]: I1005 21:14:41.614916 4754 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 05 21:14:41 crc kubenswrapper[4754]: I1005 21:14:41.685098 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"72a8e713-d71f-4339-a979-85ee309fa945","Type":"ContainerStarted","Data":"51c78451914e10b01ed0cc474d684e93b47a0baacc67214692680472f7ccafd1"} Oct 05 21:14:42 crc kubenswrapper[4754]: I1005 21:14:42.694605 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"72a8e713-d71f-4339-a979-85ee309fa945","Type":"ContainerStarted","Data":"77b01d1c78cdf77197d46e1eee58b93c48eab4fc46e38ef811c078cc4430111b"} Oct 05 21:14:43 crc kubenswrapper[4754]: I1005 21:14:43.053742 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 05 21:14:43 crc kubenswrapper[4754]: I1005 21:14:43.719112 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"72a8e713-d71f-4339-a979-85ee309fa945","Type":"ContainerStarted","Data":"c8bddf548bd07ad5a695cef4111f7c0e3a0dc07ec73040509fd14caf0a13e927"} Oct 05 21:14:44 crc kubenswrapper[4754]: I1005 21:14:44.729536 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"72a8e713-d71f-4339-a979-85ee309fa945","Type":"ContainerStarted","Data":"20142dba9bbead06a2cfa5a52cc693e2678ebe49a108996dc3c3bcee40b7019b"} Oct 05 21:14:46 crc kubenswrapper[4754]: I1005 21:14:46.753277 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"72a8e713-d71f-4339-a979-85ee309fa945","Type":"ContainerStarted","Data":"8b3b98fc3e3153b054ae575c21c671d73960d7d8303ae8963ee2933a1fe4d40b"} Oct 05 21:14:46 crc kubenswrapper[4754]: I1005 21:14:46.753724 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="72a8e713-d71f-4339-a979-85ee309fa945" containerName="ceilometer-central-agent" containerID="cri-o://77b01d1c78cdf77197d46e1eee58b93c48eab4fc46e38ef811c078cc4430111b" gracePeriod=30 Oct 05 21:14:46 crc kubenswrapper[4754]: I1005 21:14:46.754050 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 05 21:14:46 crc kubenswrapper[4754]: I1005 21:14:46.754097 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="72a8e713-d71f-4339-a979-85ee309fa945" containerName="proxy-httpd" containerID="cri-o://8b3b98fc3e3153b054ae575c21c671d73960d7d8303ae8963ee2933a1fe4d40b" gracePeriod=30 Oct 05 21:14:46 crc kubenswrapper[4754]: I1005 21:14:46.754187 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="72a8e713-d71f-4339-a979-85ee309fa945" containerName="sg-core" containerID="cri-o://20142dba9bbead06a2cfa5a52cc693e2678ebe49a108996dc3c3bcee40b7019b" gracePeriod=30 Oct 05 21:14:46 crc kubenswrapper[4754]: I1005 21:14:46.754245 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="72a8e713-d71f-4339-a979-85ee309fa945" containerName="ceilometer-notification-agent" containerID="cri-o://c8bddf548bd07ad5a695cef4111f7c0e3a0dc07ec73040509fd14caf0a13e927" gracePeriod=30 Oct 05 21:14:46 crc kubenswrapper[4754]: I1005 21:14:46.780309 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.343609916 podStartE2EDuration="6.780291239s" podCreationTimestamp="2025-10-05 21:14:40 +0000 UTC" firstStartedPulling="2025-10-05 21:14:41.614716198 +0000 UTC m=+1205.518834908" lastFinishedPulling="2025-10-05 21:14:46.051397521 +0000 UTC m=+1209.955516231" observedRunningTime="2025-10-05 21:14:46.778891061 +0000 UTC m=+1210.683009771" watchObservedRunningTime="2025-10-05 21:14:46.780291239 +0000 UTC m=+1210.684409949" Oct 05 21:14:47 crc kubenswrapper[4754]: I1005 21:14:47.767857 4754 generic.go:334] "Generic (PLEG): container finished" podID="72a8e713-d71f-4339-a979-85ee309fa945" containerID="8b3b98fc3e3153b054ae575c21c671d73960d7d8303ae8963ee2933a1fe4d40b" exitCode=0 Oct 05 21:14:47 crc kubenswrapper[4754]: I1005 21:14:47.767894 4754 generic.go:334] "Generic (PLEG): container finished" podID="72a8e713-d71f-4339-a979-85ee309fa945" containerID="20142dba9bbead06a2cfa5a52cc693e2678ebe49a108996dc3c3bcee40b7019b" exitCode=2 Oct 05 21:14:47 crc kubenswrapper[4754]: I1005 21:14:47.767903 4754 generic.go:334] "Generic (PLEG): container finished" podID="72a8e713-d71f-4339-a979-85ee309fa945" containerID="c8bddf548bd07ad5a695cef4111f7c0e3a0dc07ec73040509fd14caf0a13e927" exitCode=0 Oct 05 21:14:47 crc kubenswrapper[4754]: I1005 21:14:47.767938 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"72a8e713-d71f-4339-a979-85ee309fa945","Type":"ContainerDied","Data":"8b3b98fc3e3153b054ae575c21c671d73960d7d8303ae8963ee2933a1fe4d40b"} Oct 05 21:14:47 crc kubenswrapper[4754]: I1005 21:14:47.767970 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"72a8e713-d71f-4339-a979-85ee309fa945","Type":"ContainerDied","Data":"20142dba9bbead06a2cfa5a52cc693e2678ebe49a108996dc3c3bcee40b7019b"} Oct 05 21:14:47 crc kubenswrapper[4754]: I1005 21:14:47.767981 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"72a8e713-d71f-4339-a979-85ee309fa945","Type":"ContainerDied","Data":"c8bddf548bd07ad5a695cef4111f7c0e3a0dc07ec73040509fd14caf0a13e927"} Oct 05 21:14:50 crc kubenswrapper[4754]: I1005 21:14:50.821358 4754 generic.go:334] "Generic (PLEG): container finished" podID="72a8e713-d71f-4339-a979-85ee309fa945" containerID="77b01d1c78cdf77197d46e1eee58b93c48eab4fc46e38ef811c078cc4430111b" exitCode=0 Oct 05 21:14:50 crc kubenswrapper[4754]: I1005 21:14:50.821953 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"72a8e713-d71f-4339-a979-85ee309fa945","Type":"ContainerDied","Data":"77b01d1c78cdf77197d46e1eee58b93c48eab4fc46e38ef811c078cc4430111b"} Oct 05 21:14:50 crc kubenswrapper[4754]: I1005 21:14:50.891189 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 05 21:14:51 crc kubenswrapper[4754]: I1005 21:14:51.023575 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72a8e713-d71f-4339-a979-85ee309fa945-config-data\") pod \"72a8e713-d71f-4339-a979-85ee309fa945\" (UID: \"72a8e713-d71f-4339-a979-85ee309fa945\") " Oct 05 21:14:51 crc kubenswrapper[4754]: I1005 21:14:51.023631 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6s95q\" (UniqueName: \"kubernetes.io/projected/72a8e713-d71f-4339-a979-85ee309fa945-kube-api-access-6s95q\") pod \"72a8e713-d71f-4339-a979-85ee309fa945\" (UID: \"72a8e713-d71f-4339-a979-85ee309fa945\") " Oct 05 21:14:51 crc kubenswrapper[4754]: I1005 21:14:51.023701 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/72a8e713-d71f-4339-a979-85ee309fa945-scripts\") pod \"72a8e713-d71f-4339-a979-85ee309fa945\" (UID: \"72a8e713-d71f-4339-a979-85ee309fa945\") " Oct 05 21:14:51 crc kubenswrapper[4754]: I1005 21:14:51.023737 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/72a8e713-d71f-4339-a979-85ee309fa945-sg-core-conf-yaml\") pod \"72a8e713-d71f-4339-a979-85ee309fa945\" (UID: \"72a8e713-d71f-4339-a979-85ee309fa945\") " Oct 05 21:14:51 crc kubenswrapper[4754]: I1005 21:14:51.023764 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/72a8e713-d71f-4339-a979-85ee309fa945-run-httpd\") pod \"72a8e713-d71f-4339-a979-85ee309fa945\" (UID: \"72a8e713-d71f-4339-a979-85ee309fa945\") " Oct 05 21:14:51 crc kubenswrapper[4754]: I1005 21:14:51.023827 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/72a8e713-d71f-4339-a979-85ee309fa945-ceilometer-tls-certs\") pod \"72a8e713-d71f-4339-a979-85ee309fa945\" (UID: \"72a8e713-d71f-4339-a979-85ee309fa945\") " Oct 05 21:14:51 crc kubenswrapper[4754]: I1005 21:14:51.023857 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72a8e713-d71f-4339-a979-85ee309fa945-combined-ca-bundle\") pod \"72a8e713-d71f-4339-a979-85ee309fa945\" (UID: \"72a8e713-d71f-4339-a979-85ee309fa945\") " Oct 05 21:14:51 crc kubenswrapper[4754]: I1005 21:14:51.023909 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/72a8e713-d71f-4339-a979-85ee309fa945-log-httpd\") pod \"72a8e713-d71f-4339-a979-85ee309fa945\" (UID: \"72a8e713-d71f-4339-a979-85ee309fa945\") " Oct 05 21:14:51 crc kubenswrapper[4754]: I1005 21:14:51.025169 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/72a8e713-d71f-4339-a979-85ee309fa945-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "72a8e713-d71f-4339-a979-85ee309fa945" (UID: "72a8e713-d71f-4339-a979-85ee309fa945"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 21:14:51 crc kubenswrapper[4754]: I1005 21:14:51.025556 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/72a8e713-d71f-4339-a979-85ee309fa945-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "72a8e713-d71f-4339-a979-85ee309fa945" (UID: "72a8e713-d71f-4339-a979-85ee309fa945"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 21:14:51 crc kubenswrapper[4754]: I1005 21:14:51.047526 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/72a8e713-d71f-4339-a979-85ee309fa945-scripts" (OuterVolumeSpecName: "scripts") pod "72a8e713-d71f-4339-a979-85ee309fa945" (UID: "72a8e713-d71f-4339-a979-85ee309fa945"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:14:51 crc kubenswrapper[4754]: I1005 21:14:51.049309 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/72a8e713-d71f-4339-a979-85ee309fa945-kube-api-access-6s95q" (OuterVolumeSpecName: "kube-api-access-6s95q") pod "72a8e713-d71f-4339-a979-85ee309fa945" (UID: "72a8e713-d71f-4339-a979-85ee309fa945"). InnerVolumeSpecName "kube-api-access-6s95q". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:14:51 crc kubenswrapper[4754]: I1005 21:14:51.058094 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/72a8e713-d71f-4339-a979-85ee309fa945-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "72a8e713-d71f-4339-a979-85ee309fa945" (UID: "72a8e713-d71f-4339-a979-85ee309fa945"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:14:51 crc kubenswrapper[4754]: I1005 21:14:51.102750 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/72a8e713-d71f-4339-a979-85ee309fa945-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "72a8e713-d71f-4339-a979-85ee309fa945" (UID: "72a8e713-d71f-4339-a979-85ee309fa945"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:14:51 crc kubenswrapper[4754]: I1005 21:14:51.125860 4754 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/72a8e713-d71f-4339-a979-85ee309fa945-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 05 21:14:51 crc kubenswrapper[4754]: I1005 21:14:51.125909 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6s95q\" (UniqueName: \"kubernetes.io/projected/72a8e713-d71f-4339-a979-85ee309fa945-kube-api-access-6s95q\") on node \"crc\" DevicePath \"\"" Oct 05 21:14:51 crc kubenswrapper[4754]: I1005 21:14:51.125925 4754 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/72a8e713-d71f-4339-a979-85ee309fa945-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 21:14:51 crc kubenswrapper[4754]: I1005 21:14:51.125937 4754 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/72a8e713-d71f-4339-a979-85ee309fa945-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 05 21:14:51 crc kubenswrapper[4754]: I1005 21:14:51.125952 4754 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/72a8e713-d71f-4339-a979-85ee309fa945-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 05 21:14:51 crc kubenswrapper[4754]: I1005 21:14:51.125965 4754 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/72a8e713-d71f-4339-a979-85ee309fa945-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 05 21:14:51 crc kubenswrapper[4754]: I1005 21:14:51.146098 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/72a8e713-d71f-4339-a979-85ee309fa945-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "72a8e713-d71f-4339-a979-85ee309fa945" (UID: "72a8e713-d71f-4339-a979-85ee309fa945"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:14:51 crc kubenswrapper[4754]: I1005 21:14:51.174062 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/72a8e713-d71f-4339-a979-85ee309fa945-config-data" (OuterVolumeSpecName: "config-data") pod "72a8e713-d71f-4339-a979-85ee309fa945" (UID: "72a8e713-d71f-4339-a979-85ee309fa945"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:14:51 crc kubenswrapper[4754]: I1005 21:14:51.227532 4754 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72a8e713-d71f-4339-a979-85ee309fa945-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 21:14:51 crc kubenswrapper[4754]: I1005 21:14:51.227568 4754 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72a8e713-d71f-4339-a979-85ee309fa945-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 21:14:51 crc kubenswrapper[4754]: I1005 21:14:51.833382 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"72a8e713-d71f-4339-a979-85ee309fa945","Type":"ContainerDied","Data":"51c78451914e10b01ed0cc474d684e93b47a0baacc67214692680472f7ccafd1"} Oct 05 21:14:51 crc kubenswrapper[4754]: I1005 21:14:51.833783 4754 scope.go:117] "RemoveContainer" containerID="8b3b98fc3e3153b054ae575c21c671d73960d7d8303ae8963ee2933a1fe4d40b" Oct 05 21:14:51 crc kubenswrapper[4754]: I1005 21:14:51.834025 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 05 21:14:51 crc kubenswrapper[4754]: I1005 21:14:51.889340 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 05 21:14:51 crc kubenswrapper[4754]: I1005 21:14:51.889691 4754 scope.go:117] "RemoveContainer" containerID="20142dba9bbead06a2cfa5a52cc693e2678ebe49a108996dc3c3bcee40b7019b" Oct 05 21:14:51 crc kubenswrapper[4754]: I1005 21:14:51.907577 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 05 21:14:51 crc kubenswrapper[4754]: I1005 21:14:51.912744 4754 scope.go:117] "RemoveContainer" containerID="c8bddf548bd07ad5a695cef4111f7c0e3a0dc07ec73040509fd14caf0a13e927" Oct 05 21:14:51 crc kubenswrapper[4754]: I1005 21:14:51.917563 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 05 21:14:51 crc kubenswrapper[4754]: E1005 21:14:51.918039 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72a8e713-d71f-4339-a979-85ee309fa945" containerName="proxy-httpd" Oct 05 21:14:51 crc kubenswrapper[4754]: I1005 21:14:51.918058 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="72a8e713-d71f-4339-a979-85ee309fa945" containerName="proxy-httpd" Oct 05 21:14:51 crc kubenswrapper[4754]: E1005 21:14:51.918078 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72a8e713-d71f-4339-a979-85ee309fa945" containerName="ceilometer-central-agent" Oct 05 21:14:51 crc kubenswrapper[4754]: I1005 21:14:51.918084 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="72a8e713-d71f-4339-a979-85ee309fa945" containerName="ceilometer-central-agent" Oct 05 21:14:51 crc kubenswrapper[4754]: E1005 21:14:51.918099 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72a8e713-d71f-4339-a979-85ee309fa945" containerName="ceilometer-notification-agent" Oct 05 21:14:51 crc kubenswrapper[4754]: I1005 21:14:51.918106 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="72a8e713-d71f-4339-a979-85ee309fa945" containerName="ceilometer-notification-agent" Oct 05 21:14:51 crc kubenswrapper[4754]: E1005 21:14:51.918124 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72a8e713-d71f-4339-a979-85ee309fa945" containerName="sg-core" Oct 05 21:14:51 crc kubenswrapper[4754]: I1005 21:14:51.918132 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="72a8e713-d71f-4339-a979-85ee309fa945" containerName="sg-core" Oct 05 21:14:51 crc kubenswrapper[4754]: I1005 21:14:51.918315 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="72a8e713-d71f-4339-a979-85ee309fa945" containerName="ceilometer-central-agent" Oct 05 21:14:51 crc kubenswrapper[4754]: I1005 21:14:51.918326 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="72a8e713-d71f-4339-a979-85ee309fa945" containerName="proxy-httpd" Oct 05 21:14:51 crc kubenswrapper[4754]: I1005 21:14:51.918340 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="72a8e713-d71f-4339-a979-85ee309fa945" containerName="sg-core" Oct 05 21:14:51 crc kubenswrapper[4754]: I1005 21:14:51.918349 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="72a8e713-d71f-4339-a979-85ee309fa945" containerName="ceilometer-notification-agent" Oct 05 21:14:51 crc kubenswrapper[4754]: I1005 21:14:51.920560 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 05 21:14:51 crc kubenswrapper[4754]: I1005 21:14:51.923815 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 05 21:14:51 crc kubenswrapper[4754]: I1005 21:14:51.924014 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 05 21:14:51 crc kubenswrapper[4754]: I1005 21:14:51.925809 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 05 21:14:51 crc kubenswrapper[4754]: I1005 21:14:51.946422 4754 scope.go:117] "RemoveContainer" containerID="77b01d1c78cdf77197d46e1eee58b93c48eab4fc46e38ef811c078cc4430111b" Oct 05 21:14:51 crc kubenswrapper[4754]: I1005 21:14:51.949066 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 05 21:14:52 crc kubenswrapper[4754]: I1005 21:14:52.045792 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6ee3b3b2-07e1-4bab-9372-893eef07f250-run-httpd\") pod \"ceilometer-0\" (UID: \"6ee3b3b2-07e1-4bab-9372-893eef07f250\") " pod="openstack/ceilometer-0" Oct 05 21:14:52 crc kubenswrapper[4754]: I1005 21:14:52.045854 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6ee3b3b2-07e1-4bab-9372-893eef07f250-scripts\") pod \"ceilometer-0\" (UID: \"6ee3b3b2-07e1-4bab-9372-893eef07f250\") " pod="openstack/ceilometer-0" Oct 05 21:14:52 crc kubenswrapper[4754]: I1005 21:14:52.045917 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/6ee3b3b2-07e1-4bab-9372-893eef07f250-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"6ee3b3b2-07e1-4bab-9372-893eef07f250\") " pod="openstack/ceilometer-0" Oct 05 21:14:52 crc kubenswrapper[4754]: I1005 21:14:52.046096 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-94gmn\" (UniqueName: \"kubernetes.io/projected/6ee3b3b2-07e1-4bab-9372-893eef07f250-kube-api-access-94gmn\") pod \"ceilometer-0\" (UID: \"6ee3b3b2-07e1-4bab-9372-893eef07f250\") " pod="openstack/ceilometer-0" Oct 05 21:14:52 crc kubenswrapper[4754]: I1005 21:14:52.046276 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6ee3b3b2-07e1-4bab-9372-893eef07f250-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"6ee3b3b2-07e1-4bab-9372-893eef07f250\") " pod="openstack/ceilometer-0" Oct 05 21:14:52 crc kubenswrapper[4754]: I1005 21:14:52.046320 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6ee3b3b2-07e1-4bab-9372-893eef07f250-config-data\") pod \"ceilometer-0\" (UID: \"6ee3b3b2-07e1-4bab-9372-893eef07f250\") " pod="openstack/ceilometer-0" Oct 05 21:14:52 crc kubenswrapper[4754]: I1005 21:14:52.046371 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6ee3b3b2-07e1-4bab-9372-893eef07f250-log-httpd\") pod \"ceilometer-0\" (UID: \"6ee3b3b2-07e1-4bab-9372-893eef07f250\") " pod="openstack/ceilometer-0" Oct 05 21:14:52 crc kubenswrapper[4754]: I1005 21:14:52.046520 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ee3b3b2-07e1-4bab-9372-893eef07f250-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"6ee3b3b2-07e1-4bab-9372-893eef07f250\") " pod="openstack/ceilometer-0" Oct 05 21:14:52 crc kubenswrapper[4754]: I1005 21:14:52.147786 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/6ee3b3b2-07e1-4bab-9372-893eef07f250-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"6ee3b3b2-07e1-4bab-9372-893eef07f250\") " pod="openstack/ceilometer-0" Oct 05 21:14:52 crc kubenswrapper[4754]: I1005 21:14:52.148137 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-94gmn\" (UniqueName: \"kubernetes.io/projected/6ee3b3b2-07e1-4bab-9372-893eef07f250-kube-api-access-94gmn\") pod \"ceilometer-0\" (UID: \"6ee3b3b2-07e1-4bab-9372-893eef07f250\") " pod="openstack/ceilometer-0" Oct 05 21:14:52 crc kubenswrapper[4754]: I1005 21:14:52.148282 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6ee3b3b2-07e1-4bab-9372-893eef07f250-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"6ee3b3b2-07e1-4bab-9372-893eef07f250\") " pod="openstack/ceilometer-0" Oct 05 21:14:52 crc kubenswrapper[4754]: I1005 21:14:52.148375 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6ee3b3b2-07e1-4bab-9372-893eef07f250-config-data\") pod \"ceilometer-0\" (UID: \"6ee3b3b2-07e1-4bab-9372-893eef07f250\") " pod="openstack/ceilometer-0" Oct 05 21:14:52 crc kubenswrapper[4754]: I1005 21:14:52.148460 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6ee3b3b2-07e1-4bab-9372-893eef07f250-log-httpd\") pod \"ceilometer-0\" (UID: \"6ee3b3b2-07e1-4bab-9372-893eef07f250\") " pod="openstack/ceilometer-0" Oct 05 21:14:52 crc kubenswrapper[4754]: I1005 21:14:52.148577 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ee3b3b2-07e1-4bab-9372-893eef07f250-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"6ee3b3b2-07e1-4bab-9372-893eef07f250\") " pod="openstack/ceilometer-0" Oct 05 21:14:52 crc kubenswrapper[4754]: I1005 21:14:52.148699 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6ee3b3b2-07e1-4bab-9372-893eef07f250-run-httpd\") pod \"ceilometer-0\" (UID: \"6ee3b3b2-07e1-4bab-9372-893eef07f250\") " pod="openstack/ceilometer-0" Oct 05 21:14:52 crc kubenswrapper[4754]: I1005 21:14:52.148803 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6ee3b3b2-07e1-4bab-9372-893eef07f250-scripts\") pod \"ceilometer-0\" (UID: \"6ee3b3b2-07e1-4bab-9372-893eef07f250\") " pod="openstack/ceilometer-0" Oct 05 21:14:52 crc kubenswrapper[4754]: I1005 21:14:52.150081 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6ee3b3b2-07e1-4bab-9372-893eef07f250-log-httpd\") pod \"ceilometer-0\" (UID: \"6ee3b3b2-07e1-4bab-9372-893eef07f250\") " pod="openstack/ceilometer-0" Oct 05 21:14:52 crc kubenswrapper[4754]: I1005 21:14:52.150206 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6ee3b3b2-07e1-4bab-9372-893eef07f250-run-httpd\") pod \"ceilometer-0\" (UID: \"6ee3b3b2-07e1-4bab-9372-893eef07f250\") " pod="openstack/ceilometer-0" Oct 05 21:14:52 crc kubenswrapper[4754]: I1005 21:14:52.155090 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6ee3b3b2-07e1-4bab-9372-893eef07f250-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"6ee3b3b2-07e1-4bab-9372-893eef07f250\") " pod="openstack/ceilometer-0" Oct 05 21:14:52 crc kubenswrapper[4754]: I1005 21:14:52.155143 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ee3b3b2-07e1-4bab-9372-893eef07f250-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"6ee3b3b2-07e1-4bab-9372-893eef07f250\") " pod="openstack/ceilometer-0" Oct 05 21:14:52 crc kubenswrapper[4754]: I1005 21:14:52.156330 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6ee3b3b2-07e1-4bab-9372-893eef07f250-scripts\") pod \"ceilometer-0\" (UID: \"6ee3b3b2-07e1-4bab-9372-893eef07f250\") " pod="openstack/ceilometer-0" Oct 05 21:14:52 crc kubenswrapper[4754]: I1005 21:14:52.161517 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/6ee3b3b2-07e1-4bab-9372-893eef07f250-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"6ee3b3b2-07e1-4bab-9372-893eef07f250\") " pod="openstack/ceilometer-0" Oct 05 21:14:52 crc kubenswrapper[4754]: I1005 21:14:52.167543 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6ee3b3b2-07e1-4bab-9372-893eef07f250-config-data\") pod \"ceilometer-0\" (UID: \"6ee3b3b2-07e1-4bab-9372-893eef07f250\") " pod="openstack/ceilometer-0" Oct 05 21:14:52 crc kubenswrapper[4754]: I1005 21:14:52.167996 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-94gmn\" (UniqueName: \"kubernetes.io/projected/6ee3b3b2-07e1-4bab-9372-893eef07f250-kube-api-access-94gmn\") pod \"ceilometer-0\" (UID: \"6ee3b3b2-07e1-4bab-9372-893eef07f250\") " pod="openstack/ceilometer-0" Oct 05 21:14:52 crc kubenswrapper[4754]: I1005 21:14:52.247995 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 05 21:14:52 crc kubenswrapper[4754]: I1005 21:14:52.768157 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 05 21:14:52 crc kubenswrapper[4754]: I1005 21:14:52.852224 4754 generic.go:334] "Generic (PLEG): container finished" podID="498207e0-f869-46d9-bdc9-d83f44d02385" containerID="5c0c3769cc7e996e252f750662bfc170350b169e66fd2eb2db56ef545fd81f57" exitCode=0 Oct 05 21:14:52 crc kubenswrapper[4754]: I1005 21:14:52.857018 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="72a8e713-d71f-4339-a979-85ee309fa945" path="/var/lib/kubelet/pods/72a8e713-d71f-4339-a979-85ee309fa945/volumes" Oct 05 21:14:52 crc kubenswrapper[4754]: I1005 21:14:52.858069 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-4p4pw" event={"ID":"498207e0-f869-46d9-bdc9-d83f44d02385","Type":"ContainerDied","Data":"5c0c3769cc7e996e252f750662bfc170350b169e66fd2eb2db56ef545fd81f57"} Oct 05 21:14:52 crc kubenswrapper[4754]: I1005 21:14:52.858124 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6ee3b3b2-07e1-4bab-9372-893eef07f250","Type":"ContainerStarted","Data":"2066b48fffbebb577e4352eb48240b9ebfafe1926a62b2692b3f363961ebbf3d"} Oct 05 21:14:53 crc kubenswrapper[4754]: I1005 21:14:53.865333 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6ee3b3b2-07e1-4bab-9372-893eef07f250","Type":"ContainerStarted","Data":"3380e04f976d4766b1f52c3232e01620a9cfc06705ef8c5c6004c0e41c7f6640"} Oct 05 21:14:54 crc kubenswrapper[4754]: I1005 21:14:54.222196 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-4p4pw" Oct 05 21:14:54 crc kubenswrapper[4754]: I1005 21:14:54.297542 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/498207e0-f869-46d9-bdc9-d83f44d02385-scripts\") pod \"498207e0-f869-46d9-bdc9-d83f44d02385\" (UID: \"498207e0-f869-46d9-bdc9-d83f44d02385\") " Oct 05 21:14:54 crc kubenswrapper[4754]: I1005 21:14:54.298050 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/498207e0-f869-46d9-bdc9-d83f44d02385-combined-ca-bundle\") pod \"498207e0-f869-46d9-bdc9-d83f44d02385\" (UID: \"498207e0-f869-46d9-bdc9-d83f44d02385\") " Oct 05 21:14:54 crc kubenswrapper[4754]: I1005 21:14:54.298556 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/498207e0-f869-46d9-bdc9-d83f44d02385-config-data\") pod \"498207e0-f869-46d9-bdc9-d83f44d02385\" (UID: \"498207e0-f869-46d9-bdc9-d83f44d02385\") " Oct 05 21:14:54 crc kubenswrapper[4754]: I1005 21:14:54.298667 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wsjjx\" (UniqueName: \"kubernetes.io/projected/498207e0-f869-46d9-bdc9-d83f44d02385-kube-api-access-wsjjx\") pod \"498207e0-f869-46d9-bdc9-d83f44d02385\" (UID: \"498207e0-f869-46d9-bdc9-d83f44d02385\") " Oct 05 21:14:54 crc kubenswrapper[4754]: I1005 21:14:54.306025 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/498207e0-f869-46d9-bdc9-d83f44d02385-scripts" (OuterVolumeSpecName: "scripts") pod "498207e0-f869-46d9-bdc9-d83f44d02385" (UID: "498207e0-f869-46d9-bdc9-d83f44d02385"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:14:54 crc kubenswrapper[4754]: I1005 21:14:54.328804 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/498207e0-f869-46d9-bdc9-d83f44d02385-kube-api-access-wsjjx" (OuterVolumeSpecName: "kube-api-access-wsjjx") pod "498207e0-f869-46d9-bdc9-d83f44d02385" (UID: "498207e0-f869-46d9-bdc9-d83f44d02385"). InnerVolumeSpecName "kube-api-access-wsjjx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:14:54 crc kubenswrapper[4754]: I1005 21:14:54.363576 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/498207e0-f869-46d9-bdc9-d83f44d02385-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "498207e0-f869-46d9-bdc9-d83f44d02385" (UID: "498207e0-f869-46d9-bdc9-d83f44d02385"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:14:54 crc kubenswrapper[4754]: I1005 21:14:54.403030 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wsjjx\" (UniqueName: \"kubernetes.io/projected/498207e0-f869-46d9-bdc9-d83f44d02385-kube-api-access-wsjjx\") on node \"crc\" DevicePath \"\"" Oct 05 21:14:54 crc kubenswrapper[4754]: I1005 21:14:54.403283 4754 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/498207e0-f869-46d9-bdc9-d83f44d02385-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 21:14:54 crc kubenswrapper[4754]: I1005 21:14:54.403352 4754 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/498207e0-f869-46d9-bdc9-d83f44d02385-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 21:14:54 crc kubenswrapper[4754]: I1005 21:14:54.437137 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/498207e0-f869-46d9-bdc9-d83f44d02385-config-data" (OuterVolumeSpecName: "config-data") pod "498207e0-f869-46d9-bdc9-d83f44d02385" (UID: "498207e0-f869-46d9-bdc9-d83f44d02385"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:14:54 crc kubenswrapper[4754]: I1005 21:14:54.505158 4754 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/498207e0-f869-46d9-bdc9-d83f44d02385-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 21:14:54 crc kubenswrapper[4754]: I1005 21:14:54.875834 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6ee3b3b2-07e1-4bab-9372-893eef07f250","Type":"ContainerStarted","Data":"c1ce65ecb8ae3e72e860298633e5f6d77759ea8ef5efa531ec34ad976df4e63a"} Oct 05 21:14:54 crc kubenswrapper[4754]: I1005 21:14:54.877520 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-4p4pw" event={"ID":"498207e0-f869-46d9-bdc9-d83f44d02385","Type":"ContainerDied","Data":"1a536c303aeabab78e83ed25bc2ffc966a59b83ecea5e576a53b49267d534869"} Oct 05 21:14:54 crc kubenswrapper[4754]: I1005 21:14:54.877656 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1a536c303aeabab78e83ed25bc2ffc966a59b83ecea5e576a53b49267d534869" Oct 05 21:14:54 crc kubenswrapper[4754]: I1005 21:14:54.877572 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-4p4pw" Oct 05 21:14:54 crc kubenswrapper[4754]: I1005 21:14:54.996638 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 05 21:14:54 crc kubenswrapper[4754]: E1005 21:14:54.997584 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="498207e0-f869-46d9-bdc9-d83f44d02385" containerName="nova-cell0-conductor-db-sync" Oct 05 21:14:54 crc kubenswrapper[4754]: I1005 21:14:54.997609 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="498207e0-f869-46d9-bdc9-d83f44d02385" containerName="nova-cell0-conductor-db-sync" Oct 05 21:14:54 crc kubenswrapper[4754]: I1005 21:14:54.997821 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="498207e0-f869-46d9-bdc9-d83f44d02385" containerName="nova-cell0-conductor-db-sync" Oct 05 21:14:54 crc kubenswrapper[4754]: I1005 21:14:54.998466 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 05 21:14:55 crc kubenswrapper[4754]: I1005 21:14:55.000574 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-kg5j2" Oct 05 21:14:55 crc kubenswrapper[4754]: I1005 21:14:55.000572 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 05 21:14:55 crc kubenswrapper[4754]: I1005 21:14:55.094395 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 05 21:14:55 crc kubenswrapper[4754]: I1005 21:14:55.123073 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8e2a011-2922-4e21-8e47-b6fcebbc21e9-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"b8e2a011-2922-4e21-8e47-b6fcebbc21e9\") " pod="openstack/nova-cell0-conductor-0" Oct 05 21:14:55 crc kubenswrapper[4754]: I1005 21:14:55.123269 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b8e2a011-2922-4e21-8e47-b6fcebbc21e9-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"b8e2a011-2922-4e21-8e47-b6fcebbc21e9\") " pod="openstack/nova-cell0-conductor-0" Oct 05 21:14:55 crc kubenswrapper[4754]: I1005 21:14:55.123390 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hdst9\" (UniqueName: \"kubernetes.io/projected/b8e2a011-2922-4e21-8e47-b6fcebbc21e9-kube-api-access-hdst9\") pod \"nova-cell0-conductor-0\" (UID: \"b8e2a011-2922-4e21-8e47-b6fcebbc21e9\") " pod="openstack/nova-cell0-conductor-0" Oct 05 21:14:55 crc kubenswrapper[4754]: I1005 21:14:55.224883 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8e2a011-2922-4e21-8e47-b6fcebbc21e9-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"b8e2a011-2922-4e21-8e47-b6fcebbc21e9\") " pod="openstack/nova-cell0-conductor-0" Oct 05 21:14:55 crc kubenswrapper[4754]: I1005 21:14:55.224976 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b8e2a011-2922-4e21-8e47-b6fcebbc21e9-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"b8e2a011-2922-4e21-8e47-b6fcebbc21e9\") " pod="openstack/nova-cell0-conductor-0" Oct 05 21:14:55 crc kubenswrapper[4754]: I1005 21:14:55.225006 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hdst9\" (UniqueName: \"kubernetes.io/projected/b8e2a011-2922-4e21-8e47-b6fcebbc21e9-kube-api-access-hdst9\") pod \"nova-cell0-conductor-0\" (UID: \"b8e2a011-2922-4e21-8e47-b6fcebbc21e9\") " pod="openstack/nova-cell0-conductor-0" Oct 05 21:14:55 crc kubenswrapper[4754]: I1005 21:14:55.228895 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b8e2a011-2922-4e21-8e47-b6fcebbc21e9-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"b8e2a011-2922-4e21-8e47-b6fcebbc21e9\") " pod="openstack/nova-cell0-conductor-0" Oct 05 21:14:55 crc kubenswrapper[4754]: I1005 21:14:55.230306 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8e2a011-2922-4e21-8e47-b6fcebbc21e9-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"b8e2a011-2922-4e21-8e47-b6fcebbc21e9\") " pod="openstack/nova-cell0-conductor-0" Oct 05 21:14:55 crc kubenswrapper[4754]: I1005 21:14:55.242385 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hdst9\" (UniqueName: \"kubernetes.io/projected/b8e2a011-2922-4e21-8e47-b6fcebbc21e9-kube-api-access-hdst9\") pod \"nova-cell0-conductor-0\" (UID: \"b8e2a011-2922-4e21-8e47-b6fcebbc21e9\") " pod="openstack/nova-cell0-conductor-0" Oct 05 21:14:55 crc kubenswrapper[4754]: I1005 21:14:55.359997 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 05 21:14:55 crc kubenswrapper[4754]: I1005 21:14:55.866426 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 05 21:14:55 crc kubenswrapper[4754]: W1005 21:14:55.907831 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb8e2a011_2922_4e21_8e47_b6fcebbc21e9.slice/crio-3e5b5dbbf98939b3434e64fb7003221881c356b9516f8b4ef5207191a20a7b0f WatchSource:0}: Error finding container 3e5b5dbbf98939b3434e64fb7003221881c356b9516f8b4ef5207191a20a7b0f: Status 404 returned error can't find the container with id 3e5b5dbbf98939b3434e64fb7003221881c356b9516f8b4ef5207191a20a7b0f Oct 05 21:14:55 crc kubenswrapper[4754]: I1005 21:14:55.913624 4754 generic.go:334] "Generic (PLEG): container finished" podID="39d1aac3-14a1-45bf-9a9d-ba4cdb0afcb1" containerID="704413f5b17f48e4e3d7d104afa7e9d7beadfdcc2c562b18589f4f0a201b7281" exitCode=137 Oct 05 21:14:55 crc kubenswrapper[4754]: I1005 21:14:55.913698 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-574454d6cb-vbnk6" event={"ID":"39d1aac3-14a1-45bf-9a9d-ba4cdb0afcb1","Type":"ContainerDied","Data":"704413f5b17f48e4e3d7d104afa7e9d7beadfdcc2c562b18589f4f0a201b7281"} Oct 05 21:14:55 crc kubenswrapper[4754]: I1005 21:14:55.913757 4754 scope.go:117] "RemoveContainer" containerID="7ebc5fb1dedef86d2a80ae2a6e400e3f44aaa47d3b94955c70a007eb48092a6a" Oct 05 21:14:55 crc kubenswrapper[4754]: I1005 21:14:55.919902 4754 generic.go:334] "Generic (PLEG): container finished" podID="27cd6be5-f5eb-4158-bdd1-37f4df3d530d" containerID="8cbf678dfb968e1a5f68bc5b282346cef6ff3c4725e9f27e572e9658e525228e" exitCode=137 Oct 05 21:14:55 crc kubenswrapper[4754]: I1005 21:14:55.919966 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-68c6b67864-9msm5" event={"ID":"27cd6be5-f5eb-4158-bdd1-37f4df3d530d","Type":"ContainerDied","Data":"8cbf678dfb968e1a5f68bc5b282346cef6ff3c4725e9f27e572e9658e525228e"} Oct 05 21:14:55 crc kubenswrapper[4754]: I1005 21:14:55.934085 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6ee3b3b2-07e1-4bab-9372-893eef07f250","Type":"ContainerStarted","Data":"58202d1eec91468517c6840252471678191af733282dd96c7961df59850a5bc7"} Oct 05 21:14:56 crc kubenswrapper[4754]: I1005 21:14:56.156367 4754 scope.go:117] "RemoveContainer" containerID="7d14664c74062c1ff554d9b5e811703f4db1749cb0131849ed4d6056fb07bc02" Oct 05 21:14:56 crc kubenswrapper[4754]: I1005 21:14:56.951939 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-574454d6cb-vbnk6" event={"ID":"39d1aac3-14a1-45bf-9a9d-ba4cdb0afcb1","Type":"ContainerStarted","Data":"94d470c123c39aee1f3662387041048af9ae1b8a7d423ba36132f8bebe037cc7"} Oct 05 21:14:56 crc kubenswrapper[4754]: I1005 21:14:56.953717 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"b8e2a011-2922-4e21-8e47-b6fcebbc21e9","Type":"ContainerStarted","Data":"be4d936aacb51ea38985cd0b9cd9126deffa340cb59ef1caa9026e64c9552e0d"} Oct 05 21:14:56 crc kubenswrapper[4754]: I1005 21:14:56.953761 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"b8e2a011-2922-4e21-8e47-b6fcebbc21e9","Type":"ContainerStarted","Data":"3e5b5dbbf98939b3434e64fb7003221881c356b9516f8b4ef5207191a20a7b0f"} Oct 05 21:14:56 crc kubenswrapper[4754]: I1005 21:14:56.954168 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Oct 05 21:14:56 crc kubenswrapper[4754]: I1005 21:14:56.957109 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-68c6b67864-9msm5" event={"ID":"27cd6be5-f5eb-4158-bdd1-37f4df3d530d","Type":"ContainerStarted","Data":"e8c8cc4de7167b842bac5c879a08a541081eaf3a777d19676eea7ed48c8eae17"} Oct 05 21:14:57 crc kubenswrapper[4754]: I1005 21:14:57.030078 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=3.030059793 podStartE2EDuration="3.030059793s" podCreationTimestamp="2025-10-05 21:14:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 21:14:57.021902434 +0000 UTC m=+1220.926021144" watchObservedRunningTime="2025-10-05 21:14:57.030059793 +0000 UTC m=+1220.934178503" Oct 05 21:14:57 crc kubenswrapper[4754]: I1005 21:14:57.974584 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6ee3b3b2-07e1-4bab-9372-893eef07f250","Type":"ContainerStarted","Data":"9ed026943be03e0a1a0532c59c6308a6e590ffc346f67cf516ee3fa829679dab"} Oct 05 21:14:58 crc kubenswrapper[4754]: I1005 21:14:58.001563 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.543664853 podStartE2EDuration="7.001536826s" podCreationTimestamp="2025-10-05 21:14:51 +0000 UTC" firstStartedPulling="2025-10-05 21:14:52.780823774 +0000 UTC m=+1216.684942484" lastFinishedPulling="2025-10-05 21:14:57.238695747 +0000 UTC m=+1221.142814457" observedRunningTime="2025-10-05 21:14:57.994628681 +0000 UTC m=+1221.898747391" watchObservedRunningTime="2025-10-05 21:14:58.001536826 +0000 UTC m=+1221.905655546" Oct 05 21:14:58 crc kubenswrapper[4754]: I1005 21:14:58.985276 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 05 21:15:00 crc kubenswrapper[4754]: I1005 21:15:00.142951 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29328315-mwzdd"] Oct 05 21:15:00 crc kubenswrapper[4754]: I1005 21:15:00.144565 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29328315-mwzdd" Oct 05 21:15:00 crc kubenswrapper[4754]: I1005 21:15:00.147393 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 05 21:15:00 crc kubenswrapper[4754]: I1005 21:15:00.147476 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 05 21:15:00 crc kubenswrapper[4754]: I1005 21:15:00.204799 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29328315-mwzdd"] Oct 05 21:15:00 crc kubenswrapper[4754]: I1005 21:15:00.284465 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/545fecb5-deec-43b9-9aac-303579f1e9c5-secret-volume\") pod \"collect-profiles-29328315-mwzdd\" (UID: \"545fecb5-deec-43b9-9aac-303579f1e9c5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29328315-mwzdd" Oct 05 21:15:00 crc kubenswrapper[4754]: I1005 21:15:00.284527 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/545fecb5-deec-43b9-9aac-303579f1e9c5-config-volume\") pod \"collect-profiles-29328315-mwzdd\" (UID: \"545fecb5-deec-43b9-9aac-303579f1e9c5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29328315-mwzdd" Oct 05 21:15:00 crc kubenswrapper[4754]: I1005 21:15:00.284564 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7d2hr\" (UniqueName: \"kubernetes.io/projected/545fecb5-deec-43b9-9aac-303579f1e9c5-kube-api-access-7d2hr\") pod \"collect-profiles-29328315-mwzdd\" (UID: \"545fecb5-deec-43b9-9aac-303579f1e9c5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29328315-mwzdd" Oct 05 21:15:00 crc kubenswrapper[4754]: I1005 21:15:00.398375 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/545fecb5-deec-43b9-9aac-303579f1e9c5-secret-volume\") pod \"collect-profiles-29328315-mwzdd\" (UID: \"545fecb5-deec-43b9-9aac-303579f1e9c5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29328315-mwzdd" Oct 05 21:15:00 crc kubenswrapper[4754]: I1005 21:15:00.398928 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/545fecb5-deec-43b9-9aac-303579f1e9c5-config-volume\") pod \"collect-profiles-29328315-mwzdd\" (UID: \"545fecb5-deec-43b9-9aac-303579f1e9c5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29328315-mwzdd" Oct 05 21:15:00 crc kubenswrapper[4754]: I1005 21:15:00.399000 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7d2hr\" (UniqueName: \"kubernetes.io/projected/545fecb5-deec-43b9-9aac-303579f1e9c5-kube-api-access-7d2hr\") pod \"collect-profiles-29328315-mwzdd\" (UID: \"545fecb5-deec-43b9-9aac-303579f1e9c5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29328315-mwzdd" Oct 05 21:15:00 crc kubenswrapper[4754]: I1005 21:15:00.404071 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/545fecb5-deec-43b9-9aac-303579f1e9c5-config-volume\") pod \"collect-profiles-29328315-mwzdd\" (UID: \"545fecb5-deec-43b9-9aac-303579f1e9c5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29328315-mwzdd" Oct 05 21:15:00 crc kubenswrapper[4754]: I1005 21:15:00.425241 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/545fecb5-deec-43b9-9aac-303579f1e9c5-secret-volume\") pod \"collect-profiles-29328315-mwzdd\" (UID: \"545fecb5-deec-43b9-9aac-303579f1e9c5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29328315-mwzdd" Oct 05 21:15:00 crc kubenswrapper[4754]: I1005 21:15:00.432051 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7d2hr\" (UniqueName: \"kubernetes.io/projected/545fecb5-deec-43b9-9aac-303579f1e9c5-kube-api-access-7d2hr\") pod \"collect-profiles-29328315-mwzdd\" (UID: \"545fecb5-deec-43b9-9aac-303579f1e9c5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29328315-mwzdd" Oct 05 21:15:00 crc kubenswrapper[4754]: I1005 21:15:00.474605 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29328315-mwzdd" Oct 05 21:15:00 crc kubenswrapper[4754]: I1005 21:15:00.957145 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29328315-mwzdd"] Oct 05 21:15:01 crc kubenswrapper[4754]: I1005 21:15:01.012670 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29328315-mwzdd" event={"ID":"545fecb5-deec-43b9-9aac-303579f1e9c5","Type":"ContainerStarted","Data":"70f1752a3184cf6577c32606c7f574d807eb34d1a639a9650c025bff9a260c40"} Oct 05 21:15:02 crc kubenswrapper[4754]: I1005 21:15:02.024823 4754 generic.go:334] "Generic (PLEG): container finished" podID="545fecb5-deec-43b9-9aac-303579f1e9c5" containerID="fe2d1561f99f4ac4b0e2d046b6d8439f2b155c3f309f47b33d7b20f2ed80b389" exitCode=0 Oct 05 21:15:02 crc kubenswrapper[4754]: I1005 21:15:02.025084 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29328315-mwzdd" event={"ID":"545fecb5-deec-43b9-9aac-303579f1e9c5","Type":"ContainerDied","Data":"fe2d1561f99f4ac4b0e2d046b6d8439f2b155c3f309f47b33d7b20f2ed80b389"} Oct 05 21:15:03 crc kubenswrapper[4754]: I1005 21:15:03.437419 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29328315-mwzdd" Oct 05 21:15:03 crc kubenswrapper[4754]: I1005 21:15:03.573088 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/545fecb5-deec-43b9-9aac-303579f1e9c5-secret-volume\") pod \"545fecb5-deec-43b9-9aac-303579f1e9c5\" (UID: \"545fecb5-deec-43b9-9aac-303579f1e9c5\") " Oct 05 21:15:03 crc kubenswrapper[4754]: I1005 21:15:03.573399 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/545fecb5-deec-43b9-9aac-303579f1e9c5-config-volume\") pod \"545fecb5-deec-43b9-9aac-303579f1e9c5\" (UID: \"545fecb5-deec-43b9-9aac-303579f1e9c5\") " Oct 05 21:15:03 crc kubenswrapper[4754]: I1005 21:15:03.573449 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7d2hr\" (UniqueName: \"kubernetes.io/projected/545fecb5-deec-43b9-9aac-303579f1e9c5-kube-api-access-7d2hr\") pod \"545fecb5-deec-43b9-9aac-303579f1e9c5\" (UID: \"545fecb5-deec-43b9-9aac-303579f1e9c5\") " Oct 05 21:15:03 crc kubenswrapper[4754]: I1005 21:15:03.574150 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/545fecb5-deec-43b9-9aac-303579f1e9c5-config-volume" (OuterVolumeSpecName: "config-volume") pod "545fecb5-deec-43b9-9aac-303579f1e9c5" (UID: "545fecb5-deec-43b9-9aac-303579f1e9c5"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 21:15:03 crc kubenswrapper[4754]: I1005 21:15:03.583554 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/545fecb5-deec-43b9-9aac-303579f1e9c5-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "545fecb5-deec-43b9-9aac-303579f1e9c5" (UID: "545fecb5-deec-43b9-9aac-303579f1e9c5"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:15:03 crc kubenswrapper[4754]: I1005 21:15:03.585199 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/545fecb5-deec-43b9-9aac-303579f1e9c5-kube-api-access-7d2hr" (OuterVolumeSpecName: "kube-api-access-7d2hr") pod "545fecb5-deec-43b9-9aac-303579f1e9c5" (UID: "545fecb5-deec-43b9-9aac-303579f1e9c5"). InnerVolumeSpecName "kube-api-access-7d2hr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:15:03 crc kubenswrapper[4754]: I1005 21:15:03.675454 4754 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/545fecb5-deec-43b9-9aac-303579f1e9c5-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 05 21:15:03 crc kubenswrapper[4754]: I1005 21:15:03.675515 4754 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/545fecb5-deec-43b9-9aac-303579f1e9c5-config-volume\") on node \"crc\" DevicePath \"\"" Oct 05 21:15:03 crc kubenswrapper[4754]: I1005 21:15:03.675532 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7d2hr\" (UniqueName: \"kubernetes.io/projected/545fecb5-deec-43b9-9aac-303579f1e9c5-kube-api-access-7d2hr\") on node \"crc\" DevicePath \"\"" Oct 05 21:15:04 crc kubenswrapper[4754]: I1005 21:15:04.045301 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29328315-mwzdd" event={"ID":"545fecb5-deec-43b9-9aac-303579f1e9c5","Type":"ContainerDied","Data":"70f1752a3184cf6577c32606c7f574d807eb34d1a639a9650c025bff9a260c40"} Oct 05 21:15:04 crc kubenswrapper[4754]: I1005 21:15:04.045343 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="70f1752a3184cf6577c32606c7f574d807eb34d1a639a9650c025bff9a260c40" Oct 05 21:15:04 crc kubenswrapper[4754]: I1005 21:15:04.045402 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29328315-mwzdd" Oct 05 21:15:05 crc kubenswrapper[4754]: I1005 21:15:05.417603 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Oct 05 21:15:05 crc kubenswrapper[4754]: I1005 21:15:05.603460 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-68c6b67864-9msm5" Oct 05 21:15:05 crc kubenswrapper[4754]: I1005 21:15:05.604764 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-68c6b67864-9msm5" Oct 05 21:15:05 crc kubenswrapper[4754]: I1005 21:15:05.722362 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-574454d6cb-vbnk6" Oct 05 21:15:05 crc kubenswrapper[4754]: I1005 21:15:05.725040 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-574454d6cb-vbnk6" Oct 05 21:15:05 crc kubenswrapper[4754]: I1005 21:15:05.984781 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-fpsrz"] Oct 05 21:15:05 crc kubenswrapper[4754]: E1005 21:15:05.985574 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="545fecb5-deec-43b9-9aac-303579f1e9c5" containerName="collect-profiles" Oct 05 21:15:05 crc kubenswrapper[4754]: I1005 21:15:05.985592 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="545fecb5-deec-43b9-9aac-303579f1e9c5" containerName="collect-profiles" Oct 05 21:15:05 crc kubenswrapper[4754]: I1005 21:15:05.985777 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="545fecb5-deec-43b9-9aac-303579f1e9c5" containerName="collect-profiles" Oct 05 21:15:05 crc kubenswrapper[4754]: I1005 21:15:05.986420 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-fpsrz" Oct 05 21:15:05 crc kubenswrapper[4754]: I1005 21:15:05.996057 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Oct 05 21:15:05 crc kubenswrapper[4754]: I1005 21:15:05.996909 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Oct 05 21:15:06 crc kubenswrapper[4754]: I1005 21:15:06.010417 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-fpsrz"] Oct 05 21:15:06 crc kubenswrapper[4754]: I1005 21:15:06.143627 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d2683cf4-4f3b-4843-b859-75f9c38dfc4d-scripts\") pod \"nova-cell0-cell-mapping-fpsrz\" (UID: \"d2683cf4-4f3b-4843-b859-75f9c38dfc4d\") " pod="openstack/nova-cell0-cell-mapping-fpsrz" Oct 05 21:15:06 crc kubenswrapper[4754]: I1005 21:15:06.143686 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2683cf4-4f3b-4843-b859-75f9c38dfc4d-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-fpsrz\" (UID: \"d2683cf4-4f3b-4843-b859-75f9c38dfc4d\") " pod="openstack/nova-cell0-cell-mapping-fpsrz" Oct 05 21:15:06 crc kubenswrapper[4754]: I1005 21:15:06.143737 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m5qzn\" (UniqueName: \"kubernetes.io/projected/d2683cf4-4f3b-4843-b859-75f9c38dfc4d-kube-api-access-m5qzn\") pod \"nova-cell0-cell-mapping-fpsrz\" (UID: \"d2683cf4-4f3b-4843-b859-75f9c38dfc4d\") " pod="openstack/nova-cell0-cell-mapping-fpsrz" Oct 05 21:15:06 crc kubenswrapper[4754]: I1005 21:15:06.143897 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2683cf4-4f3b-4843-b859-75f9c38dfc4d-config-data\") pod \"nova-cell0-cell-mapping-fpsrz\" (UID: \"d2683cf4-4f3b-4843-b859-75f9c38dfc4d\") " pod="openstack/nova-cell0-cell-mapping-fpsrz" Oct 05 21:15:06 crc kubenswrapper[4754]: I1005 21:15:06.216891 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 05 21:15:06 crc kubenswrapper[4754]: I1005 21:15:06.218579 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 05 21:15:06 crc kubenswrapper[4754]: I1005 21:15:06.242164 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 05 21:15:06 crc kubenswrapper[4754]: I1005 21:15:06.245621 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2683cf4-4f3b-4843-b859-75f9c38dfc4d-config-data\") pod \"nova-cell0-cell-mapping-fpsrz\" (UID: \"d2683cf4-4f3b-4843-b859-75f9c38dfc4d\") " pod="openstack/nova-cell0-cell-mapping-fpsrz" Oct 05 21:15:06 crc kubenswrapper[4754]: I1005 21:15:06.245817 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d2683cf4-4f3b-4843-b859-75f9c38dfc4d-scripts\") pod \"nova-cell0-cell-mapping-fpsrz\" (UID: \"d2683cf4-4f3b-4843-b859-75f9c38dfc4d\") " pod="openstack/nova-cell0-cell-mapping-fpsrz" Oct 05 21:15:06 crc kubenswrapper[4754]: I1005 21:15:06.245888 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2683cf4-4f3b-4843-b859-75f9c38dfc4d-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-fpsrz\" (UID: \"d2683cf4-4f3b-4843-b859-75f9c38dfc4d\") " pod="openstack/nova-cell0-cell-mapping-fpsrz" Oct 05 21:15:06 crc kubenswrapper[4754]: I1005 21:15:06.245979 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m5qzn\" (UniqueName: \"kubernetes.io/projected/d2683cf4-4f3b-4843-b859-75f9c38dfc4d-kube-api-access-m5qzn\") pod \"nova-cell0-cell-mapping-fpsrz\" (UID: \"d2683cf4-4f3b-4843-b859-75f9c38dfc4d\") " pod="openstack/nova-cell0-cell-mapping-fpsrz" Oct 05 21:15:06 crc kubenswrapper[4754]: I1005 21:15:06.246779 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 05 21:15:06 crc kubenswrapper[4754]: I1005 21:15:06.279003 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d2683cf4-4f3b-4843-b859-75f9c38dfc4d-scripts\") pod \"nova-cell0-cell-mapping-fpsrz\" (UID: \"d2683cf4-4f3b-4843-b859-75f9c38dfc4d\") " pod="openstack/nova-cell0-cell-mapping-fpsrz" Oct 05 21:15:06 crc kubenswrapper[4754]: I1005 21:15:06.282083 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2683cf4-4f3b-4843-b859-75f9c38dfc4d-config-data\") pod \"nova-cell0-cell-mapping-fpsrz\" (UID: \"d2683cf4-4f3b-4843-b859-75f9c38dfc4d\") " pod="openstack/nova-cell0-cell-mapping-fpsrz" Oct 05 21:15:06 crc kubenswrapper[4754]: I1005 21:15:06.283309 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2683cf4-4f3b-4843-b859-75f9c38dfc4d-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-fpsrz\" (UID: \"d2683cf4-4f3b-4843-b859-75f9c38dfc4d\") " pod="openstack/nova-cell0-cell-mapping-fpsrz" Oct 05 21:15:06 crc kubenswrapper[4754]: I1005 21:15:06.305139 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m5qzn\" (UniqueName: \"kubernetes.io/projected/d2683cf4-4f3b-4843-b859-75f9c38dfc4d-kube-api-access-m5qzn\") pod \"nova-cell0-cell-mapping-fpsrz\" (UID: \"d2683cf4-4f3b-4843-b859-75f9c38dfc4d\") " pod="openstack/nova-cell0-cell-mapping-fpsrz" Oct 05 21:15:06 crc kubenswrapper[4754]: I1005 21:15:06.317971 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-fpsrz" Oct 05 21:15:06 crc kubenswrapper[4754]: I1005 21:15:06.353142 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74b91ec9-bc00-4917-869f-a84f3cdab6fb-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"74b91ec9-bc00-4917-869f-a84f3cdab6fb\") " pod="openstack/nova-metadata-0" Oct 05 21:15:06 crc kubenswrapper[4754]: I1005 21:15:06.353209 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74b91ec9-bc00-4917-869f-a84f3cdab6fb-config-data\") pod \"nova-metadata-0\" (UID: \"74b91ec9-bc00-4917-869f-a84f3cdab6fb\") " pod="openstack/nova-metadata-0" Oct 05 21:15:06 crc kubenswrapper[4754]: I1005 21:15:06.353261 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vz7vw\" (UniqueName: \"kubernetes.io/projected/74b91ec9-bc00-4917-869f-a84f3cdab6fb-kube-api-access-vz7vw\") pod \"nova-metadata-0\" (UID: \"74b91ec9-bc00-4917-869f-a84f3cdab6fb\") " pod="openstack/nova-metadata-0" Oct 05 21:15:06 crc kubenswrapper[4754]: I1005 21:15:06.353289 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/74b91ec9-bc00-4917-869f-a84f3cdab6fb-logs\") pod \"nova-metadata-0\" (UID: \"74b91ec9-bc00-4917-869f-a84f3cdab6fb\") " pod="openstack/nova-metadata-0" Oct 05 21:15:06 crc kubenswrapper[4754]: I1005 21:15:06.455762 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74b91ec9-bc00-4917-869f-a84f3cdab6fb-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"74b91ec9-bc00-4917-869f-a84f3cdab6fb\") " pod="openstack/nova-metadata-0" Oct 05 21:15:06 crc kubenswrapper[4754]: I1005 21:15:06.456120 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74b91ec9-bc00-4917-869f-a84f3cdab6fb-config-data\") pod \"nova-metadata-0\" (UID: \"74b91ec9-bc00-4917-869f-a84f3cdab6fb\") " pod="openstack/nova-metadata-0" Oct 05 21:15:06 crc kubenswrapper[4754]: I1005 21:15:06.456170 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vz7vw\" (UniqueName: \"kubernetes.io/projected/74b91ec9-bc00-4917-869f-a84f3cdab6fb-kube-api-access-vz7vw\") pod \"nova-metadata-0\" (UID: \"74b91ec9-bc00-4917-869f-a84f3cdab6fb\") " pod="openstack/nova-metadata-0" Oct 05 21:15:06 crc kubenswrapper[4754]: I1005 21:15:06.456200 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/74b91ec9-bc00-4917-869f-a84f3cdab6fb-logs\") pod \"nova-metadata-0\" (UID: \"74b91ec9-bc00-4917-869f-a84f3cdab6fb\") " pod="openstack/nova-metadata-0" Oct 05 21:15:06 crc kubenswrapper[4754]: I1005 21:15:06.456588 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/74b91ec9-bc00-4917-869f-a84f3cdab6fb-logs\") pod \"nova-metadata-0\" (UID: \"74b91ec9-bc00-4917-869f-a84f3cdab6fb\") " pod="openstack/nova-metadata-0" Oct 05 21:15:06 crc kubenswrapper[4754]: I1005 21:15:06.475267 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74b91ec9-bc00-4917-869f-a84f3cdab6fb-config-data\") pod \"nova-metadata-0\" (UID: \"74b91ec9-bc00-4917-869f-a84f3cdab6fb\") " pod="openstack/nova-metadata-0" Oct 05 21:15:06 crc kubenswrapper[4754]: I1005 21:15:06.496295 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74b91ec9-bc00-4917-869f-a84f3cdab6fb-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"74b91ec9-bc00-4917-869f-a84f3cdab6fb\") " pod="openstack/nova-metadata-0" Oct 05 21:15:06 crc kubenswrapper[4754]: I1005 21:15:06.527330 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vz7vw\" (UniqueName: \"kubernetes.io/projected/74b91ec9-bc00-4917-869f-a84f3cdab6fb-kube-api-access-vz7vw\") pod \"nova-metadata-0\" (UID: \"74b91ec9-bc00-4917-869f-a84f3cdab6fb\") " pod="openstack/nova-metadata-0" Oct 05 21:15:06 crc kubenswrapper[4754]: I1005 21:15:06.546362 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 05 21:15:06 crc kubenswrapper[4754]: I1005 21:15:06.566120 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 05 21:15:06 crc kubenswrapper[4754]: I1005 21:15:06.573832 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 05 21:15:06 crc kubenswrapper[4754]: I1005 21:15:06.596234 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Oct 05 21:15:06 crc kubenswrapper[4754]: I1005 21:15:06.644522 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 05 21:15:06 crc kubenswrapper[4754]: I1005 21:15:06.646182 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 05 21:15:06 crc kubenswrapper[4754]: I1005 21:15:06.653158 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 05 21:15:06 crc kubenswrapper[4754]: I1005 21:15:06.661961 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d27z7\" (UniqueName: \"kubernetes.io/projected/746c64fc-7a41-440f-894f-79af6edcc9ee-kube-api-access-d27z7\") pod \"nova-cell1-novncproxy-0\" (UID: \"746c64fc-7a41-440f-894f-79af6edcc9ee\") " pod="openstack/nova-cell1-novncproxy-0" Oct 05 21:15:06 crc kubenswrapper[4754]: I1005 21:15:06.662008 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/746c64fc-7a41-440f-894f-79af6edcc9ee-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"746c64fc-7a41-440f-894f-79af6edcc9ee\") " pod="openstack/nova-cell1-novncproxy-0" Oct 05 21:15:06 crc kubenswrapper[4754]: I1005 21:15:06.662067 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/746c64fc-7a41-440f-894f-79af6edcc9ee-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"746c64fc-7a41-440f-894f-79af6edcc9ee\") " pod="openstack/nova-cell1-novncproxy-0" Oct 05 21:15:06 crc kubenswrapper[4754]: I1005 21:15:06.662217 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 05 21:15:06 crc kubenswrapper[4754]: I1005 21:15:06.694563 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 05 21:15:06 crc kubenswrapper[4754]: I1005 21:15:06.695948 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 05 21:15:06 crc kubenswrapper[4754]: I1005 21:15:06.703929 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 05 21:15:06 crc kubenswrapper[4754]: I1005 21:15:06.720078 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-886x9"] Oct 05 21:15:06 crc kubenswrapper[4754]: I1005 21:15:06.748387 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-845d6d6f59-886x9" Oct 05 21:15:06 crc kubenswrapper[4754]: I1005 21:15:06.755030 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 05 21:15:06 crc kubenswrapper[4754]: I1005 21:15:06.773181 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d27z7\" (UniqueName: \"kubernetes.io/projected/746c64fc-7a41-440f-894f-79af6edcc9ee-kube-api-access-d27z7\") pod \"nova-cell1-novncproxy-0\" (UID: \"746c64fc-7a41-440f-894f-79af6edcc9ee\") " pod="openstack/nova-cell1-novncproxy-0" Oct 05 21:15:06 crc kubenswrapper[4754]: I1005 21:15:06.773225 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/746c64fc-7a41-440f-894f-79af6edcc9ee-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"746c64fc-7a41-440f-894f-79af6edcc9ee\") " pod="openstack/nova-cell1-novncproxy-0" Oct 05 21:15:06 crc kubenswrapper[4754]: I1005 21:15:06.773678 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76f0c9ec-6a29-4250-ae7a-21d4d29d3084-config-data\") pod \"nova-api-0\" (UID: \"76f0c9ec-6a29-4250-ae7a-21d4d29d3084\") " pod="openstack/nova-api-0" Oct 05 21:15:06 crc kubenswrapper[4754]: I1005 21:15:06.773726 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76f0c9ec-6a29-4250-ae7a-21d4d29d3084-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"76f0c9ec-6a29-4250-ae7a-21d4d29d3084\") " pod="openstack/nova-api-0" Oct 05 21:15:06 crc kubenswrapper[4754]: I1005 21:15:06.773797 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/76f0c9ec-6a29-4250-ae7a-21d4d29d3084-logs\") pod \"nova-api-0\" (UID: \"76f0c9ec-6a29-4250-ae7a-21d4d29d3084\") " pod="openstack/nova-api-0" Oct 05 21:15:06 crc kubenswrapper[4754]: I1005 21:15:06.773835 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/746c64fc-7a41-440f-894f-79af6edcc9ee-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"746c64fc-7a41-440f-894f-79af6edcc9ee\") " pod="openstack/nova-cell1-novncproxy-0" Oct 05 21:15:06 crc kubenswrapper[4754]: I1005 21:15:06.773928 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mhfsg\" (UniqueName: \"kubernetes.io/projected/76f0c9ec-6a29-4250-ae7a-21d4d29d3084-kube-api-access-mhfsg\") pod \"nova-api-0\" (UID: \"76f0c9ec-6a29-4250-ae7a-21d4d29d3084\") " pod="openstack/nova-api-0" Oct 05 21:15:06 crc kubenswrapper[4754]: I1005 21:15:06.798249 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/746c64fc-7a41-440f-894f-79af6edcc9ee-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"746c64fc-7a41-440f-894f-79af6edcc9ee\") " pod="openstack/nova-cell1-novncproxy-0" Oct 05 21:15:06 crc kubenswrapper[4754]: I1005 21:15:06.799259 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/746c64fc-7a41-440f-894f-79af6edcc9ee-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"746c64fc-7a41-440f-894f-79af6edcc9ee\") " pod="openstack/nova-cell1-novncproxy-0" Oct 05 21:15:06 crc kubenswrapper[4754]: I1005 21:15:06.850545 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d27z7\" (UniqueName: \"kubernetes.io/projected/746c64fc-7a41-440f-894f-79af6edcc9ee-kube-api-access-d27z7\") pod \"nova-cell1-novncproxy-0\" (UID: \"746c64fc-7a41-440f-894f-79af6edcc9ee\") " pod="openstack/nova-cell1-novncproxy-0" Oct 05 21:15:06 crc kubenswrapper[4754]: I1005 21:15:06.876685 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5prtn\" (UniqueName: \"kubernetes.io/projected/22fd2da3-30b4-4e02-92b7-455f64232af5-kube-api-access-5prtn\") pod \"nova-scheduler-0\" (UID: \"22fd2da3-30b4-4e02-92b7-455f64232af5\") " pod="openstack/nova-scheduler-0" Oct 05 21:15:06 crc kubenswrapper[4754]: I1005 21:15:06.877117 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1bb3793f-57ce-4f10-9452-30ff52baabda-dns-swift-storage-0\") pod \"dnsmasq-dns-845d6d6f59-886x9\" (UID: \"1bb3793f-57ce-4f10-9452-30ff52baabda\") " pod="openstack/dnsmasq-dns-845d6d6f59-886x9" Oct 05 21:15:06 crc kubenswrapper[4754]: I1005 21:15:06.877140 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bb3793f-57ce-4f10-9452-30ff52baabda-config\") pod \"dnsmasq-dns-845d6d6f59-886x9\" (UID: \"1bb3793f-57ce-4f10-9452-30ff52baabda\") " pod="openstack/dnsmasq-dns-845d6d6f59-886x9" Oct 05 21:15:06 crc kubenswrapper[4754]: I1005 21:15:06.877168 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1bb3793f-57ce-4f10-9452-30ff52baabda-dns-svc\") pod \"dnsmasq-dns-845d6d6f59-886x9\" (UID: \"1bb3793f-57ce-4f10-9452-30ff52baabda\") " pod="openstack/dnsmasq-dns-845d6d6f59-886x9" Oct 05 21:15:06 crc kubenswrapper[4754]: I1005 21:15:06.877214 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76f0c9ec-6a29-4250-ae7a-21d4d29d3084-config-data\") pod \"nova-api-0\" (UID: \"76f0c9ec-6a29-4250-ae7a-21d4d29d3084\") " pod="openstack/nova-api-0" Oct 05 21:15:06 crc kubenswrapper[4754]: I1005 21:15:06.877241 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76f0c9ec-6a29-4250-ae7a-21d4d29d3084-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"76f0c9ec-6a29-4250-ae7a-21d4d29d3084\") " pod="openstack/nova-api-0" Oct 05 21:15:06 crc kubenswrapper[4754]: I1005 21:15:06.877259 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1bb3793f-57ce-4f10-9452-30ff52baabda-ovsdbserver-sb\") pod \"dnsmasq-dns-845d6d6f59-886x9\" (UID: \"1bb3793f-57ce-4f10-9452-30ff52baabda\") " pod="openstack/dnsmasq-dns-845d6d6f59-886x9" Oct 05 21:15:06 crc kubenswrapper[4754]: I1005 21:15:06.877283 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1bb3793f-57ce-4f10-9452-30ff52baabda-ovsdbserver-nb\") pod \"dnsmasq-dns-845d6d6f59-886x9\" (UID: \"1bb3793f-57ce-4f10-9452-30ff52baabda\") " pod="openstack/dnsmasq-dns-845d6d6f59-886x9" Oct 05 21:15:06 crc kubenswrapper[4754]: I1005 21:15:06.877314 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-74fxg\" (UniqueName: \"kubernetes.io/projected/1bb3793f-57ce-4f10-9452-30ff52baabda-kube-api-access-74fxg\") pod \"dnsmasq-dns-845d6d6f59-886x9\" (UID: \"1bb3793f-57ce-4f10-9452-30ff52baabda\") " pod="openstack/dnsmasq-dns-845d6d6f59-886x9" Oct 05 21:15:06 crc kubenswrapper[4754]: I1005 21:15:06.877336 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/76f0c9ec-6a29-4250-ae7a-21d4d29d3084-logs\") pod \"nova-api-0\" (UID: \"76f0c9ec-6a29-4250-ae7a-21d4d29d3084\") " pod="openstack/nova-api-0" Oct 05 21:15:06 crc kubenswrapper[4754]: I1005 21:15:06.877379 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mhfsg\" (UniqueName: \"kubernetes.io/projected/76f0c9ec-6a29-4250-ae7a-21d4d29d3084-kube-api-access-mhfsg\") pod \"nova-api-0\" (UID: \"76f0c9ec-6a29-4250-ae7a-21d4d29d3084\") " pod="openstack/nova-api-0" Oct 05 21:15:06 crc kubenswrapper[4754]: I1005 21:15:06.877396 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22fd2da3-30b4-4e02-92b7-455f64232af5-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"22fd2da3-30b4-4e02-92b7-455f64232af5\") " pod="openstack/nova-scheduler-0" Oct 05 21:15:06 crc kubenswrapper[4754]: I1005 21:15:06.877428 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/22fd2da3-30b4-4e02-92b7-455f64232af5-config-data\") pod \"nova-scheduler-0\" (UID: \"22fd2da3-30b4-4e02-92b7-455f64232af5\") " pod="openstack/nova-scheduler-0" Oct 05 21:15:06 crc kubenswrapper[4754]: I1005 21:15:06.879956 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/76f0c9ec-6a29-4250-ae7a-21d4d29d3084-logs\") pod \"nova-api-0\" (UID: \"76f0c9ec-6a29-4250-ae7a-21d4d29d3084\") " pod="openstack/nova-api-0" Oct 05 21:15:06 crc kubenswrapper[4754]: I1005 21:15:06.889258 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 05 21:15:06 crc kubenswrapper[4754]: I1005 21:15:06.890824 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76f0c9ec-6a29-4250-ae7a-21d4d29d3084-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"76f0c9ec-6a29-4250-ae7a-21d4d29d3084\") " pod="openstack/nova-api-0" Oct 05 21:15:06 crc kubenswrapper[4754]: I1005 21:15:06.894413 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76f0c9ec-6a29-4250-ae7a-21d4d29d3084-config-data\") pod \"nova-api-0\" (UID: \"76f0c9ec-6a29-4250-ae7a-21d4d29d3084\") " pod="openstack/nova-api-0" Oct 05 21:15:06 crc kubenswrapper[4754]: I1005 21:15:06.900973 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mhfsg\" (UniqueName: \"kubernetes.io/projected/76f0c9ec-6a29-4250-ae7a-21d4d29d3084-kube-api-access-mhfsg\") pod \"nova-api-0\" (UID: \"76f0c9ec-6a29-4250-ae7a-21d4d29d3084\") " pod="openstack/nova-api-0" Oct 05 21:15:06 crc kubenswrapper[4754]: I1005 21:15:06.910149 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 05 21:15:06 crc kubenswrapper[4754]: I1005 21:15:06.948347 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-886x9"] Oct 05 21:15:07 crc kubenswrapper[4754]: I1005 21:15:07.015970 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5prtn\" (UniqueName: \"kubernetes.io/projected/22fd2da3-30b4-4e02-92b7-455f64232af5-kube-api-access-5prtn\") pod \"nova-scheduler-0\" (UID: \"22fd2da3-30b4-4e02-92b7-455f64232af5\") " pod="openstack/nova-scheduler-0" Oct 05 21:15:07 crc kubenswrapper[4754]: I1005 21:15:07.016082 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1bb3793f-57ce-4f10-9452-30ff52baabda-dns-swift-storage-0\") pod \"dnsmasq-dns-845d6d6f59-886x9\" (UID: \"1bb3793f-57ce-4f10-9452-30ff52baabda\") " pod="openstack/dnsmasq-dns-845d6d6f59-886x9" Oct 05 21:15:07 crc kubenswrapper[4754]: I1005 21:15:07.016111 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bb3793f-57ce-4f10-9452-30ff52baabda-config\") pod \"dnsmasq-dns-845d6d6f59-886x9\" (UID: \"1bb3793f-57ce-4f10-9452-30ff52baabda\") " pod="openstack/dnsmasq-dns-845d6d6f59-886x9" Oct 05 21:15:07 crc kubenswrapper[4754]: I1005 21:15:07.016172 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1bb3793f-57ce-4f10-9452-30ff52baabda-dns-svc\") pod \"dnsmasq-dns-845d6d6f59-886x9\" (UID: \"1bb3793f-57ce-4f10-9452-30ff52baabda\") " pod="openstack/dnsmasq-dns-845d6d6f59-886x9" Oct 05 21:15:07 crc kubenswrapper[4754]: I1005 21:15:07.016244 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1bb3793f-57ce-4f10-9452-30ff52baabda-ovsdbserver-sb\") pod \"dnsmasq-dns-845d6d6f59-886x9\" (UID: \"1bb3793f-57ce-4f10-9452-30ff52baabda\") " pod="openstack/dnsmasq-dns-845d6d6f59-886x9" Oct 05 21:15:07 crc kubenswrapper[4754]: I1005 21:15:07.016276 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1bb3793f-57ce-4f10-9452-30ff52baabda-ovsdbserver-nb\") pod \"dnsmasq-dns-845d6d6f59-886x9\" (UID: \"1bb3793f-57ce-4f10-9452-30ff52baabda\") " pod="openstack/dnsmasq-dns-845d6d6f59-886x9" Oct 05 21:15:07 crc kubenswrapper[4754]: I1005 21:15:07.016310 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-74fxg\" (UniqueName: \"kubernetes.io/projected/1bb3793f-57ce-4f10-9452-30ff52baabda-kube-api-access-74fxg\") pod \"dnsmasq-dns-845d6d6f59-886x9\" (UID: \"1bb3793f-57ce-4f10-9452-30ff52baabda\") " pod="openstack/dnsmasq-dns-845d6d6f59-886x9" Oct 05 21:15:07 crc kubenswrapper[4754]: I1005 21:15:07.016375 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22fd2da3-30b4-4e02-92b7-455f64232af5-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"22fd2da3-30b4-4e02-92b7-455f64232af5\") " pod="openstack/nova-scheduler-0" Oct 05 21:15:07 crc kubenswrapper[4754]: I1005 21:15:07.016408 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/22fd2da3-30b4-4e02-92b7-455f64232af5-config-data\") pod \"nova-scheduler-0\" (UID: \"22fd2da3-30b4-4e02-92b7-455f64232af5\") " pod="openstack/nova-scheduler-0" Oct 05 21:15:07 crc kubenswrapper[4754]: I1005 21:15:07.018266 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1bb3793f-57ce-4f10-9452-30ff52baabda-ovsdbserver-nb\") pod \"dnsmasq-dns-845d6d6f59-886x9\" (UID: \"1bb3793f-57ce-4f10-9452-30ff52baabda\") " pod="openstack/dnsmasq-dns-845d6d6f59-886x9" Oct 05 21:15:07 crc kubenswrapper[4754]: I1005 21:15:07.028295 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1bb3793f-57ce-4f10-9452-30ff52baabda-dns-svc\") pod \"dnsmasq-dns-845d6d6f59-886x9\" (UID: \"1bb3793f-57ce-4f10-9452-30ff52baabda\") " pod="openstack/dnsmasq-dns-845d6d6f59-886x9" Oct 05 21:15:07 crc kubenswrapper[4754]: I1005 21:15:07.028946 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1bb3793f-57ce-4f10-9452-30ff52baabda-ovsdbserver-sb\") pod \"dnsmasq-dns-845d6d6f59-886x9\" (UID: \"1bb3793f-57ce-4f10-9452-30ff52baabda\") " pod="openstack/dnsmasq-dns-845d6d6f59-886x9" Oct 05 21:15:07 crc kubenswrapper[4754]: I1005 21:15:07.035556 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1bb3793f-57ce-4f10-9452-30ff52baabda-dns-swift-storage-0\") pod \"dnsmasq-dns-845d6d6f59-886x9\" (UID: \"1bb3793f-57ce-4f10-9452-30ff52baabda\") " pod="openstack/dnsmasq-dns-845d6d6f59-886x9" Oct 05 21:15:07 crc kubenswrapper[4754]: I1005 21:15:07.036134 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bb3793f-57ce-4f10-9452-30ff52baabda-config\") pod \"dnsmasq-dns-845d6d6f59-886x9\" (UID: \"1bb3793f-57ce-4f10-9452-30ff52baabda\") " pod="openstack/dnsmasq-dns-845d6d6f59-886x9" Oct 05 21:15:07 crc kubenswrapper[4754]: I1005 21:15:07.058004 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22fd2da3-30b4-4e02-92b7-455f64232af5-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"22fd2da3-30b4-4e02-92b7-455f64232af5\") " pod="openstack/nova-scheduler-0" Oct 05 21:15:07 crc kubenswrapper[4754]: I1005 21:15:07.059869 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 05 21:15:07 crc kubenswrapper[4754]: I1005 21:15:07.061174 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-74fxg\" (UniqueName: \"kubernetes.io/projected/1bb3793f-57ce-4f10-9452-30ff52baabda-kube-api-access-74fxg\") pod \"dnsmasq-dns-845d6d6f59-886x9\" (UID: \"1bb3793f-57ce-4f10-9452-30ff52baabda\") " pod="openstack/dnsmasq-dns-845d6d6f59-886x9" Oct 05 21:15:07 crc kubenswrapper[4754]: I1005 21:15:07.082256 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/22fd2da3-30b4-4e02-92b7-455f64232af5-config-data\") pod \"nova-scheduler-0\" (UID: \"22fd2da3-30b4-4e02-92b7-455f64232af5\") " pod="openstack/nova-scheduler-0" Oct 05 21:15:07 crc kubenswrapper[4754]: I1005 21:15:07.088713 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5prtn\" (UniqueName: \"kubernetes.io/projected/22fd2da3-30b4-4e02-92b7-455f64232af5-kube-api-access-5prtn\") pod \"nova-scheduler-0\" (UID: \"22fd2da3-30b4-4e02-92b7-455f64232af5\") " pod="openstack/nova-scheduler-0" Oct 05 21:15:07 crc kubenswrapper[4754]: I1005 21:15:07.101311 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 05 21:15:07 crc kubenswrapper[4754]: I1005 21:15:07.131486 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-845d6d6f59-886x9" Oct 05 21:15:07 crc kubenswrapper[4754]: I1005 21:15:07.341679 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-fpsrz"] Oct 05 21:15:07 crc kubenswrapper[4754]: I1005 21:15:07.525020 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 05 21:15:07 crc kubenswrapper[4754]: I1005 21:15:07.732829 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 05 21:15:07 crc kubenswrapper[4754]: I1005 21:15:07.980559 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 05 21:15:07 crc kubenswrapper[4754]: I1005 21:15:07.990595 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 05 21:15:08 crc kubenswrapper[4754]: I1005 21:15:08.123777 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-khclj"] Oct 05 21:15:08 crc kubenswrapper[4754]: I1005 21:15:08.125537 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-khclj" Oct 05 21:15:08 crc kubenswrapper[4754]: I1005 21:15:08.137759 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 05 21:15:08 crc kubenswrapper[4754]: I1005 21:15:08.137931 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Oct 05 21:15:08 crc kubenswrapper[4754]: I1005 21:15:08.168033 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"22fd2da3-30b4-4e02-92b7-455f64232af5","Type":"ContainerStarted","Data":"813912eba303928a5cde7b67147b1d2baf5eed70a42ceb5dbf71362655d69e78"} Oct 05 21:15:08 crc kubenswrapper[4754]: I1005 21:15:08.173057 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"76f0c9ec-6a29-4250-ae7a-21d4d29d3084","Type":"ContainerStarted","Data":"1ccda2c1f8437a08067bf69caf4ae11e2a4978bcc60b6c4fafb416ab84dce2b3"} Oct 05 21:15:08 crc kubenswrapper[4754]: I1005 21:15:08.179998 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"746c64fc-7a41-440f-894f-79af6edcc9ee","Type":"ContainerStarted","Data":"cf2fa3aa5225a0d8cc439448a9af9c63aa870c32a91d9202f3fe8b79040feb33"} Oct 05 21:15:08 crc kubenswrapper[4754]: I1005 21:15:08.183549 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-khclj"] Oct 05 21:15:08 crc kubenswrapper[4754]: I1005 21:15:08.190573 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-fpsrz" event={"ID":"d2683cf4-4f3b-4843-b859-75f9c38dfc4d","Type":"ContainerStarted","Data":"b93f6db8a9b3fdf25a9dcf220ebb50e302b178a1bef18437f5f0cf17d7d6d3a4"} Oct 05 21:15:08 crc kubenswrapper[4754]: I1005 21:15:08.190617 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-fpsrz" event={"ID":"d2683cf4-4f3b-4843-b859-75f9c38dfc4d","Type":"ContainerStarted","Data":"8b86079a0c90e781400317d391dcfa80ead7b4fa1d45d74dea5268d8f9970790"} Oct 05 21:15:08 crc kubenswrapper[4754]: I1005 21:15:08.198099 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"74b91ec9-bc00-4917-869f-a84f3cdab6fb","Type":"ContainerStarted","Data":"ec2b2b378f2570820f469143f5a5ca8731d69a979c142828bdcf083c7681b858"} Oct 05 21:15:08 crc kubenswrapper[4754]: I1005 21:15:08.204975 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-886x9"] Oct 05 21:15:08 crc kubenswrapper[4754]: I1005 21:15:08.222261 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-fpsrz" podStartSLOduration=3.22223776 podStartE2EDuration="3.22223776s" podCreationTimestamp="2025-10-05 21:15:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 21:15:08.208519242 +0000 UTC m=+1232.112637952" watchObservedRunningTime="2025-10-05 21:15:08.22223776 +0000 UTC m=+1232.126356470" Oct 05 21:15:08 crc kubenswrapper[4754]: I1005 21:15:08.256678 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5bc88d1c-ec79-415b-a441-86b68fdf064f-config-data\") pod \"nova-cell1-conductor-db-sync-khclj\" (UID: \"5bc88d1c-ec79-415b-a441-86b68fdf064f\") " pod="openstack/nova-cell1-conductor-db-sync-khclj" Oct 05 21:15:08 crc kubenswrapper[4754]: I1005 21:15:08.256797 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5bc88d1c-ec79-415b-a441-86b68fdf064f-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-khclj\" (UID: \"5bc88d1c-ec79-415b-a441-86b68fdf064f\") " pod="openstack/nova-cell1-conductor-db-sync-khclj" Oct 05 21:15:08 crc kubenswrapper[4754]: I1005 21:15:08.256868 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5bc88d1c-ec79-415b-a441-86b68fdf064f-scripts\") pod \"nova-cell1-conductor-db-sync-khclj\" (UID: \"5bc88d1c-ec79-415b-a441-86b68fdf064f\") " pod="openstack/nova-cell1-conductor-db-sync-khclj" Oct 05 21:15:08 crc kubenswrapper[4754]: I1005 21:15:08.257006 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dvbmc\" (UniqueName: \"kubernetes.io/projected/5bc88d1c-ec79-415b-a441-86b68fdf064f-kube-api-access-dvbmc\") pod \"nova-cell1-conductor-db-sync-khclj\" (UID: \"5bc88d1c-ec79-415b-a441-86b68fdf064f\") " pod="openstack/nova-cell1-conductor-db-sync-khclj" Oct 05 21:15:08 crc kubenswrapper[4754]: I1005 21:15:08.378366 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5bc88d1c-ec79-415b-a441-86b68fdf064f-scripts\") pod \"nova-cell1-conductor-db-sync-khclj\" (UID: \"5bc88d1c-ec79-415b-a441-86b68fdf064f\") " pod="openstack/nova-cell1-conductor-db-sync-khclj" Oct 05 21:15:08 crc kubenswrapper[4754]: I1005 21:15:08.378761 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dvbmc\" (UniqueName: \"kubernetes.io/projected/5bc88d1c-ec79-415b-a441-86b68fdf064f-kube-api-access-dvbmc\") pod \"nova-cell1-conductor-db-sync-khclj\" (UID: \"5bc88d1c-ec79-415b-a441-86b68fdf064f\") " pod="openstack/nova-cell1-conductor-db-sync-khclj" Oct 05 21:15:08 crc kubenswrapper[4754]: I1005 21:15:08.378836 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5bc88d1c-ec79-415b-a441-86b68fdf064f-config-data\") pod \"nova-cell1-conductor-db-sync-khclj\" (UID: \"5bc88d1c-ec79-415b-a441-86b68fdf064f\") " pod="openstack/nova-cell1-conductor-db-sync-khclj" Oct 05 21:15:08 crc kubenswrapper[4754]: I1005 21:15:08.378881 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5bc88d1c-ec79-415b-a441-86b68fdf064f-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-khclj\" (UID: \"5bc88d1c-ec79-415b-a441-86b68fdf064f\") " pod="openstack/nova-cell1-conductor-db-sync-khclj" Oct 05 21:15:08 crc kubenswrapper[4754]: I1005 21:15:08.393057 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5bc88d1c-ec79-415b-a441-86b68fdf064f-config-data\") pod \"nova-cell1-conductor-db-sync-khclj\" (UID: \"5bc88d1c-ec79-415b-a441-86b68fdf064f\") " pod="openstack/nova-cell1-conductor-db-sync-khclj" Oct 05 21:15:08 crc kubenswrapper[4754]: I1005 21:15:08.394524 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5bc88d1c-ec79-415b-a441-86b68fdf064f-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-khclj\" (UID: \"5bc88d1c-ec79-415b-a441-86b68fdf064f\") " pod="openstack/nova-cell1-conductor-db-sync-khclj" Oct 05 21:15:08 crc kubenswrapper[4754]: I1005 21:15:08.413076 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5bc88d1c-ec79-415b-a441-86b68fdf064f-scripts\") pod \"nova-cell1-conductor-db-sync-khclj\" (UID: \"5bc88d1c-ec79-415b-a441-86b68fdf064f\") " pod="openstack/nova-cell1-conductor-db-sync-khclj" Oct 05 21:15:08 crc kubenswrapper[4754]: I1005 21:15:08.426030 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dvbmc\" (UniqueName: \"kubernetes.io/projected/5bc88d1c-ec79-415b-a441-86b68fdf064f-kube-api-access-dvbmc\") pod \"nova-cell1-conductor-db-sync-khclj\" (UID: \"5bc88d1c-ec79-415b-a441-86b68fdf064f\") " pod="openstack/nova-cell1-conductor-db-sync-khclj" Oct 05 21:15:08 crc kubenswrapper[4754]: I1005 21:15:08.464752 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-khclj" Oct 05 21:15:09 crc kubenswrapper[4754]: I1005 21:15:09.035751 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-khclj"] Oct 05 21:15:09 crc kubenswrapper[4754]: W1005 21:15:09.061403 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5bc88d1c_ec79_415b_a441_86b68fdf064f.slice/crio-600e4af57f542ebc7ca6e35823cbfec29b82349b08843ffd78c22c401608aa1f WatchSource:0}: Error finding container 600e4af57f542ebc7ca6e35823cbfec29b82349b08843ffd78c22c401608aa1f: Status 404 returned error can't find the container with id 600e4af57f542ebc7ca6e35823cbfec29b82349b08843ffd78c22c401608aa1f Oct 05 21:15:09 crc kubenswrapper[4754]: I1005 21:15:09.249592 4754 generic.go:334] "Generic (PLEG): container finished" podID="1bb3793f-57ce-4f10-9452-30ff52baabda" containerID="0b410ed9181022b6ebd1849808b46acb9dd06c414a2cfbeabb1c6c56fabbf760" exitCode=0 Oct 05 21:15:09 crc kubenswrapper[4754]: I1005 21:15:09.249734 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-886x9" event={"ID":"1bb3793f-57ce-4f10-9452-30ff52baabda","Type":"ContainerDied","Data":"0b410ed9181022b6ebd1849808b46acb9dd06c414a2cfbeabb1c6c56fabbf760"} Oct 05 21:15:09 crc kubenswrapper[4754]: I1005 21:15:09.249766 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-886x9" event={"ID":"1bb3793f-57ce-4f10-9452-30ff52baabda","Type":"ContainerStarted","Data":"81a35e8f3d10eab42b989d814dcb355c9d07d4154711205819ed25e2f470bc43"} Oct 05 21:15:09 crc kubenswrapper[4754]: I1005 21:15:09.267205 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-khclj" event={"ID":"5bc88d1c-ec79-415b-a441-86b68fdf064f","Type":"ContainerStarted","Data":"600e4af57f542ebc7ca6e35823cbfec29b82349b08843ffd78c22c401608aa1f"} Oct 05 21:15:10 crc kubenswrapper[4754]: I1005 21:15:10.280909 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-886x9" event={"ID":"1bb3793f-57ce-4f10-9452-30ff52baabda","Type":"ContainerStarted","Data":"b744b5d6b2440cfdb316893cc10ee6f27f818aef3a954222567332e928539f0f"} Oct 05 21:15:10 crc kubenswrapper[4754]: I1005 21:15:10.281562 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-845d6d6f59-886x9" Oct 05 21:15:10 crc kubenswrapper[4754]: I1005 21:15:10.284374 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-khclj" event={"ID":"5bc88d1c-ec79-415b-a441-86b68fdf064f","Type":"ContainerStarted","Data":"9ead6f708e274a7942637dece32316d7cf2de9ff003c6fa19eb76d7bc9ad792b"} Oct 05 21:15:10 crc kubenswrapper[4754]: I1005 21:15:10.314099 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-845d6d6f59-886x9" podStartSLOduration=4.314081584 podStartE2EDuration="4.314081584s" podCreationTimestamp="2025-10-05 21:15:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 21:15:10.312636445 +0000 UTC m=+1234.216755155" watchObservedRunningTime="2025-10-05 21:15:10.314081584 +0000 UTC m=+1234.218200294" Oct 05 21:15:10 crc kubenswrapper[4754]: I1005 21:15:10.337850 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-khclj" podStartSLOduration=2.337831572 podStartE2EDuration="2.337831572s" podCreationTimestamp="2025-10-05 21:15:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 21:15:10.330723291 +0000 UTC m=+1234.234842001" watchObservedRunningTime="2025-10-05 21:15:10.337831572 +0000 UTC m=+1234.241950302" Oct 05 21:15:11 crc kubenswrapper[4754]: I1005 21:15:11.263550 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 05 21:15:11 crc kubenswrapper[4754]: I1005 21:15:11.274330 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 05 21:15:14 crc kubenswrapper[4754]: I1005 21:15:14.354317 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"22fd2da3-30b4-4e02-92b7-455f64232af5","Type":"ContainerStarted","Data":"f47314e363525c96da1da620ef441697711979100de891a7e34eabd4f6742bde"} Oct 05 21:15:14 crc kubenswrapper[4754]: I1005 21:15:14.357086 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"76f0c9ec-6a29-4250-ae7a-21d4d29d3084","Type":"ContainerStarted","Data":"467abfb0ec7ef0fceea5eb09ef62362880eb56a402ab37779344042f10ed0f38"} Oct 05 21:15:14 crc kubenswrapper[4754]: I1005 21:15:14.357133 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"76f0c9ec-6a29-4250-ae7a-21d4d29d3084","Type":"ContainerStarted","Data":"49d5f2a3799144f8697097a4a091fd392234670f1db76e178a69a345c3bca985"} Oct 05 21:15:14 crc kubenswrapper[4754]: I1005 21:15:14.360005 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"746c64fc-7a41-440f-894f-79af6edcc9ee","Type":"ContainerStarted","Data":"cd4f36a301547763e21743d5818ec98066ba379572e58a96b42a2076dabd0c67"} Oct 05 21:15:14 crc kubenswrapper[4754]: I1005 21:15:14.360056 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="746c64fc-7a41-440f-894f-79af6edcc9ee" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://cd4f36a301547763e21743d5818ec98066ba379572e58a96b42a2076dabd0c67" gracePeriod=30 Oct 05 21:15:14 crc kubenswrapper[4754]: I1005 21:15:14.362189 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"74b91ec9-bc00-4917-869f-a84f3cdab6fb","Type":"ContainerStarted","Data":"4caf4202a0b54fbf81141ad3ff9fa01130d8bfd83ee3df50d975152893d9575f"} Oct 05 21:15:14 crc kubenswrapper[4754]: I1005 21:15:14.362218 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"74b91ec9-bc00-4917-869f-a84f3cdab6fb","Type":"ContainerStarted","Data":"28610ec1687d334c0e3c7e02696191cf547a18fe01b026529e4398c23dcef92b"} Oct 05 21:15:14 crc kubenswrapper[4754]: I1005 21:15:14.362340 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="74b91ec9-bc00-4917-869f-a84f3cdab6fb" containerName="nova-metadata-log" containerID="cri-o://28610ec1687d334c0e3c7e02696191cf547a18fe01b026529e4398c23dcef92b" gracePeriod=30 Oct 05 21:15:14 crc kubenswrapper[4754]: I1005 21:15:14.362448 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="74b91ec9-bc00-4917-869f-a84f3cdab6fb" containerName="nova-metadata-metadata" containerID="cri-o://4caf4202a0b54fbf81141ad3ff9fa01130d8bfd83ee3df50d975152893d9575f" gracePeriod=30 Oct 05 21:15:14 crc kubenswrapper[4754]: I1005 21:15:14.379405 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=3.190063415 podStartE2EDuration="8.379376902s" podCreationTimestamp="2025-10-05 21:15:06 +0000 UTC" firstStartedPulling="2025-10-05 21:15:07.991582285 +0000 UTC m=+1231.895700995" lastFinishedPulling="2025-10-05 21:15:13.180895752 +0000 UTC m=+1237.085014482" observedRunningTime="2025-10-05 21:15:14.373913175 +0000 UTC m=+1238.278031885" watchObservedRunningTime="2025-10-05 21:15:14.379376902 +0000 UTC m=+1238.283495612" Oct 05 21:15:14 crc kubenswrapper[4754]: I1005 21:15:14.401915 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.218250292 podStartE2EDuration="8.401894787s" podCreationTimestamp="2025-10-05 21:15:06 +0000 UTC" firstStartedPulling="2025-10-05 21:15:07.998577253 +0000 UTC m=+1231.902695963" lastFinishedPulling="2025-10-05 21:15:13.182221748 +0000 UTC m=+1237.086340458" observedRunningTime="2025-10-05 21:15:14.392230367 +0000 UTC m=+1238.296349157" watchObservedRunningTime="2025-10-05 21:15:14.401894787 +0000 UTC m=+1238.306013497" Oct 05 21:15:14 crc kubenswrapper[4754]: I1005 21:15:14.412367 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.764938676 podStartE2EDuration="8.412337027s" podCreationTimestamp="2025-10-05 21:15:06 +0000 UTC" firstStartedPulling="2025-10-05 21:15:07.548786092 +0000 UTC m=+1231.452904802" lastFinishedPulling="2025-10-05 21:15:13.196184443 +0000 UTC m=+1237.100303153" observedRunningTime="2025-10-05 21:15:14.410719244 +0000 UTC m=+1238.314837954" watchObservedRunningTime="2025-10-05 21:15:14.412337027 +0000 UTC m=+1238.316455737" Oct 05 21:15:14 crc kubenswrapper[4754]: I1005 21:15:14.432357 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=3.014555971 podStartE2EDuration="8.432341865s" podCreationTimestamp="2025-10-05 21:15:06 +0000 UTC" firstStartedPulling="2025-10-05 21:15:07.763668143 +0000 UTC m=+1231.667786853" lastFinishedPulling="2025-10-05 21:15:13.181454027 +0000 UTC m=+1237.085572747" observedRunningTime="2025-10-05 21:15:14.428487441 +0000 UTC m=+1238.332606151" watchObservedRunningTime="2025-10-05 21:15:14.432341865 +0000 UTC m=+1238.336460575" Oct 05 21:15:14 crc kubenswrapper[4754]: I1005 21:15:14.929263 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 05 21:15:15 crc kubenswrapper[4754]: I1005 21:15:15.051535 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74b91ec9-bc00-4917-869f-a84f3cdab6fb-config-data\") pod \"74b91ec9-bc00-4917-869f-a84f3cdab6fb\" (UID: \"74b91ec9-bc00-4917-869f-a84f3cdab6fb\") " Oct 05 21:15:15 crc kubenswrapper[4754]: I1005 21:15:15.051766 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74b91ec9-bc00-4917-869f-a84f3cdab6fb-combined-ca-bundle\") pod \"74b91ec9-bc00-4917-869f-a84f3cdab6fb\" (UID: \"74b91ec9-bc00-4917-869f-a84f3cdab6fb\") " Oct 05 21:15:15 crc kubenswrapper[4754]: I1005 21:15:15.051866 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/74b91ec9-bc00-4917-869f-a84f3cdab6fb-logs\") pod \"74b91ec9-bc00-4917-869f-a84f3cdab6fb\" (UID: \"74b91ec9-bc00-4917-869f-a84f3cdab6fb\") " Oct 05 21:15:15 crc kubenswrapper[4754]: I1005 21:15:15.051891 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vz7vw\" (UniqueName: \"kubernetes.io/projected/74b91ec9-bc00-4917-869f-a84f3cdab6fb-kube-api-access-vz7vw\") pod \"74b91ec9-bc00-4917-869f-a84f3cdab6fb\" (UID: \"74b91ec9-bc00-4917-869f-a84f3cdab6fb\") " Oct 05 21:15:15 crc kubenswrapper[4754]: I1005 21:15:15.052170 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/74b91ec9-bc00-4917-869f-a84f3cdab6fb-logs" (OuterVolumeSpecName: "logs") pod "74b91ec9-bc00-4917-869f-a84f3cdab6fb" (UID: "74b91ec9-bc00-4917-869f-a84f3cdab6fb"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 21:15:15 crc kubenswrapper[4754]: I1005 21:15:15.052638 4754 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/74b91ec9-bc00-4917-869f-a84f3cdab6fb-logs\") on node \"crc\" DevicePath \"\"" Oct 05 21:15:15 crc kubenswrapper[4754]: I1005 21:15:15.058315 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/74b91ec9-bc00-4917-869f-a84f3cdab6fb-kube-api-access-vz7vw" (OuterVolumeSpecName: "kube-api-access-vz7vw") pod "74b91ec9-bc00-4917-869f-a84f3cdab6fb" (UID: "74b91ec9-bc00-4917-869f-a84f3cdab6fb"). InnerVolumeSpecName "kube-api-access-vz7vw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:15:15 crc kubenswrapper[4754]: I1005 21:15:15.083108 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74b91ec9-bc00-4917-869f-a84f3cdab6fb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "74b91ec9-bc00-4917-869f-a84f3cdab6fb" (UID: "74b91ec9-bc00-4917-869f-a84f3cdab6fb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:15:15 crc kubenswrapper[4754]: I1005 21:15:15.094596 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74b91ec9-bc00-4917-869f-a84f3cdab6fb-config-data" (OuterVolumeSpecName: "config-data") pod "74b91ec9-bc00-4917-869f-a84f3cdab6fb" (UID: "74b91ec9-bc00-4917-869f-a84f3cdab6fb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:15:15 crc kubenswrapper[4754]: I1005 21:15:15.155226 4754 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74b91ec9-bc00-4917-869f-a84f3cdab6fb-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 21:15:15 crc kubenswrapper[4754]: I1005 21:15:15.155284 4754 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74b91ec9-bc00-4917-869f-a84f3cdab6fb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 21:15:15 crc kubenswrapper[4754]: I1005 21:15:15.155306 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vz7vw\" (UniqueName: \"kubernetes.io/projected/74b91ec9-bc00-4917-869f-a84f3cdab6fb-kube-api-access-vz7vw\") on node \"crc\" DevicePath \"\"" Oct 05 21:15:15 crc kubenswrapper[4754]: I1005 21:15:15.379865 4754 generic.go:334] "Generic (PLEG): container finished" podID="74b91ec9-bc00-4917-869f-a84f3cdab6fb" containerID="4caf4202a0b54fbf81141ad3ff9fa01130d8bfd83ee3df50d975152893d9575f" exitCode=0 Oct 05 21:15:15 crc kubenswrapper[4754]: I1005 21:15:15.379898 4754 generic.go:334] "Generic (PLEG): container finished" podID="74b91ec9-bc00-4917-869f-a84f3cdab6fb" containerID="28610ec1687d334c0e3c7e02696191cf547a18fe01b026529e4398c23dcef92b" exitCode=143 Oct 05 21:15:15 crc kubenswrapper[4754]: I1005 21:15:15.381304 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 05 21:15:15 crc kubenswrapper[4754]: I1005 21:15:15.381591 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"74b91ec9-bc00-4917-869f-a84f3cdab6fb","Type":"ContainerDied","Data":"4caf4202a0b54fbf81141ad3ff9fa01130d8bfd83ee3df50d975152893d9575f"} Oct 05 21:15:15 crc kubenswrapper[4754]: I1005 21:15:15.382058 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"74b91ec9-bc00-4917-869f-a84f3cdab6fb","Type":"ContainerDied","Data":"28610ec1687d334c0e3c7e02696191cf547a18fe01b026529e4398c23dcef92b"} Oct 05 21:15:15 crc kubenswrapper[4754]: I1005 21:15:15.382211 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"74b91ec9-bc00-4917-869f-a84f3cdab6fb","Type":"ContainerDied","Data":"ec2b2b378f2570820f469143f5a5ca8731d69a979c142828bdcf083c7681b858"} Oct 05 21:15:15 crc kubenswrapper[4754]: I1005 21:15:15.382093 4754 scope.go:117] "RemoveContainer" containerID="4caf4202a0b54fbf81141ad3ff9fa01130d8bfd83ee3df50d975152893d9575f" Oct 05 21:15:15 crc kubenswrapper[4754]: I1005 21:15:15.414984 4754 scope.go:117] "RemoveContainer" containerID="28610ec1687d334c0e3c7e02696191cf547a18fe01b026529e4398c23dcef92b" Oct 05 21:15:15 crc kubenswrapper[4754]: I1005 21:15:15.428699 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 05 21:15:15 crc kubenswrapper[4754]: I1005 21:15:15.438838 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 05 21:15:15 crc kubenswrapper[4754]: I1005 21:15:15.453520 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 05 21:15:15 crc kubenswrapper[4754]: E1005 21:15:15.454196 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74b91ec9-bc00-4917-869f-a84f3cdab6fb" containerName="nova-metadata-log" Oct 05 21:15:15 crc kubenswrapper[4754]: I1005 21:15:15.454216 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="74b91ec9-bc00-4917-869f-a84f3cdab6fb" containerName="nova-metadata-log" Oct 05 21:15:15 crc kubenswrapper[4754]: E1005 21:15:15.454233 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74b91ec9-bc00-4917-869f-a84f3cdab6fb" containerName="nova-metadata-metadata" Oct 05 21:15:15 crc kubenswrapper[4754]: I1005 21:15:15.454243 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="74b91ec9-bc00-4917-869f-a84f3cdab6fb" containerName="nova-metadata-metadata" Oct 05 21:15:15 crc kubenswrapper[4754]: I1005 21:15:15.454415 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="74b91ec9-bc00-4917-869f-a84f3cdab6fb" containerName="nova-metadata-metadata" Oct 05 21:15:15 crc kubenswrapper[4754]: I1005 21:15:15.454432 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="74b91ec9-bc00-4917-869f-a84f3cdab6fb" containerName="nova-metadata-log" Oct 05 21:15:15 crc kubenswrapper[4754]: I1005 21:15:15.455425 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 05 21:15:15 crc kubenswrapper[4754]: I1005 21:15:15.461131 4754 scope.go:117] "RemoveContainer" containerID="4caf4202a0b54fbf81141ad3ff9fa01130d8bfd83ee3df50d975152893d9575f" Oct 05 21:15:15 crc kubenswrapper[4754]: I1005 21:15:15.461883 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 05 21:15:15 crc kubenswrapper[4754]: I1005 21:15:15.462300 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 05 21:15:15 crc kubenswrapper[4754]: E1005 21:15:15.462451 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4caf4202a0b54fbf81141ad3ff9fa01130d8bfd83ee3df50d975152893d9575f\": container with ID starting with 4caf4202a0b54fbf81141ad3ff9fa01130d8bfd83ee3df50d975152893d9575f not found: ID does not exist" containerID="4caf4202a0b54fbf81141ad3ff9fa01130d8bfd83ee3df50d975152893d9575f" Oct 05 21:15:15 crc kubenswrapper[4754]: I1005 21:15:15.462488 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4caf4202a0b54fbf81141ad3ff9fa01130d8bfd83ee3df50d975152893d9575f"} err="failed to get container status \"4caf4202a0b54fbf81141ad3ff9fa01130d8bfd83ee3df50d975152893d9575f\": rpc error: code = NotFound desc = could not find container \"4caf4202a0b54fbf81141ad3ff9fa01130d8bfd83ee3df50d975152893d9575f\": container with ID starting with 4caf4202a0b54fbf81141ad3ff9fa01130d8bfd83ee3df50d975152893d9575f not found: ID does not exist" Oct 05 21:15:15 crc kubenswrapper[4754]: I1005 21:15:15.462529 4754 scope.go:117] "RemoveContainer" containerID="28610ec1687d334c0e3c7e02696191cf547a18fe01b026529e4398c23dcef92b" Oct 05 21:15:15 crc kubenswrapper[4754]: E1005 21:15:15.466030 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"28610ec1687d334c0e3c7e02696191cf547a18fe01b026529e4398c23dcef92b\": container with ID starting with 28610ec1687d334c0e3c7e02696191cf547a18fe01b026529e4398c23dcef92b not found: ID does not exist" containerID="28610ec1687d334c0e3c7e02696191cf547a18fe01b026529e4398c23dcef92b" Oct 05 21:15:15 crc kubenswrapper[4754]: I1005 21:15:15.466177 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"28610ec1687d334c0e3c7e02696191cf547a18fe01b026529e4398c23dcef92b"} err="failed to get container status \"28610ec1687d334c0e3c7e02696191cf547a18fe01b026529e4398c23dcef92b\": rpc error: code = NotFound desc = could not find container \"28610ec1687d334c0e3c7e02696191cf547a18fe01b026529e4398c23dcef92b\": container with ID starting with 28610ec1687d334c0e3c7e02696191cf547a18fe01b026529e4398c23dcef92b not found: ID does not exist" Oct 05 21:15:15 crc kubenswrapper[4754]: I1005 21:15:15.466308 4754 scope.go:117] "RemoveContainer" containerID="4caf4202a0b54fbf81141ad3ff9fa01130d8bfd83ee3df50d975152893d9575f" Oct 05 21:15:15 crc kubenswrapper[4754]: I1005 21:15:15.467828 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4caf4202a0b54fbf81141ad3ff9fa01130d8bfd83ee3df50d975152893d9575f"} err="failed to get container status \"4caf4202a0b54fbf81141ad3ff9fa01130d8bfd83ee3df50d975152893d9575f\": rpc error: code = NotFound desc = could not find container \"4caf4202a0b54fbf81141ad3ff9fa01130d8bfd83ee3df50d975152893d9575f\": container with ID starting with 4caf4202a0b54fbf81141ad3ff9fa01130d8bfd83ee3df50d975152893d9575f not found: ID does not exist" Oct 05 21:15:15 crc kubenswrapper[4754]: I1005 21:15:15.467858 4754 scope.go:117] "RemoveContainer" containerID="28610ec1687d334c0e3c7e02696191cf547a18fe01b026529e4398c23dcef92b" Oct 05 21:15:15 crc kubenswrapper[4754]: I1005 21:15:15.468178 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"28610ec1687d334c0e3c7e02696191cf547a18fe01b026529e4398c23dcef92b"} err="failed to get container status \"28610ec1687d334c0e3c7e02696191cf547a18fe01b026529e4398c23dcef92b\": rpc error: code = NotFound desc = could not find container \"28610ec1687d334c0e3c7e02696191cf547a18fe01b026529e4398c23dcef92b\": container with ID starting with 28610ec1687d334c0e3c7e02696191cf547a18fe01b026529e4398c23dcef92b not found: ID does not exist" Oct 05 21:15:15 crc kubenswrapper[4754]: I1005 21:15:15.475390 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 05 21:15:15 crc kubenswrapper[4754]: I1005 21:15:15.564539 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/195e1064-1e39-4e74-a37b-6b01481cc471-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"195e1064-1e39-4e74-a37b-6b01481cc471\") " pod="openstack/nova-metadata-0" Oct 05 21:15:15 crc kubenswrapper[4754]: I1005 21:15:15.564883 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/195e1064-1e39-4e74-a37b-6b01481cc471-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"195e1064-1e39-4e74-a37b-6b01481cc471\") " pod="openstack/nova-metadata-0" Oct 05 21:15:15 crc kubenswrapper[4754]: I1005 21:15:15.565091 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mdkv6\" (UniqueName: \"kubernetes.io/projected/195e1064-1e39-4e74-a37b-6b01481cc471-kube-api-access-mdkv6\") pod \"nova-metadata-0\" (UID: \"195e1064-1e39-4e74-a37b-6b01481cc471\") " pod="openstack/nova-metadata-0" Oct 05 21:15:15 crc kubenswrapper[4754]: I1005 21:15:15.565200 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/195e1064-1e39-4e74-a37b-6b01481cc471-config-data\") pod \"nova-metadata-0\" (UID: \"195e1064-1e39-4e74-a37b-6b01481cc471\") " pod="openstack/nova-metadata-0" Oct 05 21:15:15 crc kubenswrapper[4754]: I1005 21:15:15.565316 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/195e1064-1e39-4e74-a37b-6b01481cc471-logs\") pod \"nova-metadata-0\" (UID: \"195e1064-1e39-4e74-a37b-6b01481cc471\") " pod="openstack/nova-metadata-0" Oct 05 21:15:15 crc kubenswrapper[4754]: I1005 21:15:15.605628 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-68c6b67864-9msm5" podUID="27cd6be5-f5eb-4158-bdd1-37f4df3d530d" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.148:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.148:8443: connect: connection refused" Oct 05 21:15:15 crc kubenswrapper[4754]: I1005 21:15:15.667614 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mdkv6\" (UniqueName: \"kubernetes.io/projected/195e1064-1e39-4e74-a37b-6b01481cc471-kube-api-access-mdkv6\") pod \"nova-metadata-0\" (UID: \"195e1064-1e39-4e74-a37b-6b01481cc471\") " pod="openstack/nova-metadata-0" Oct 05 21:15:15 crc kubenswrapper[4754]: I1005 21:15:15.667669 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/195e1064-1e39-4e74-a37b-6b01481cc471-config-data\") pod \"nova-metadata-0\" (UID: \"195e1064-1e39-4e74-a37b-6b01481cc471\") " pod="openstack/nova-metadata-0" Oct 05 21:15:15 crc kubenswrapper[4754]: I1005 21:15:15.667699 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/195e1064-1e39-4e74-a37b-6b01481cc471-logs\") pod \"nova-metadata-0\" (UID: \"195e1064-1e39-4e74-a37b-6b01481cc471\") " pod="openstack/nova-metadata-0" Oct 05 21:15:15 crc kubenswrapper[4754]: I1005 21:15:15.667773 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/195e1064-1e39-4e74-a37b-6b01481cc471-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"195e1064-1e39-4e74-a37b-6b01481cc471\") " pod="openstack/nova-metadata-0" Oct 05 21:15:15 crc kubenswrapper[4754]: I1005 21:15:15.667801 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/195e1064-1e39-4e74-a37b-6b01481cc471-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"195e1064-1e39-4e74-a37b-6b01481cc471\") " pod="openstack/nova-metadata-0" Oct 05 21:15:15 crc kubenswrapper[4754]: I1005 21:15:15.668826 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/195e1064-1e39-4e74-a37b-6b01481cc471-logs\") pod \"nova-metadata-0\" (UID: \"195e1064-1e39-4e74-a37b-6b01481cc471\") " pod="openstack/nova-metadata-0" Oct 05 21:15:15 crc kubenswrapper[4754]: I1005 21:15:15.674077 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/195e1064-1e39-4e74-a37b-6b01481cc471-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"195e1064-1e39-4e74-a37b-6b01481cc471\") " pod="openstack/nova-metadata-0" Oct 05 21:15:15 crc kubenswrapper[4754]: I1005 21:15:15.675308 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/195e1064-1e39-4e74-a37b-6b01481cc471-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"195e1064-1e39-4e74-a37b-6b01481cc471\") " pod="openstack/nova-metadata-0" Oct 05 21:15:15 crc kubenswrapper[4754]: I1005 21:15:15.688010 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/195e1064-1e39-4e74-a37b-6b01481cc471-config-data\") pod \"nova-metadata-0\" (UID: \"195e1064-1e39-4e74-a37b-6b01481cc471\") " pod="openstack/nova-metadata-0" Oct 05 21:15:15 crc kubenswrapper[4754]: I1005 21:15:15.692328 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mdkv6\" (UniqueName: \"kubernetes.io/projected/195e1064-1e39-4e74-a37b-6b01481cc471-kube-api-access-mdkv6\") pod \"nova-metadata-0\" (UID: \"195e1064-1e39-4e74-a37b-6b01481cc471\") " pod="openstack/nova-metadata-0" Oct 05 21:15:15 crc kubenswrapper[4754]: I1005 21:15:15.722448 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-574454d6cb-vbnk6" podUID="39d1aac3-14a1-45bf-9a9d-ba4cdb0afcb1" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.149:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.149:8443: connect: connection refused" Oct 05 21:15:15 crc kubenswrapper[4754]: I1005 21:15:15.785284 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 05 21:15:16 crc kubenswrapper[4754]: I1005 21:15:16.422345 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 05 21:15:16 crc kubenswrapper[4754]: I1005 21:15:16.857836 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="74b91ec9-bc00-4917-869f-a84f3cdab6fb" path="/var/lib/kubelet/pods/74b91ec9-bc00-4917-869f-a84f3cdab6fb/volumes" Oct 05 21:15:16 crc kubenswrapper[4754]: I1005 21:15:16.911529 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Oct 05 21:15:17 crc kubenswrapper[4754]: I1005 21:15:17.062120 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 05 21:15:17 crc kubenswrapper[4754]: I1005 21:15:17.062180 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 05 21:15:17 crc kubenswrapper[4754]: I1005 21:15:17.102672 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 05 21:15:17 crc kubenswrapper[4754]: I1005 21:15:17.103098 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 05 21:15:17 crc kubenswrapper[4754]: I1005 21:15:17.135762 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-845d6d6f59-886x9" Oct 05 21:15:17 crc kubenswrapper[4754]: I1005 21:15:17.245608 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-dvhns"] Oct 05 21:15:17 crc kubenswrapper[4754]: I1005 21:15:17.245878 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5784cf869f-dvhns" podUID="8a5ef4bc-52a1-4ab7-bc00-5bae67d18232" containerName="dnsmasq-dns" containerID="cri-o://8232f51f3147300f4ec3dd8d79475ed0e37e904ce8b0d2ba0df89f73c1725c50" gracePeriod=10 Oct 05 21:15:17 crc kubenswrapper[4754]: I1005 21:15:17.249559 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 05 21:15:17 crc kubenswrapper[4754]: I1005 21:15:17.440177 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"195e1064-1e39-4e74-a37b-6b01481cc471","Type":"ContainerStarted","Data":"fefbbb38754912ce2f1e91f14f3f85c67f28136e18281ab60e288d7537414c7e"} Oct 05 21:15:17 crc kubenswrapper[4754]: I1005 21:15:17.440223 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"195e1064-1e39-4e74-a37b-6b01481cc471","Type":"ContainerStarted","Data":"841dbe4c29096a07f00b3aeeeb458d2b7acbe8b90244a952145375b05c81140b"} Oct 05 21:15:17 crc kubenswrapper[4754]: I1005 21:15:17.440235 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"195e1064-1e39-4e74-a37b-6b01481cc471","Type":"ContainerStarted","Data":"cac0f841f2bbd979c25ec997f46ab1b40a91d2363024539c9d4b4104901ed1ac"} Oct 05 21:15:17 crc kubenswrapper[4754]: I1005 21:15:17.452733 4754 generic.go:334] "Generic (PLEG): container finished" podID="8a5ef4bc-52a1-4ab7-bc00-5bae67d18232" containerID="8232f51f3147300f4ec3dd8d79475ed0e37e904ce8b0d2ba0df89f73c1725c50" exitCode=0 Oct 05 21:15:17 crc kubenswrapper[4754]: I1005 21:15:17.452823 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-dvhns" event={"ID":"8a5ef4bc-52a1-4ab7-bc00-5bae67d18232","Type":"ContainerDied","Data":"8232f51f3147300f4ec3dd8d79475ed0e37e904ce8b0d2ba0df89f73c1725c50"} Oct 05 21:15:17 crc kubenswrapper[4754]: I1005 21:15:17.492607 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 05 21:15:17 crc kubenswrapper[4754]: I1005 21:15:17.511026 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.511003193 podStartE2EDuration="2.511003193s" podCreationTimestamp="2025-10-05 21:15:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 21:15:17.473860855 +0000 UTC m=+1241.377979565" watchObservedRunningTime="2025-10-05 21:15:17.511003193 +0000 UTC m=+1241.415121903" Oct 05 21:15:17 crc kubenswrapper[4754]: I1005 21:15:17.877754 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5784cf869f-dvhns" Oct 05 21:15:17 crc kubenswrapper[4754]: I1005 21:15:17.948386 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8a5ef4bc-52a1-4ab7-bc00-5bae67d18232-ovsdbserver-nb\") pod \"8a5ef4bc-52a1-4ab7-bc00-5bae67d18232\" (UID: \"8a5ef4bc-52a1-4ab7-bc00-5bae67d18232\") " Oct 05 21:15:17 crc kubenswrapper[4754]: I1005 21:15:17.948505 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g84v8\" (UniqueName: \"kubernetes.io/projected/8a5ef4bc-52a1-4ab7-bc00-5bae67d18232-kube-api-access-g84v8\") pod \"8a5ef4bc-52a1-4ab7-bc00-5bae67d18232\" (UID: \"8a5ef4bc-52a1-4ab7-bc00-5bae67d18232\") " Oct 05 21:15:17 crc kubenswrapper[4754]: I1005 21:15:17.948540 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8a5ef4bc-52a1-4ab7-bc00-5bae67d18232-config\") pod \"8a5ef4bc-52a1-4ab7-bc00-5bae67d18232\" (UID: \"8a5ef4bc-52a1-4ab7-bc00-5bae67d18232\") " Oct 05 21:15:17 crc kubenswrapper[4754]: I1005 21:15:17.948565 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8a5ef4bc-52a1-4ab7-bc00-5bae67d18232-dns-swift-storage-0\") pod \"8a5ef4bc-52a1-4ab7-bc00-5bae67d18232\" (UID: \"8a5ef4bc-52a1-4ab7-bc00-5bae67d18232\") " Oct 05 21:15:17 crc kubenswrapper[4754]: I1005 21:15:17.948747 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8a5ef4bc-52a1-4ab7-bc00-5bae67d18232-ovsdbserver-sb\") pod \"8a5ef4bc-52a1-4ab7-bc00-5bae67d18232\" (UID: \"8a5ef4bc-52a1-4ab7-bc00-5bae67d18232\") " Oct 05 21:15:17 crc kubenswrapper[4754]: I1005 21:15:17.948786 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8a5ef4bc-52a1-4ab7-bc00-5bae67d18232-dns-svc\") pod \"8a5ef4bc-52a1-4ab7-bc00-5bae67d18232\" (UID: \"8a5ef4bc-52a1-4ab7-bc00-5bae67d18232\") " Oct 05 21:15:17 crc kubenswrapper[4754]: I1005 21:15:17.965005 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a5ef4bc-52a1-4ab7-bc00-5bae67d18232-kube-api-access-g84v8" (OuterVolumeSpecName: "kube-api-access-g84v8") pod "8a5ef4bc-52a1-4ab7-bc00-5bae67d18232" (UID: "8a5ef4bc-52a1-4ab7-bc00-5bae67d18232"). InnerVolumeSpecName "kube-api-access-g84v8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:15:18 crc kubenswrapper[4754]: I1005 21:15:18.033000 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8a5ef4bc-52a1-4ab7-bc00-5bae67d18232-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "8a5ef4bc-52a1-4ab7-bc00-5bae67d18232" (UID: "8a5ef4bc-52a1-4ab7-bc00-5bae67d18232"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 21:15:18 crc kubenswrapper[4754]: I1005 21:15:18.051032 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g84v8\" (UniqueName: \"kubernetes.io/projected/8a5ef4bc-52a1-4ab7-bc00-5bae67d18232-kube-api-access-g84v8\") on node \"crc\" DevicePath \"\"" Oct 05 21:15:18 crc kubenswrapper[4754]: I1005 21:15:18.051058 4754 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8a5ef4bc-52a1-4ab7-bc00-5bae67d18232-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 05 21:15:18 crc kubenswrapper[4754]: I1005 21:15:18.078716 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8a5ef4bc-52a1-4ab7-bc00-5bae67d18232-config" (OuterVolumeSpecName: "config") pod "8a5ef4bc-52a1-4ab7-bc00-5bae67d18232" (UID: "8a5ef4bc-52a1-4ab7-bc00-5bae67d18232"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 21:15:18 crc kubenswrapper[4754]: I1005 21:15:18.113275 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8a5ef4bc-52a1-4ab7-bc00-5bae67d18232-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "8a5ef4bc-52a1-4ab7-bc00-5bae67d18232" (UID: "8a5ef4bc-52a1-4ab7-bc00-5bae67d18232"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 21:15:18 crc kubenswrapper[4754]: I1005 21:15:18.113456 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8a5ef4bc-52a1-4ab7-bc00-5bae67d18232-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "8a5ef4bc-52a1-4ab7-bc00-5bae67d18232" (UID: "8a5ef4bc-52a1-4ab7-bc00-5bae67d18232"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 21:15:18 crc kubenswrapper[4754]: I1005 21:15:18.147154 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8a5ef4bc-52a1-4ab7-bc00-5bae67d18232-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "8a5ef4bc-52a1-4ab7-bc00-5bae67d18232" (UID: "8a5ef4bc-52a1-4ab7-bc00-5bae67d18232"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 21:15:18 crc kubenswrapper[4754]: I1005 21:15:18.152855 4754 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8a5ef4bc-52a1-4ab7-bc00-5bae67d18232-config\") on node \"crc\" DevicePath \"\"" Oct 05 21:15:18 crc kubenswrapper[4754]: I1005 21:15:18.152893 4754 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8a5ef4bc-52a1-4ab7-bc00-5bae67d18232-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 05 21:15:18 crc kubenswrapper[4754]: I1005 21:15:18.152907 4754 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8a5ef4bc-52a1-4ab7-bc00-5bae67d18232-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 05 21:15:18 crc kubenswrapper[4754]: I1005 21:15:18.152919 4754 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8a5ef4bc-52a1-4ab7-bc00-5bae67d18232-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 05 21:15:18 crc kubenswrapper[4754]: I1005 21:15:18.164755 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="76f0c9ec-6a29-4250-ae7a-21d4d29d3084" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.190:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 05 21:15:18 crc kubenswrapper[4754]: I1005 21:15:18.165214 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="76f0c9ec-6a29-4250-ae7a-21d4d29d3084" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.190:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 05 21:15:18 crc kubenswrapper[4754]: I1005 21:15:18.480112 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-dvhns" event={"ID":"8a5ef4bc-52a1-4ab7-bc00-5bae67d18232","Type":"ContainerDied","Data":"40ca378de3b0f6e31c8a7ea645e6d5239dd825b948855c118d4eb90cb8278546"} Oct 05 21:15:18 crc kubenswrapper[4754]: I1005 21:15:18.480210 4754 scope.go:117] "RemoveContainer" containerID="8232f51f3147300f4ec3dd8d79475ed0e37e904ce8b0d2ba0df89f73c1725c50" Oct 05 21:15:18 crc kubenswrapper[4754]: I1005 21:15:18.480774 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5784cf869f-dvhns" Oct 05 21:15:18 crc kubenswrapper[4754]: I1005 21:15:18.553715 4754 scope.go:117] "RemoveContainer" containerID="d5a18ce4db7c6a5664879e2181b9630036565632163d929a740c53352dc23536" Oct 05 21:15:18 crc kubenswrapper[4754]: I1005 21:15:18.558055 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-dvhns"] Oct 05 21:15:18 crc kubenswrapper[4754]: I1005 21:15:18.595074 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-dvhns"] Oct 05 21:15:18 crc kubenswrapper[4754]: I1005 21:15:18.850644 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8a5ef4bc-52a1-4ab7-bc00-5bae67d18232" path="/var/lib/kubelet/pods/8a5ef4bc-52a1-4ab7-bc00-5bae67d18232/volumes" Oct 05 21:15:20 crc kubenswrapper[4754]: I1005 21:15:20.501556 4754 generic.go:334] "Generic (PLEG): container finished" podID="d2683cf4-4f3b-4843-b859-75f9c38dfc4d" containerID="b93f6db8a9b3fdf25a9dcf220ebb50e302b178a1bef18437f5f0cf17d7d6d3a4" exitCode=0 Oct 05 21:15:20 crc kubenswrapper[4754]: I1005 21:15:20.501942 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-fpsrz" event={"ID":"d2683cf4-4f3b-4843-b859-75f9c38dfc4d","Type":"ContainerDied","Data":"b93f6db8a9b3fdf25a9dcf220ebb50e302b178a1bef18437f5f0cf17d7d6d3a4"} Oct 05 21:15:20 crc kubenswrapper[4754]: I1005 21:15:20.504766 4754 generic.go:334] "Generic (PLEG): container finished" podID="5bc88d1c-ec79-415b-a441-86b68fdf064f" containerID="9ead6f708e274a7942637dece32316d7cf2de9ff003c6fa19eb76d7bc9ad792b" exitCode=0 Oct 05 21:15:20 crc kubenswrapper[4754]: I1005 21:15:20.504809 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-khclj" event={"ID":"5bc88d1c-ec79-415b-a441-86b68fdf064f","Type":"ContainerDied","Data":"9ead6f708e274a7942637dece32316d7cf2de9ff003c6fa19eb76d7bc9ad792b"} Oct 05 21:15:20 crc kubenswrapper[4754]: I1005 21:15:20.786253 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 05 21:15:20 crc kubenswrapper[4754]: I1005 21:15:20.786320 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 05 21:15:22 crc kubenswrapper[4754]: I1005 21:15:22.004864 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-khclj" Oct 05 21:15:22 crc kubenswrapper[4754]: I1005 21:15:22.008483 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-fpsrz" Oct 05 21:15:22 crc kubenswrapper[4754]: I1005 21:15:22.045304 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5bc88d1c-ec79-415b-a441-86b68fdf064f-config-data\") pod \"5bc88d1c-ec79-415b-a441-86b68fdf064f\" (UID: \"5bc88d1c-ec79-415b-a441-86b68fdf064f\") " Oct 05 21:15:22 crc kubenswrapper[4754]: I1005 21:15:22.045362 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d2683cf4-4f3b-4843-b859-75f9c38dfc4d-scripts\") pod \"d2683cf4-4f3b-4843-b859-75f9c38dfc4d\" (UID: \"d2683cf4-4f3b-4843-b859-75f9c38dfc4d\") " Oct 05 21:15:22 crc kubenswrapper[4754]: I1005 21:15:22.045486 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5bc88d1c-ec79-415b-a441-86b68fdf064f-scripts\") pod \"5bc88d1c-ec79-415b-a441-86b68fdf064f\" (UID: \"5bc88d1c-ec79-415b-a441-86b68fdf064f\") " Oct 05 21:15:22 crc kubenswrapper[4754]: I1005 21:15:22.045540 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5bc88d1c-ec79-415b-a441-86b68fdf064f-combined-ca-bundle\") pod \"5bc88d1c-ec79-415b-a441-86b68fdf064f\" (UID: \"5bc88d1c-ec79-415b-a441-86b68fdf064f\") " Oct 05 21:15:22 crc kubenswrapper[4754]: I1005 21:15:22.045709 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dvbmc\" (UniqueName: \"kubernetes.io/projected/5bc88d1c-ec79-415b-a441-86b68fdf064f-kube-api-access-dvbmc\") pod \"5bc88d1c-ec79-415b-a441-86b68fdf064f\" (UID: \"5bc88d1c-ec79-415b-a441-86b68fdf064f\") " Oct 05 21:15:22 crc kubenswrapper[4754]: I1005 21:15:22.045728 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2683cf4-4f3b-4843-b859-75f9c38dfc4d-combined-ca-bundle\") pod \"d2683cf4-4f3b-4843-b859-75f9c38dfc4d\" (UID: \"d2683cf4-4f3b-4843-b859-75f9c38dfc4d\") " Oct 05 21:15:22 crc kubenswrapper[4754]: I1005 21:15:22.045783 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m5qzn\" (UniqueName: \"kubernetes.io/projected/d2683cf4-4f3b-4843-b859-75f9c38dfc4d-kube-api-access-m5qzn\") pod \"d2683cf4-4f3b-4843-b859-75f9c38dfc4d\" (UID: \"d2683cf4-4f3b-4843-b859-75f9c38dfc4d\") " Oct 05 21:15:22 crc kubenswrapper[4754]: I1005 21:15:22.045822 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2683cf4-4f3b-4843-b859-75f9c38dfc4d-config-data\") pod \"d2683cf4-4f3b-4843-b859-75f9c38dfc4d\" (UID: \"d2683cf4-4f3b-4843-b859-75f9c38dfc4d\") " Oct 05 21:15:22 crc kubenswrapper[4754]: I1005 21:15:22.103721 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2683cf4-4f3b-4843-b859-75f9c38dfc4d-scripts" (OuterVolumeSpecName: "scripts") pod "d2683cf4-4f3b-4843-b859-75f9c38dfc4d" (UID: "d2683cf4-4f3b-4843-b859-75f9c38dfc4d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:15:22 crc kubenswrapper[4754]: I1005 21:15:22.113674 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5bc88d1c-ec79-415b-a441-86b68fdf064f-scripts" (OuterVolumeSpecName: "scripts") pod "5bc88d1c-ec79-415b-a441-86b68fdf064f" (UID: "5bc88d1c-ec79-415b-a441-86b68fdf064f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:15:22 crc kubenswrapper[4754]: I1005 21:15:22.113864 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d2683cf4-4f3b-4843-b859-75f9c38dfc4d-kube-api-access-m5qzn" (OuterVolumeSpecName: "kube-api-access-m5qzn") pod "d2683cf4-4f3b-4843-b859-75f9c38dfc4d" (UID: "d2683cf4-4f3b-4843-b859-75f9c38dfc4d"). InnerVolumeSpecName "kube-api-access-m5qzn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:15:22 crc kubenswrapper[4754]: I1005 21:15:22.120967 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5bc88d1c-ec79-415b-a441-86b68fdf064f-kube-api-access-dvbmc" (OuterVolumeSpecName: "kube-api-access-dvbmc") pod "5bc88d1c-ec79-415b-a441-86b68fdf064f" (UID: "5bc88d1c-ec79-415b-a441-86b68fdf064f"). InnerVolumeSpecName "kube-api-access-dvbmc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:15:22 crc kubenswrapper[4754]: I1005 21:15:22.132568 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2683cf4-4f3b-4843-b859-75f9c38dfc4d-config-data" (OuterVolumeSpecName: "config-data") pod "d2683cf4-4f3b-4843-b859-75f9c38dfc4d" (UID: "d2683cf4-4f3b-4843-b859-75f9c38dfc4d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:15:22 crc kubenswrapper[4754]: I1005 21:15:22.150277 4754 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5bc88d1c-ec79-415b-a441-86b68fdf064f-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 21:15:22 crc kubenswrapper[4754]: I1005 21:15:22.150799 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dvbmc\" (UniqueName: \"kubernetes.io/projected/5bc88d1c-ec79-415b-a441-86b68fdf064f-kube-api-access-dvbmc\") on node \"crc\" DevicePath \"\"" Oct 05 21:15:22 crc kubenswrapper[4754]: I1005 21:15:22.150823 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m5qzn\" (UniqueName: \"kubernetes.io/projected/d2683cf4-4f3b-4843-b859-75f9c38dfc4d-kube-api-access-m5qzn\") on node \"crc\" DevicePath \"\"" Oct 05 21:15:22 crc kubenswrapper[4754]: I1005 21:15:22.150834 4754 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2683cf4-4f3b-4843-b859-75f9c38dfc4d-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 21:15:22 crc kubenswrapper[4754]: I1005 21:15:22.150845 4754 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d2683cf4-4f3b-4843-b859-75f9c38dfc4d-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 21:15:22 crc kubenswrapper[4754]: I1005 21:15:22.153935 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5bc88d1c-ec79-415b-a441-86b68fdf064f-config-data" (OuterVolumeSpecName: "config-data") pod "5bc88d1c-ec79-415b-a441-86b68fdf064f" (UID: "5bc88d1c-ec79-415b-a441-86b68fdf064f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:15:22 crc kubenswrapper[4754]: I1005 21:15:22.169403 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5bc88d1c-ec79-415b-a441-86b68fdf064f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5bc88d1c-ec79-415b-a441-86b68fdf064f" (UID: "5bc88d1c-ec79-415b-a441-86b68fdf064f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:15:22 crc kubenswrapper[4754]: I1005 21:15:22.172802 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2683cf4-4f3b-4843-b859-75f9c38dfc4d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d2683cf4-4f3b-4843-b859-75f9c38dfc4d" (UID: "d2683cf4-4f3b-4843-b859-75f9c38dfc4d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:15:22 crc kubenswrapper[4754]: I1005 21:15:22.252564 4754 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2683cf4-4f3b-4843-b859-75f9c38dfc4d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 21:15:22 crc kubenswrapper[4754]: I1005 21:15:22.252623 4754 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5bc88d1c-ec79-415b-a441-86b68fdf064f-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 21:15:22 crc kubenswrapper[4754]: I1005 21:15:22.252637 4754 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5bc88d1c-ec79-415b-a441-86b68fdf064f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 21:15:22 crc kubenswrapper[4754]: I1005 21:15:22.264893 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 05 21:15:22 crc kubenswrapper[4754]: I1005 21:15:22.531991 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-fpsrz" event={"ID":"d2683cf4-4f3b-4843-b859-75f9c38dfc4d","Type":"ContainerDied","Data":"8b86079a0c90e781400317d391dcfa80ead7b4fa1d45d74dea5268d8f9970790"} Oct 05 21:15:22 crc kubenswrapper[4754]: I1005 21:15:22.532329 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8b86079a0c90e781400317d391dcfa80ead7b4fa1d45d74dea5268d8f9970790" Oct 05 21:15:22 crc kubenswrapper[4754]: I1005 21:15:22.532305 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-fpsrz" Oct 05 21:15:22 crc kubenswrapper[4754]: I1005 21:15:22.560097 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-khclj" event={"ID":"5bc88d1c-ec79-415b-a441-86b68fdf064f","Type":"ContainerDied","Data":"600e4af57f542ebc7ca6e35823cbfec29b82349b08843ffd78c22c401608aa1f"} Oct 05 21:15:22 crc kubenswrapper[4754]: I1005 21:15:22.560307 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="600e4af57f542ebc7ca6e35823cbfec29b82349b08843ffd78c22c401608aa1f" Oct 05 21:15:22 crc kubenswrapper[4754]: I1005 21:15:22.560451 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-khclj" Oct 05 21:15:22 crc kubenswrapper[4754]: I1005 21:15:22.633305 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 05 21:15:22 crc kubenswrapper[4754]: E1005 21:15:22.633874 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a5ef4bc-52a1-4ab7-bc00-5bae67d18232" containerName="dnsmasq-dns" Oct 05 21:15:22 crc kubenswrapper[4754]: I1005 21:15:22.633896 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a5ef4bc-52a1-4ab7-bc00-5bae67d18232" containerName="dnsmasq-dns" Oct 05 21:15:22 crc kubenswrapper[4754]: E1005 21:15:22.633917 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2683cf4-4f3b-4843-b859-75f9c38dfc4d" containerName="nova-manage" Oct 05 21:15:22 crc kubenswrapper[4754]: I1005 21:15:22.633923 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2683cf4-4f3b-4843-b859-75f9c38dfc4d" containerName="nova-manage" Oct 05 21:15:22 crc kubenswrapper[4754]: E1005 21:15:22.633936 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5bc88d1c-ec79-415b-a441-86b68fdf064f" containerName="nova-cell1-conductor-db-sync" Oct 05 21:15:22 crc kubenswrapper[4754]: I1005 21:15:22.633943 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="5bc88d1c-ec79-415b-a441-86b68fdf064f" containerName="nova-cell1-conductor-db-sync" Oct 05 21:15:22 crc kubenswrapper[4754]: E1005 21:15:22.633972 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a5ef4bc-52a1-4ab7-bc00-5bae67d18232" containerName="init" Oct 05 21:15:22 crc kubenswrapper[4754]: I1005 21:15:22.633980 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a5ef4bc-52a1-4ab7-bc00-5bae67d18232" containerName="init" Oct 05 21:15:22 crc kubenswrapper[4754]: I1005 21:15:22.634384 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="d2683cf4-4f3b-4843-b859-75f9c38dfc4d" containerName="nova-manage" Oct 05 21:15:22 crc kubenswrapper[4754]: I1005 21:15:22.634409 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a5ef4bc-52a1-4ab7-bc00-5bae67d18232" containerName="dnsmasq-dns" Oct 05 21:15:22 crc kubenswrapper[4754]: I1005 21:15:22.634423 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="5bc88d1c-ec79-415b-a441-86b68fdf064f" containerName="nova-cell1-conductor-db-sync" Oct 05 21:15:22 crc kubenswrapper[4754]: I1005 21:15:22.635261 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 05 21:15:22 crc kubenswrapper[4754]: I1005 21:15:22.638716 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 05 21:15:22 crc kubenswrapper[4754]: I1005 21:15:22.683675 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8pmzl\" (UniqueName: \"kubernetes.io/projected/9cecb9f9-02d8-42fc-b9ff-b38d3a5aea1b-kube-api-access-8pmzl\") pod \"nova-cell1-conductor-0\" (UID: \"9cecb9f9-02d8-42fc-b9ff-b38d3a5aea1b\") " pod="openstack/nova-cell1-conductor-0" Oct 05 21:15:22 crc kubenswrapper[4754]: I1005 21:15:22.683735 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9cecb9f9-02d8-42fc-b9ff-b38d3a5aea1b-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"9cecb9f9-02d8-42fc-b9ff-b38d3a5aea1b\") " pod="openstack/nova-cell1-conductor-0" Oct 05 21:15:22 crc kubenswrapper[4754]: I1005 21:15:22.683760 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9cecb9f9-02d8-42fc-b9ff-b38d3a5aea1b-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"9cecb9f9-02d8-42fc-b9ff-b38d3a5aea1b\") " pod="openstack/nova-cell1-conductor-0" Oct 05 21:15:22 crc kubenswrapper[4754]: I1005 21:15:22.697921 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 05 21:15:22 crc kubenswrapper[4754]: I1005 21:15:22.755718 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 05 21:15:22 crc kubenswrapper[4754]: I1005 21:15:22.756148 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="76f0c9ec-6a29-4250-ae7a-21d4d29d3084" containerName="nova-api-log" containerID="cri-o://49d5f2a3799144f8697097a4a091fd392234670f1db76e178a69a345c3bca985" gracePeriod=30 Oct 05 21:15:22 crc kubenswrapper[4754]: I1005 21:15:22.756232 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="76f0c9ec-6a29-4250-ae7a-21d4d29d3084" containerName="nova-api-api" containerID="cri-o://467abfb0ec7ef0fceea5eb09ef62362880eb56a402ab37779344042f10ed0f38" gracePeriod=30 Oct 05 21:15:22 crc kubenswrapper[4754]: I1005 21:15:22.771780 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 05 21:15:22 crc kubenswrapper[4754]: I1005 21:15:22.772081 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="22fd2da3-30b4-4e02-92b7-455f64232af5" containerName="nova-scheduler-scheduler" containerID="cri-o://f47314e363525c96da1da620ef441697711979100de891a7e34eabd4f6742bde" gracePeriod=30 Oct 05 21:15:22 crc kubenswrapper[4754]: I1005 21:15:22.787209 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9cecb9f9-02d8-42fc-b9ff-b38d3a5aea1b-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"9cecb9f9-02d8-42fc-b9ff-b38d3a5aea1b\") " pod="openstack/nova-cell1-conductor-0" Oct 05 21:15:22 crc kubenswrapper[4754]: I1005 21:15:22.787273 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9cecb9f9-02d8-42fc-b9ff-b38d3a5aea1b-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"9cecb9f9-02d8-42fc-b9ff-b38d3a5aea1b\") " pod="openstack/nova-cell1-conductor-0" Oct 05 21:15:22 crc kubenswrapper[4754]: I1005 21:15:22.787449 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8pmzl\" (UniqueName: \"kubernetes.io/projected/9cecb9f9-02d8-42fc-b9ff-b38d3a5aea1b-kube-api-access-8pmzl\") pod \"nova-cell1-conductor-0\" (UID: \"9cecb9f9-02d8-42fc-b9ff-b38d3a5aea1b\") " pod="openstack/nova-cell1-conductor-0" Oct 05 21:15:22 crc kubenswrapper[4754]: I1005 21:15:22.795807 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9cecb9f9-02d8-42fc-b9ff-b38d3a5aea1b-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"9cecb9f9-02d8-42fc-b9ff-b38d3a5aea1b\") " pod="openstack/nova-cell1-conductor-0" Oct 05 21:15:22 crc kubenswrapper[4754]: I1005 21:15:22.796276 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9cecb9f9-02d8-42fc-b9ff-b38d3a5aea1b-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"9cecb9f9-02d8-42fc-b9ff-b38d3a5aea1b\") " pod="openstack/nova-cell1-conductor-0" Oct 05 21:15:22 crc kubenswrapper[4754]: I1005 21:15:22.810034 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8pmzl\" (UniqueName: \"kubernetes.io/projected/9cecb9f9-02d8-42fc-b9ff-b38d3a5aea1b-kube-api-access-8pmzl\") pod \"nova-cell1-conductor-0\" (UID: \"9cecb9f9-02d8-42fc-b9ff-b38d3a5aea1b\") " pod="openstack/nova-cell1-conductor-0" Oct 05 21:15:22 crc kubenswrapper[4754]: I1005 21:15:22.869576 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 05 21:15:22 crc kubenswrapper[4754]: I1005 21:15:22.869900 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="195e1064-1e39-4e74-a37b-6b01481cc471" containerName="nova-metadata-log" containerID="cri-o://841dbe4c29096a07f00b3aeeeb458d2b7acbe8b90244a952145375b05c81140b" gracePeriod=30 Oct 05 21:15:22 crc kubenswrapper[4754]: I1005 21:15:22.870521 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="195e1064-1e39-4e74-a37b-6b01481cc471" containerName="nova-metadata-metadata" containerID="cri-o://fefbbb38754912ce2f1e91f14f3f85c67f28136e18281ab60e288d7537414c7e" gracePeriod=30 Oct 05 21:15:23 crc kubenswrapper[4754]: I1005 21:15:23.014071 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 05 21:15:23 crc kubenswrapper[4754]: I1005 21:15:23.529320 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 05 21:15:23 crc kubenswrapper[4754]: W1005 21:15:23.537651 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9cecb9f9_02d8_42fc_b9ff_b38d3a5aea1b.slice/crio-a2065dc2f3b0084693141a6f6c5efae6ffd7df2e76a0e4a6011614ff15190452 WatchSource:0}: Error finding container a2065dc2f3b0084693141a6f6c5efae6ffd7df2e76a0e4a6011614ff15190452: Status 404 returned error can't find the container with id a2065dc2f3b0084693141a6f6c5efae6ffd7df2e76a0e4a6011614ff15190452 Oct 05 21:15:23 crc kubenswrapper[4754]: I1005 21:15:23.579252 4754 generic.go:334] "Generic (PLEG): container finished" podID="195e1064-1e39-4e74-a37b-6b01481cc471" containerID="fefbbb38754912ce2f1e91f14f3f85c67f28136e18281ab60e288d7537414c7e" exitCode=0 Oct 05 21:15:23 crc kubenswrapper[4754]: I1005 21:15:23.579274 4754 generic.go:334] "Generic (PLEG): container finished" podID="195e1064-1e39-4e74-a37b-6b01481cc471" containerID="841dbe4c29096a07f00b3aeeeb458d2b7acbe8b90244a952145375b05c81140b" exitCode=143 Oct 05 21:15:23 crc kubenswrapper[4754]: I1005 21:15:23.579325 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"195e1064-1e39-4e74-a37b-6b01481cc471","Type":"ContainerDied","Data":"fefbbb38754912ce2f1e91f14f3f85c67f28136e18281ab60e288d7537414c7e"} Oct 05 21:15:23 crc kubenswrapper[4754]: I1005 21:15:23.579353 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"195e1064-1e39-4e74-a37b-6b01481cc471","Type":"ContainerDied","Data":"841dbe4c29096a07f00b3aeeeb458d2b7acbe8b90244a952145375b05c81140b"} Oct 05 21:15:23 crc kubenswrapper[4754]: I1005 21:15:23.608732 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"9cecb9f9-02d8-42fc-b9ff-b38d3a5aea1b","Type":"ContainerStarted","Data":"a2065dc2f3b0084693141a6f6c5efae6ffd7df2e76a0e4a6011614ff15190452"} Oct 05 21:15:23 crc kubenswrapper[4754]: I1005 21:15:23.641202 4754 generic.go:334] "Generic (PLEG): container finished" podID="76f0c9ec-6a29-4250-ae7a-21d4d29d3084" containerID="49d5f2a3799144f8697097a4a091fd392234670f1db76e178a69a345c3bca985" exitCode=143 Oct 05 21:15:23 crc kubenswrapper[4754]: I1005 21:15:23.641249 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"76f0c9ec-6a29-4250-ae7a-21d4d29d3084","Type":"ContainerDied","Data":"49d5f2a3799144f8697097a4a091fd392234670f1db76e178a69a345c3bca985"} Oct 05 21:15:23 crc kubenswrapper[4754]: I1005 21:15:23.761847 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 05 21:15:23 crc kubenswrapper[4754]: I1005 21:15:23.831052 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/195e1064-1e39-4e74-a37b-6b01481cc471-nova-metadata-tls-certs\") pod \"195e1064-1e39-4e74-a37b-6b01481cc471\" (UID: \"195e1064-1e39-4e74-a37b-6b01481cc471\") " Oct 05 21:15:23 crc kubenswrapper[4754]: I1005 21:15:23.831149 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/195e1064-1e39-4e74-a37b-6b01481cc471-combined-ca-bundle\") pod \"195e1064-1e39-4e74-a37b-6b01481cc471\" (UID: \"195e1064-1e39-4e74-a37b-6b01481cc471\") " Oct 05 21:15:23 crc kubenswrapper[4754]: I1005 21:15:23.831174 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/195e1064-1e39-4e74-a37b-6b01481cc471-config-data\") pod \"195e1064-1e39-4e74-a37b-6b01481cc471\" (UID: \"195e1064-1e39-4e74-a37b-6b01481cc471\") " Oct 05 21:15:23 crc kubenswrapper[4754]: I1005 21:15:23.831364 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mdkv6\" (UniqueName: \"kubernetes.io/projected/195e1064-1e39-4e74-a37b-6b01481cc471-kube-api-access-mdkv6\") pod \"195e1064-1e39-4e74-a37b-6b01481cc471\" (UID: \"195e1064-1e39-4e74-a37b-6b01481cc471\") " Oct 05 21:15:23 crc kubenswrapper[4754]: I1005 21:15:23.831474 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/195e1064-1e39-4e74-a37b-6b01481cc471-logs\") pod \"195e1064-1e39-4e74-a37b-6b01481cc471\" (UID: \"195e1064-1e39-4e74-a37b-6b01481cc471\") " Oct 05 21:15:23 crc kubenswrapper[4754]: I1005 21:15:23.832513 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/195e1064-1e39-4e74-a37b-6b01481cc471-logs" (OuterVolumeSpecName: "logs") pod "195e1064-1e39-4e74-a37b-6b01481cc471" (UID: "195e1064-1e39-4e74-a37b-6b01481cc471"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 21:15:23 crc kubenswrapper[4754]: I1005 21:15:23.842678 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/195e1064-1e39-4e74-a37b-6b01481cc471-kube-api-access-mdkv6" (OuterVolumeSpecName: "kube-api-access-mdkv6") pod "195e1064-1e39-4e74-a37b-6b01481cc471" (UID: "195e1064-1e39-4e74-a37b-6b01481cc471"). InnerVolumeSpecName "kube-api-access-mdkv6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:15:23 crc kubenswrapper[4754]: I1005 21:15:23.891777 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/195e1064-1e39-4e74-a37b-6b01481cc471-config-data" (OuterVolumeSpecName: "config-data") pod "195e1064-1e39-4e74-a37b-6b01481cc471" (UID: "195e1064-1e39-4e74-a37b-6b01481cc471"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:15:23 crc kubenswrapper[4754]: I1005 21:15:23.925073 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/195e1064-1e39-4e74-a37b-6b01481cc471-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "195e1064-1e39-4e74-a37b-6b01481cc471" (UID: "195e1064-1e39-4e74-a37b-6b01481cc471"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:15:23 crc kubenswrapper[4754]: I1005 21:15:23.934187 4754 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/195e1064-1e39-4e74-a37b-6b01481cc471-logs\") on node \"crc\" DevicePath \"\"" Oct 05 21:15:23 crc kubenswrapper[4754]: I1005 21:15:23.934245 4754 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/195e1064-1e39-4e74-a37b-6b01481cc471-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 21:15:23 crc kubenswrapper[4754]: I1005 21:15:23.934259 4754 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/195e1064-1e39-4e74-a37b-6b01481cc471-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 21:15:23 crc kubenswrapper[4754]: I1005 21:15:23.934270 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mdkv6\" (UniqueName: \"kubernetes.io/projected/195e1064-1e39-4e74-a37b-6b01481cc471-kube-api-access-mdkv6\") on node \"crc\" DevicePath \"\"" Oct 05 21:15:23 crc kubenswrapper[4754]: I1005 21:15:23.941658 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/195e1064-1e39-4e74-a37b-6b01481cc471-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "195e1064-1e39-4e74-a37b-6b01481cc471" (UID: "195e1064-1e39-4e74-a37b-6b01481cc471"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:15:24 crc kubenswrapper[4754]: I1005 21:15:24.036537 4754 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/195e1064-1e39-4e74-a37b-6b01481cc471-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 05 21:15:24 crc kubenswrapper[4754]: I1005 21:15:24.654322 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"195e1064-1e39-4e74-a37b-6b01481cc471","Type":"ContainerDied","Data":"cac0f841f2bbd979c25ec997f46ab1b40a91d2363024539c9d4b4104901ed1ac"} Oct 05 21:15:24 crc kubenswrapper[4754]: I1005 21:15:24.654378 4754 scope.go:117] "RemoveContainer" containerID="fefbbb38754912ce2f1e91f14f3f85c67f28136e18281ab60e288d7537414c7e" Oct 05 21:15:24 crc kubenswrapper[4754]: I1005 21:15:24.654535 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 05 21:15:24 crc kubenswrapper[4754]: I1005 21:15:24.659865 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"9cecb9f9-02d8-42fc-b9ff-b38d3a5aea1b","Type":"ContainerStarted","Data":"b691fbcc37d09d65599c465737ccfc51a8133db87915baf71fc105629cd44833"} Oct 05 21:15:24 crc kubenswrapper[4754]: I1005 21:15:24.660076 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Oct 05 21:15:24 crc kubenswrapper[4754]: I1005 21:15:24.680128 4754 scope.go:117] "RemoveContainer" containerID="841dbe4c29096a07f00b3aeeeb458d2b7acbe8b90244a952145375b05c81140b" Oct 05 21:15:24 crc kubenswrapper[4754]: I1005 21:15:24.696874 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.696851335 podStartE2EDuration="2.696851335s" podCreationTimestamp="2025-10-05 21:15:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 21:15:24.679658853 +0000 UTC m=+1248.583777563" watchObservedRunningTime="2025-10-05 21:15:24.696851335 +0000 UTC m=+1248.600970055" Oct 05 21:15:24 crc kubenswrapper[4754]: I1005 21:15:24.707398 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 05 21:15:24 crc kubenswrapper[4754]: I1005 21:15:24.718678 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 05 21:15:24 crc kubenswrapper[4754]: I1005 21:15:24.727393 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 05 21:15:24 crc kubenswrapper[4754]: E1005 21:15:24.727788 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="195e1064-1e39-4e74-a37b-6b01481cc471" containerName="nova-metadata-log" Oct 05 21:15:24 crc kubenswrapper[4754]: I1005 21:15:24.727805 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="195e1064-1e39-4e74-a37b-6b01481cc471" containerName="nova-metadata-log" Oct 05 21:15:24 crc kubenswrapper[4754]: E1005 21:15:24.727816 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="195e1064-1e39-4e74-a37b-6b01481cc471" containerName="nova-metadata-metadata" Oct 05 21:15:24 crc kubenswrapper[4754]: I1005 21:15:24.727822 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="195e1064-1e39-4e74-a37b-6b01481cc471" containerName="nova-metadata-metadata" Oct 05 21:15:24 crc kubenswrapper[4754]: I1005 21:15:24.728015 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="195e1064-1e39-4e74-a37b-6b01481cc471" containerName="nova-metadata-log" Oct 05 21:15:24 crc kubenswrapper[4754]: I1005 21:15:24.728039 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="195e1064-1e39-4e74-a37b-6b01481cc471" containerName="nova-metadata-metadata" Oct 05 21:15:24 crc kubenswrapper[4754]: I1005 21:15:24.728973 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 05 21:15:24 crc kubenswrapper[4754]: I1005 21:15:24.733258 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 05 21:15:24 crc kubenswrapper[4754]: I1005 21:15:24.744766 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 05 21:15:24 crc kubenswrapper[4754]: I1005 21:15:24.761784 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c45e75d9-997b-4b2b-a923-8ee9739fcbbc-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"c45e75d9-997b-4b2b-a923-8ee9739fcbbc\") " pod="openstack/nova-metadata-0" Oct 05 21:15:24 crc kubenswrapper[4754]: I1005 21:15:24.762253 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c45e75d9-997b-4b2b-a923-8ee9739fcbbc-config-data\") pod \"nova-metadata-0\" (UID: \"c45e75d9-997b-4b2b-a923-8ee9739fcbbc\") " pod="openstack/nova-metadata-0" Oct 05 21:15:24 crc kubenswrapper[4754]: I1005 21:15:24.762462 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s45ll\" (UniqueName: \"kubernetes.io/projected/c45e75d9-997b-4b2b-a923-8ee9739fcbbc-kube-api-access-s45ll\") pod \"nova-metadata-0\" (UID: \"c45e75d9-997b-4b2b-a923-8ee9739fcbbc\") " pod="openstack/nova-metadata-0" Oct 05 21:15:24 crc kubenswrapper[4754]: I1005 21:15:24.762759 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c45e75d9-997b-4b2b-a923-8ee9739fcbbc-logs\") pod \"nova-metadata-0\" (UID: \"c45e75d9-997b-4b2b-a923-8ee9739fcbbc\") " pod="openstack/nova-metadata-0" Oct 05 21:15:24 crc kubenswrapper[4754]: I1005 21:15:24.762976 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/c45e75d9-997b-4b2b-a923-8ee9739fcbbc-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"c45e75d9-997b-4b2b-a923-8ee9739fcbbc\") " pod="openstack/nova-metadata-0" Oct 05 21:15:24 crc kubenswrapper[4754]: I1005 21:15:24.775961 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 05 21:15:24 crc kubenswrapper[4754]: I1005 21:15:24.852224 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="195e1064-1e39-4e74-a37b-6b01481cc471" path="/var/lib/kubelet/pods/195e1064-1e39-4e74-a37b-6b01481cc471/volumes" Oct 05 21:15:24 crc kubenswrapper[4754]: I1005 21:15:24.865700 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c45e75d9-997b-4b2b-a923-8ee9739fcbbc-logs\") pod \"nova-metadata-0\" (UID: \"c45e75d9-997b-4b2b-a923-8ee9739fcbbc\") " pod="openstack/nova-metadata-0" Oct 05 21:15:24 crc kubenswrapper[4754]: I1005 21:15:24.865805 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/c45e75d9-997b-4b2b-a923-8ee9739fcbbc-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"c45e75d9-997b-4b2b-a923-8ee9739fcbbc\") " pod="openstack/nova-metadata-0" Oct 05 21:15:24 crc kubenswrapper[4754]: I1005 21:15:24.865849 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c45e75d9-997b-4b2b-a923-8ee9739fcbbc-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"c45e75d9-997b-4b2b-a923-8ee9739fcbbc\") " pod="openstack/nova-metadata-0" Oct 05 21:15:24 crc kubenswrapper[4754]: I1005 21:15:24.865881 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c45e75d9-997b-4b2b-a923-8ee9739fcbbc-config-data\") pod \"nova-metadata-0\" (UID: \"c45e75d9-997b-4b2b-a923-8ee9739fcbbc\") " pod="openstack/nova-metadata-0" Oct 05 21:15:24 crc kubenswrapper[4754]: I1005 21:15:24.865927 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s45ll\" (UniqueName: \"kubernetes.io/projected/c45e75d9-997b-4b2b-a923-8ee9739fcbbc-kube-api-access-s45ll\") pod \"nova-metadata-0\" (UID: \"c45e75d9-997b-4b2b-a923-8ee9739fcbbc\") " pod="openstack/nova-metadata-0" Oct 05 21:15:24 crc kubenswrapper[4754]: I1005 21:15:24.866669 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c45e75d9-997b-4b2b-a923-8ee9739fcbbc-logs\") pod \"nova-metadata-0\" (UID: \"c45e75d9-997b-4b2b-a923-8ee9739fcbbc\") " pod="openstack/nova-metadata-0" Oct 05 21:15:24 crc kubenswrapper[4754]: I1005 21:15:24.874078 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c45e75d9-997b-4b2b-a923-8ee9739fcbbc-config-data\") pod \"nova-metadata-0\" (UID: \"c45e75d9-997b-4b2b-a923-8ee9739fcbbc\") " pod="openstack/nova-metadata-0" Oct 05 21:15:24 crc kubenswrapper[4754]: I1005 21:15:24.875605 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/c45e75d9-997b-4b2b-a923-8ee9739fcbbc-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"c45e75d9-997b-4b2b-a923-8ee9739fcbbc\") " pod="openstack/nova-metadata-0" Oct 05 21:15:24 crc kubenswrapper[4754]: I1005 21:15:24.884522 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c45e75d9-997b-4b2b-a923-8ee9739fcbbc-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"c45e75d9-997b-4b2b-a923-8ee9739fcbbc\") " pod="openstack/nova-metadata-0" Oct 05 21:15:24 crc kubenswrapper[4754]: I1005 21:15:24.890111 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s45ll\" (UniqueName: \"kubernetes.io/projected/c45e75d9-997b-4b2b-a923-8ee9739fcbbc-kube-api-access-s45ll\") pod \"nova-metadata-0\" (UID: \"c45e75d9-997b-4b2b-a923-8ee9739fcbbc\") " pod="openstack/nova-metadata-0" Oct 05 21:15:25 crc kubenswrapper[4754]: I1005 21:15:25.044553 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 05 21:15:25 crc kubenswrapper[4754]: I1005 21:15:25.561919 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 05 21:15:25 crc kubenswrapper[4754]: I1005 21:15:25.604220 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-68c6b67864-9msm5" podUID="27cd6be5-f5eb-4158-bdd1-37f4df3d530d" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.148:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.148:8443: connect: connection refused" Oct 05 21:15:25 crc kubenswrapper[4754]: I1005 21:15:25.669794 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c45e75d9-997b-4b2b-a923-8ee9739fcbbc","Type":"ContainerStarted","Data":"54e5975bfd60a82b4331cc54db1b409d2c2bab2841856c7a404ebbc218129086"} Oct 05 21:15:25 crc kubenswrapper[4754]: I1005 21:15:25.721051 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-574454d6cb-vbnk6" podUID="39d1aac3-14a1-45bf-9a9d-ba4cdb0afcb1" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.149:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.149:8443: connect: connection refused" Oct 05 21:15:26 crc kubenswrapper[4754]: I1005 21:15:26.683347 4754 generic.go:334] "Generic (PLEG): container finished" podID="76f0c9ec-6a29-4250-ae7a-21d4d29d3084" containerID="467abfb0ec7ef0fceea5eb09ef62362880eb56a402ab37779344042f10ed0f38" exitCode=0 Oct 05 21:15:26 crc kubenswrapper[4754]: I1005 21:15:26.683434 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"76f0c9ec-6a29-4250-ae7a-21d4d29d3084","Type":"ContainerDied","Data":"467abfb0ec7ef0fceea5eb09ef62362880eb56a402ab37779344042f10ed0f38"} Oct 05 21:15:26 crc kubenswrapper[4754]: I1005 21:15:26.686663 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c45e75d9-997b-4b2b-a923-8ee9739fcbbc","Type":"ContainerStarted","Data":"40f7e9db9c4cfc1a4743fb8a28bcd85bf3946e7cfabbb175345526f616f3cd1d"} Oct 05 21:15:26 crc kubenswrapper[4754]: I1005 21:15:26.686704 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c45e75d9-997b-4b2b-a923-8ee9739fcbbc","Type":"ContainerStarted","Data":"c6039287394c53d920bccc08c1a5fc75ad1e47d196696fb83bcc68a3e88c13b5"} Oct 05 21:15:26 crc kubenswrapper[4754]: I1005 21:15:26.695168 4754 generic.go:334] "Generic (PLEG): container finished" podID="22fd2da3-30b4-4e02-92b7-455f64232af5" containerID="f47314e363525c96da1da620ef441697711979100de891a7e34eabd4f6742bde" exitCode=0 Oct 05 21:15:26 crc kubenswrapper[4754]: I1005 21:15:26.695212 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"22fd2da3-30b4-4e02-92b7-455f64232af5","Type":"ContainerDied","Data":"f47314e363525c96da1da620ef441697711979100de891a7e34eabd4f6742bde"} Oct 05 21:15:26 crc kubenswrapper[4754]: I1005 21:15:26.731944 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.731905574 podStartE2EDuration="2.731905574s" podCreationTimestamp="2025-10-05 21:15:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 21:15:26.718745301 +0000 UTC m=+1250.622864011" watchObservedRunningTime="2025-10-05 21:15:26.731905574 +0000 UTC m=+1250.636024284" Oct 05 21:15:26 crc kubenswrapper[4754]: I1005 21:15:26.815191 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 05 21:15:26 crc kubenswrapper[4754]: I1005 21:15:26.871725 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 05 21:15:26 crc kubenswrapper[4754]: I1005 21:15:26.928703 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/76f0c9ec-6a29-4250-ae7a-21d4d29d3084-logs\") pod \"76f0c9ec-6a29-4250-ae7a-21d4d29d3084\" (UID: \"76f0c9ec-6a29-4250-ae7a-21d4d29d3084\") " Oct 05 21:15:26 crc kubenswrapper[4754]: I1005 21:15:26.928757 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76f0c9ec-6a29-4250-ae7a-21d4d29d3084-combined-ca-bundle\") pod \"76f0c9ec-6a29-4250-ae7a-21d4d29d3084\" (UID: \"76f0c9ec-6a29-4250-ae7a-21d4d29d3084\") " Oct 05 21:15:26 crc kubenswrapper[4754]: I1005 21:15:26.928797 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mhfsg\" (UniqueName: \"kubernetes.io/projected/76f0c9ec-6a29-4250-ae7a-21d4d29d3084-kube-api-access-mhfsg\") pod \"76f0c9ec-6a29-4250-ae7a-21d4d29d3084\" (UID: \"76f0c9ec-6a29-4250-ae7a-21d4d29d3084\") " Oct 05 21:15:26 crc kubenswrapper[4754]: I1005 21:15:26.928930 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76f0c9ec-6a29-4250-ae7a-21d4d29d3084-config-data\") pod \"76f0c9ec-6a29-4250-ae7a-21d4d29d3084\" (UID: \"76f0c9ec-6a29-4250-ae7a-21d4d29d3084\") " Oct 05 21:15:26 crc kubenswrapper[4754]: I1005 21:15:26.929942 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/76f0c9ec-6a29-4250-ae7a-21d4d29d3084-logs" (OuterVolumeSpecName: "logs") pod "76f0c9ec-6a29-4250-ae7a-21d4d29d3084" (UID: "76f0c9ec-6a29-4250-ae7a-21d4d29d3084"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 21:15:26 crc kubenswrapper[4754]: I1005 21:15:26.969295 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/76f0c9ec-6a29-4250-ae7a-21d4d29d3084-kube-api-access-mhfsg" (OuterVolumeSpecName: "kube-api-access-mhfsg") pod "76f0c9ec-6a29-4250-ae7a-21d4d29d3084" (UID: "76f0c9ec-6a29-4250-ae7a-21d4d29d3084"). InnerVolumeSpecName "kube-api-access-mhfsg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:15:27 crc kubenswrapper[4754]: I1005 21:15:27.016143 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76f0c9ec-6a29-4250-ae7a-21d4d29d3084-config-data" (OuterVolumeSpecName: "config-data") pod "76f0c9ec-6a29-4250-ae7a-21d4d29d3084" (UID: "76f0c9ec-6a29-4250-ae7a-21d4d29d3084"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:15:27 crc kubenswrapper[4754]: I1005 21:15:27.027623 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76f0c9ec-6a29-4250-ae7a-21d4d29d3084-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "76f0c9ec-6a29-4250-ae7a-21d4d29d3084" (UID: "76f0c9ec-6a29-4250-ae7a-21d4d29d3084"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:15:27 crc kubenswrapper[4754]: I1005 21:15:27.031108 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22fd2da3-30b4-4e02-92b7-455f64232af5-combined-ca-bundle\") pod \"22fd2da3-30b4-4e02-92b7-455f64232af5\" (UID: \"22fd2da3-30b4-4e02-92b7-455f64232af5\") " Oct 05 21:15:27 crc kubenswrapper[4754]: I1005 21:15:27.031193 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5prtn\" (UniqueName: \"kubernetes.io/projected/22fd2da3-30b4-4e02-92b7-455f64232af5-kube-api-access-5prtn\") pod \"22fd2da3-30b4-4e02-92b7-455f64232af5\" (UID: \"22fd2da3-30b4-4e02-92b7-455f64232af5\") " Oct 05 21:15:27 crc kubenswrapper[4754]: I1005 21:15:27.031325 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/22fd2da3-30b4-4e02-92b7-455f64232af5-config-data\") pod \"22fd2da3-30b4-4e02-92b7-455f64232af5\" (UID: \"22fd2da3-30b4-4e02-92b7-455f64232af5\") " Oct 05 21:15:27 crc kubenswrapper[4754]: I1005 21:15:27.031721 4754 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/76f0c9ec-6a29-4250-ae7a-21d4d29d3084-logs\") on node \"crc\" DevicePath \"\"" Oct 05 21:15:27 crc kubenswrapper[4754]: I1005 21:15:27.031733 4754 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76f0c9ec-6a29-4250-ae7a-21d4d29d3084-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 21:15:27 crc kubenswrapper[4754]: I1005 21:15:27.031745 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mhfsg\" (UniqueName: \"kubernetes.io/projected/76f0c9ec-6a29-4250-ae7a-21d4d29d3084-kube-api-access-mhfsg\") on node \"crc\" DevicePath \"\"" Oct 05 21:15:27 crc kubenswrapper[4754]: I1005 21:15:27.031753 4754 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76f0c9ec-6a29-4250-ae7a-21d4d29d3084-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 21:15:27 crc kubenswrapper[4754]: I1005 21:15:27.042682 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22fd2da3-30b4-4e02-92b7-455f64232af5-kube-api-access-5prtn" (OuterVolumeSpecName: "kube-api-access-5prtn") pod "22fd2da3-30b4-4e02-92b7-455f64232af5" (UID: "22fd2da3-30b4-4e02-92b7-455f64232af5"). InnerVolumeSpecName "kube-api-access-5prtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:15:27 crc kubenswrapper[4754]: I1005 21:15:27.060576 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22fd2da3-30b4-4e02-92b7-455f64232af5-config-data" (OuterVolumeSpecName: "config-data") pod "22fd2da3-30b4-4e02-92b7-455f64232af5" (UID: "22fd2da3-30b4-4e02-92b7-455f64232af5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:15:27 crc kubenswrapper[4754]: I1005 21:15:27.073641 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22fd2da3-30b4-4e02-92b7-455f64232af5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "22fd2da3-30b4-4e02-92b7-455f64232af5" (UID: "22fd2da3-30b4-4e02-92b7-455f64232af5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:15:27 crc kubenswrapper[4754]: I1005 21:15:27.133393 4754 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/22fd2da3-30b4-4e02-92b7-455f64232af5-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 21:15:27 crc kubenswrapper[4754]: I1005 21:15:27.133423 4754 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22fd2da3-30b4-4e02-92b7-455f64232af5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 21:15:27 crc kubenswrapper[4754]: I1005 21:15:27.133436 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5prtn\" (UniqueName: \"kubernetes.io/projected/22fd2da3-30b4-4e02-92b7-455f64232af5-kube-api-access-5prtn\") on node \"crc\" DevicePath \"\"" Oct 05 21:15:27 crc kubenswrapper[4754]: I1005 21:15:27.707460 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 05 21:15:27 crc kubenswrapper[4754]: I1005 21:15:27.707482 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"22fd2da3-30b4-4e02-92b7-455f64232af5","Type":"ContainerDied","Data":"813912eba303928a5cde7b67147b1d2baf5eed70a42ceb5dbf71362655d69e78"} Oct 05 21:15:27 crc kubenswrapper[4754]: I1005 21:15:27.711084 4754 scope.go:117] "RemoveContainer" containerID="f47314e363525c96da1da620ef441697711979100de891a7e34eabd4f6742bde" Oct 05 21:15:27 crc kubenswrapper[4754]: I1005 21:15:27.717167 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 05 21:15:27 crc kubenswrapper[4754]: I1005 21:15:27.717190 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"76f0c9ec-6a29-4250-ae7a-21d4d29d3084","Type":"ContainerDied","Data":"1ccda2c1f8437a08067bf69caf4ae11e2a4978bcc60b6c4fafb416ab84dce2b3"} Oct 05 21:15:27 crc kubenswrapper[4754]: I1005 21:15:27.744805 4754 scope.go:117] "RemoveContainer" containerID="467abfb0ec7ef0fceea5eb09ef62362880eb56a402ab37779344042f10ed0f38" Oct 05 21:15:27 crc kubenswrapper[4754]: I1005 21:15:27.750580 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 05 21:15:27 crc kubenswrapper[4754]: I1005 21:15:27.775596 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 05 21:15:27 crc kubenswrapper[4754]: I1005 21:15:27.783445 4754 scope.go:117] "RemoveContainer" containerID="49d5f2a3799144f8697097a4a091fd392234670f1db76e178a69a345c3bca985" Oct 05 21:15:27 crc kubenswrapper[4754]: I1005 21:15:27.790065 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 05 21:15:27 crc kubenswrapper[4754]: I1005 21:15:27.801581 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 05 21:15:27 crc kubenswrapper[4754]: I1005 21:15:27.814564 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 05 21:15:27 crc kubenswrapper[4754]: E1005 21:15:27.815004 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76f0c9ec-6a29-4250-ae7a-21d4d29d3084" containerName="nova-api-log" Oct 05 21:15:27 crc kubenswrapper[4754]: I1005 21:15:27.815016 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="76f0c9ec-6a29-4250-ae7a-21d4d29d3084" containerName="nova-api-log" Oct 05 21:15:27 crc kubenswrapper[4754]: E1005 21:15:27.815030 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22fd2da3-30b4-4e02-92b7-455f64232af5" containerName="nova-scheduler-scheduler" Oct 05 21:15:27 crc kubenswrapper[4754]: I1005 21:15:27.815037 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="22fd2da3-30b4-4e02-92b7-455f64232af5" containerName="nova-scheduler-scheduler" Oct 05 21:15:27 crc kubenswrapper[4754]: E1005 21:15:27.815051 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76f0c9ec-6a29-4250-ae7a-21d4d29d3084" containerName="nova-api-api" Oct 05 21:15:27 crc kubenswrapper[4754]: I1005 21:15:27.815057 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="76f0c9ec-6a29-4250-ae7a-21d4d29d3084" containerName="nova-api-api" Oct 05 21:15:27 crc kubenswrapper[4754]: I1005 21:15:27.815268 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="22fd2da3-30b4-4e02-92b7-455f64232af5" containerName="nova-scheduler-scheduler" Oct 05 21:15:27 crc kubenswrapper[4754]: I1005 21:15:27.815284 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="76f0c9ec-6a29-4250-ae7a-21d4d29d3084" containerName="nova-api-api" Oct 05 21:15:27 crc kubenswrapper[4754]: I1005 21:15:27.815296 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="76f0c9ec-6a29-4250-ae7a-21d4d29d3084" containerName="nova-api-log" Oct 05 21:15:27 crc kubenswrapper[4754]: I1005 21:15:27.815935 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 05 21:15:27 crc kubenswrapper[4754]: I1005 21:15:27.818406 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 05 21:15:27 crc kubenswrapper[4754]: I1005 21:15:27.835711 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 05 21:15:27 crc kubenswrapper[4754]: I1005 21:15:27.843592 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 05 21:15:27 crc kubenswrapper[4754]: I1005 21:15:27.845277 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 05 21:15:27 crc kubenswrapper[4754]: I1005 21:15:27.858141 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 05 21:15:27 crc kubenswrapper[4754]: I1005 21:15:27.889675 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 05 21:15:27 crc kubenswrapper[4754]: I1005 21:15:27.947865 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b820b69-1578-4f07-9819-246603cdd777-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"4b820b69-1578-4f07-9819-246603cdd777\") " pod="openstack/nova-scheduler-0" Oct 05 21:15:27 crc kubenswrapper[4754]: I1005 21:15:27.947954 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9fz7t\" (UniqueName: \"kubernetes.io/projected/74c27a73-083c-4713-9345-4dc0ebf9a00e-kube-api-access-9fz7t\") pod \"nova-api-0\" (UID: \"74c27a73-083c-4713-9345-4dc0ebf9a00e\") " pod="openstack/nova-api-0" Oct 05 21:15:27 crc kubenswrapper[4754]: I1005 21:15:27.947997 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74c27a73-083c-4713-9345-4dc0ebf9a00e-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"74c27a73-083c-4713-9345-4dc0ebf9a00e\") " pod="openstack/nova-api-0" Oct 05 21:15:27 crc kubenswrapper[4754]: I1005 21:15:27.948038 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b820b69-1578-4f07-9819-246603cdd777-config-data\") pod \"nova-scheduler-0\" (UID: \"4b820b69-1578-4f07-9819-246603cdd777\") " pod="openstack/nova-scheduler-0" Oct 05 21:15:27 crc kubenswrapper[4754]: I1005 21:15:27.948063 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/74c27a73-083c-4713-9345-4dc0ebf9a00e-logs\") pod \"nova-api-0\" (UID: \"74c27a73-083c-4713-9345-4dc0ebf9a00e\") " pod="openstack/nova-api-0" Oct 05 21:15:27 crc kubenswrapper[4754]: I1005 21:15:27.948110 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74c27a73-083c-4713-9345-4dc0ebf9a00e-config-data\") pod \"nova-api-0\" (UID: \"74c27a73-083c-4713-9345-4dc0ebf9a00e\") " pod="openstack/nova-api-0" Oct 05 21:15:27 crc kubenswrapper[4754]: I1005 21:15:27.948163 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-22p2b\" (UniqueName: \"kubernetes.io/projected/4b820b69-1578-4f07-9819-246603cdd777-kube-api-access-22p2b\") pod \"nova-scheduler-0\" (UID: \"4b820b69-1578-4f07-9819-246603cdd777\") " pod="openstack/nova-scheduler-0" Oct 05 21:15:28 crc kubenswrapper[4754]: I1005 21:15:28.045049 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Oct 05 21:15:28 crc kubenswrapper[4754]: I1005 21:15:28.049740 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b820b69-1578-4f07-9819-246603cdd777-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"4b820b69-1578-4f07-9819-246603cdd777\") " pod="openstack/nova-scheduler-0" Oct 05 21:15:28 crc kubenswrapper[4754]: I1005 21:15:28.049783 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9fz7t\" (UniqueName: \"kubernetes.io/projected/74c27a73-083c-4713-9345-4dc0ebf9a00e-kube-api-access-9fz7t\") pod \"nova-api-0\" (UID: \"74c27a73-083c-4713-9345-4dc0ebf9a00e\") " pod="openstack/nova-api-0" Oct 05 21:15:28 crc kubenswrapper[4754]: I1005 21:15:28.049819 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74c27a73-083c-4713-9345-4dc0ebf9a00e-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"74c27a73-083c-4713-9345-4dc0ebf9a00e\") " pod="openstack/nova-api-0" Oct 05 21:15:28 crc kubenswrapper[4754]: I1005 21:15:28.049853 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b820b69-1578-4f07-9819-246603cdd777-config-data\") pod \"nova-scheduler-0\" (UID: \"4b820b69-1578-4f07-9819-246603cdd777\") " pod="openstack/nova-scheduler-0" Oct 05 21:15:28 crc kubenswrapper[4754]: I1005 21:15:28.049876 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/74c27a73-083c-4713-9345-4dc0ebf9a00e-logs\") pod \"nova-api-0\" (UID: \"74c27a73-083c-4713-9345-4dc0ebf9a00e\") " pod="openstack/nova-api-0" Oct 05 21:15:28 crc kubenswrapper[4754]: I1005 21:15:28.049903 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74c27a73-083c-4713-9345-4dc0ebf9a00e-config-data\") pod \"nova-api-0\" (UID: \"74c27a73-083c-4713-9345-4dc0ebf9a00e\") " pod="openstack/nova-api-0" Oct 05 21:15:28 crc kubenswrapper[4754]: I1005 21:15:28.049939 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-22p2b\" (UniqueName: \"kubernetes.io/projected/4b820b69-1578-4f07-9819-246603cdd777-kube-api-access-22p2b\") pod \"nova-scheduler-0\" (UID: \"4b820b69-1578-4f07-9819-246603cdd777\") " pod="openstack/nova-scheduler-0" Oct 05 21:15:28 crc kubenswrapper[4754]: I1005 21:15:28.050467 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/74c27a73-083c-4713-9345-4dc0ebf9a00e-logs\") pod \"nova-api-0\" (UID: \"74c27a73-083c-4713-9345-4dc0ebf9a00e\") " pod="openstack/nova-api-0" Oct 05 21:15:28 crc kubenswrapper[4754]: I1005 21:15:28.054974 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b820b69-1578-4f07-9819-246603cdd777-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"4b820b69-1578-4f07-9819-246603cdd777\") " pod="openstack/nova-scheduler-0" Oct 05 21:15:28 crc kubenswrapper[4754]: I1005 21:15:28.056914 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74c27a73-083c-4713-9345-4dc0ebf9a00e-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"74c27a73-083c-4713-9345-4dc0ebf9a00e\") " pod="openstack/nova-api-0" Oct 05 21:15:28 crc kubenswrapper[4754]: I1005 21:15:28.058523 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b820b69-1578-4f07-9819-246603cdd777-config-data\") pod \"nova-scheduler-0\" (UID: \"4b820b69-1578-4f07-9819-246603cdd777\") " pod="openstack/nova-scheduler-0" Oct 05 21:15:28 crc kubenswrapper[4754]: I1005 21:15:28.069959 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74c27a73-083c-4713-9345-4dc0ebf9a00e-config-data\") pod \"nova-api-0\" (UID: \"74c27a73-083c-4713-9345-4dc0ebf9a00e\") " pod="openstack/nova-api-0" Oct 05 21:15:28 crc kubenswrapper[4754]: I1005 21:15:28.070727 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9fz7t\" (UniqueName: \"kubernetes.io/projected/74c27a73-083c-4713-9345-4dc0ebf9a00e-kube-api-access-9fz7t\") pod \"nova-api-0\" (UID: \"74c27a73-083c-4713-9345-4dc0ebf9a00e\") " pod="openstack/nova-api-0" Oct 05 21:15:28 crc kubenswrapper[4754]: I1005 21:15:28.074406 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-22p2b\" (UniqueName: \"kubernetes.io/projected/4b820b69-1578-4f07-9819-246603cdd777-kube-api-access-22p2b\") pod \"nova-scheduler-0\" (UID: \"4b820b69-1578-4f07-9819-246603cdd777\") " pod="openstack/nova-scheduler-0" Oct 05 21:15:28 crc kubenswrapper[4754]: I1005 21:15:28.134604 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 05 21:15:28 crc kubenswrapper[4754]: I1005 21:15:28.173425 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 05 21:15:28 crc kubenswrapper[4754]: I1005 21:15:28.617922 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 05 21:15:28 crc kubenswrapper[4754]: W1005 21:15:28.620688 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4b820b69_1578_4f07_9819_246603cdd777.slice/crio-f3949d2edb0f65cf87106622eed86f928b5d8ed4e9919654a5be798926a8e8c8 WatchSource:0}: Error finding container f3949d2edb0f65cf87106622eed86f928b5d8ed4e9919654a5be798926a8e8c8: Status 404 returned error can't find the container with id f3949d2edb0f65cf87106622eed86f928b5d8ed4e9919654a5be798926a8e8c8 Oct 05 21:15:28 crc kubenswrapper[4754]: I1005 21:15:28.690592 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 05 21:15:28 crc kubenswrapper[4754]: W1005 21:15:28.709835 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod74c27a73_083c_4713_9345_4dc0ebf9a00e.slice/crio-8cad5f9883aab31b721c9ac8c8226809cfd817ac8b46963d70883f1d9f472f47 WatchSource:0}: Error finding container 8cad5f9883aab31b721c9ac8c8226809cfd817ac8b46963d70883f1d9f472f47: Status 404 returned error can't find the container with id 8cad5f9883aab31b721c9ac8c8226809cfd817ac8b46963d70883f1d9f472f47 Oct 05 21:15:28 crc kubenswrapper[4754]: I1005 21:15:28.729626 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"74c27a73-083c-4713-9345-4dc0ebf9a00e","Type":"ContainerStarted","Data":"8cad5f9883aab31b721c9ac8c8226809cfd817ac8b46963d70883f1d9f472f47"} Oct 05 21:15:28 crc kubenswrapper[4754]: I1005 21:15:28.732136 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"4b820b69-1578-4f07-9819-246603cdd777","Type":"ContainerStarted","Data":"f3949d2edb0f65cf87106622eed86f928b5d8ed4e9919654a5be798926a8e8c8"} Oct 05 21:15:28 crc kubenswrapper[4754]: I1005 21:15:28.885441 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22fd2da3-30b4-4e02-92b7-455f64232af5" path="/var/lib/kubelet/pods/22fd2da3-30b4-4e02-92b7-455f64232af5/volumes" Oct 05 21:15:28 crc kubenswrapper[4754]: I1005 21:15:28.886559 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="76f0c9ec-6a29-4250-ae7a-21d4d29d3084" path="/var/lib/kubelet/pods/76f0c9ec-6a29-4250-ae7a-21d4d29d3084/volumes" Oct 05 21:15:29 crc kubenswrapper[4754]: I1005 21:15:29.762272 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"4b820b69-1578-4f07-9819-246603cdd777","Type":"ContainerStarted","Data":"06e6bd3415112bee3d9524ee46f595f8d3cca1a787b692ef83d0b0bc782a9aae"} Oct 05 21:15:29 crc kubenswrapper[4754]: I1005 21:15:29.766212 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"74c27a73-083c-4713-9345-4dc0ebf9a00e","Type":"ContainerStarted","Data":"414aa4ae6d587a775d944f425af1a445b4ad194a78e1c8fcf619395ec0531910"} Oct 05 21:15:29 crc kubenswrapper[4754]: I1005 21:15:29.766264 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"74c27a73-083c-4713-9345-4dc0ebf9a00e","Type":"ContainerStarted","Data":"535d1c814f0ab307a5a2a202d80fde5006ea9450607082f6ec36dccab9eec72f"} Oct 05 21:15:29 crc kubenswrapper[4754]: I1005 21:15:29.804365 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.804337405 podStartE2EDuration="2.804337405s" podCreationTimestamp="2025-10-05 21:15:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 21:15:29.790051652 +0000 UTC m=+1253.694170402" watchObservedRunningTime="2025-10-05 21:15:29.804337405 +0000 UTC m=+1253.708456145" Oct 05 21:15:29 crc kubenswrapper[4754]: I1005 21:15:29.832314 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.8322818659999998 podStartE2EDuration="2.832281866s" podCreationTimestamp="2025-10-05 21:15:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 21:15:29.826201033 +0000 UTC m=+1253.730319783" watchObservedRunningTime="2025-10-05 21:15:29.832281866 +0000 UTC m=+1253.736400606" Oct 05 21:15:30 crc kubenswrapper[4754]: I1005 21:15:30.045095 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 05 21:15:30 crc kubenswrapper[4754]: I1005 21:15:30.045937 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 05 21:15:33 crc kubenswrapper[4754]: I1005 21:15:33.135298 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 05 21:15:35 crc kubenswrapper[4754]: I1005 21:15:35.045340 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 05 21:15:35 crc kubenswrapper[4754]: I1005 21:15:35.045846 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 05 21:15:36 crc kubenswrapper[4754]: I1005 21:15:36.061676 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="c45e75d9-997b-4b2b-a923-8ee9739fcbbc" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.196:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 05 21:15:36 crc kubenswrapper[4754]: I1005 21:15:36.061722 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="c45e75d9-997b-4b2b-a923-8ee9739fcbbc" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.196:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 05 21:15:38 crc kubenswrapper[4754]: I1005 21:15:38.134796 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 05 21:15:38 crc kubenswrapper[4754]: I1005 21:15:38.162352 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-68c6b67864-9msm5" Oct 05 21:15:38 crc kubenswrapper[4754]: I1005 21:15:38.177229 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 05 21:15:38 crc kubenswrapper[4754]: I1005 21:15:38.189647 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 05 21:15:38 crc kubenswrapper[4754]: I1005 21:15:38.220916 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 05 21:15:38 crc kubenswrapper[4754]: I1005 21:15:38.247310 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-574454d6cb-vbnk6" Oct 05 21:15:38 crc kubenswrapper[4754]: I1005 21:15:38.888803 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 05 21:15:39 crc kubenswrapper[4754]: I1005 21:15:39.259728 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="74c27a73-083c-4713-9345-4dc0ebf9a00e" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.198:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 05 21:15:39 crc kubenswrapper[4754]: I1005 21:15:39.259745 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="74c27a73-083c-4713-9345-4dc0ebf9a00e" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.198:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 05 21:15:40 crc kubenswrapper[4754]: I1005 21:15:40.386221 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-68c6b67864-9msm5" Oct 05 21:15:40 crc kubenswrapper[4754]: I1005 21:15:40.958157 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-574454d6cb-vbnk6" Oct 05 21:15:41 crc kubenswrapper[4754]: I1005 21:15:41.072777 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-68c6b67864-9msm5"] Oct 05 21:15:41 crc kubenswrapper[4754]: I1005 21:15:41.073097 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-68c6b67864-9msm5" podUID="27cd6be5-f5eb-4158-bdd1-37f4df3d530d" containerName="horizon-log" containerID="cri-o://48ce6ae318ddb68e70258170c00936a21c53d9535eb9576fac3f4ad34120b991" gracePeriod=30 Oct 05 21:15:41 crc kubenswrapper[4754]: I1005 21:15:41.073796 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-68c6b67864-9msm5" podUID="27cd6be5-f5eb-4158-bdd1-37f4df3d530d" containerName="horizon" containerID="cri-o://e8c8cc4de7167b842bac5c879a08a541081eaf3a777d19676eea7ed48c8eae17" gracePeriod=30 Oct 05 21:15:44 crc kubenswrapper[4754]: I1005 21:15:44.864340 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 05 21:15:44 crc kubenswrapper[4754]: I1005 21:15:44.939404 4754 generic.go:334] "Generic (PLEG): container finished" podID="746c64fc-7a41-440f-894f-79af6edcc9ee" containerID="cd4f36a301547763e21743d5818ec98066ba379572e58a96b42a2076dabd0c67" exitCode=137 Oct 05 21:15:44 crc kubenswrapper[4754]: I1005 21:15:44.939460 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 05 21:15:44 crc kubenswrapper[4754]: I1005 21:15:44.939539 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"746c64fc-7a41-440f-894f-79af6edcc9ee","Type":"ContainerDied","Data":"cd4f36a301547763e21743d5818ec98066ba379572e58a96b42a2076dabd0c67"} Oct 05 21:15:44 crc kubenswrapper[4754]: I1005 21:15:44.939600 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"746c64fc-7a41-440f-894f-79af6edcc9ee","Type":"ContainerDied","Data":"cf2fa3aa5225a0d8cc439448a9af9c63aa870c32a91d9202f3fe8b79040feb33"} Oct 05 21:15:44 crc kubenswrapper[4754]: I1005 21:15:44.939622 4754 scope.go:117] "RemoveContainer" containerID="cd4f36a301547763e21743d5818ec98066ba379572e58a96b42a2076dabd0c67" Oct 05 21:15:44 crc kubenswrapper[4754]: I1005 21:15:44.944695 4754 generic.go:334] "Generic (PLEG): container finished" podID="27cd6be5-f5eb-4158-bdd1-37f4df3d530d" containerID="e8c8cc4de7167b842bac5c879a08a541081eaf3a777d19676eea7ed48c8eae17" exitCode=0 Oct 05 21:15:44 crc kubenswrapper[4754]: I1005 21:15:44.944736 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-68c6b67864-9msm5" event={"ID":"27cd6be5-f5eb-4158-bdd1-37f4df3d530d","Type":"ContainerDied","Data":"e8c8cc4de7167b842bac5c879a08a541081eaf3a777d19676eea7ed48c8eae17"} Oct 05 21:15:44 crc kubenswrapper[4754]: I1005 21:15:44.950870 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/746c64fc-7a41-440f-894f-79af6edcc9ee-combined-ca-bundle\") pod \"746c64fc-7a41-440f-894f-79af6edcc9ee\" (UID: \"746c64fc-7a41-440f-894f-79af6edcc9ee\") " Oct 05 21:15:44 crc kubenswrapper[4754]: I1005 21:15:44.951015 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d27z7\" (UniqueName: \"kubernetes.io/projected/746c64fc-7a41-440f-894f-79af6edcc9ee-kube-api-access-d27z7\") pod \"746c64fc-7a41-440f-894f-79af6edcc9ee\" (UID: \"746c64fc-7a41-440f-894f-79af6edcc9ee\") " Oct 05 21:15:44 crc kubenswrapper[4754]: I1005 21:15:44.951068 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/746c64fc-7a41-440f-894f-79af6edcc9ee-config-data\") pod \"746c64fc-7a41-440f-894f-79af6edcc9ee\" (UID: \"746c64fc-7a41-440f-894f-79af6edcc9ee\") " Oct 05 21:15:44 crc kubenswrapper[4754]: I1005 21:15:44.970501 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/746c64fc-7a41-440f-894f-79af6edcc9ee-kube-api-access-d27z7" (OuterVolumeSpecName: "kube-api-access-d27z7") pod "746c64fc-7a41-440f-894f-79af6edcc9ee" (UID: "746c64fc-7a41-440f-894f-79af6edcc9ee"). InnerVolumeSpecName "kube-api-access-d27z7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:15:44 crc kubenswrapper[4754]: I1005 21:15:44.971941 4754 scope.go:117] "RemoveContainer" containerID="cd4f36a301547763e21743d5818ec98066ba379572e58a96b42a2076dabd0c67" Oct 05 21:15:44 crc kubenswrapper[4754]: E1005 21:15:44.972835 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cd4f36a301547763e21743d5818ec98066ba379572e58a96b42a2076dabd0c67\": container with ID starting with cd4f36a301547763e21743d5818ec98066ba379572e58a96b42a2076dabd0c67 not found: ID does not exist" containerID="cd4f36a301547763e21743d5818ec98066ba379572e58a96b42a2076dabd0c67" Oct 05 21:15:44 crc kubenswrapper[4754]: I1005 21:15:44.972879 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cd4f36a301547763e21743d5818ec98066ba379572e58a96b42a2076dabd0c67"} err="failed to get container status \"cd4f36a301547763e21743d5818ec98066ba379572e58a96b42a2076dabd0c67\": rpc error: code = NotFound desc = could not find container \"cd4f36a301547763e21743d5818ec98066ba379572e58a96b42a2076dabd0c67\": container with ID starting with cd4f36a301547763e21743d5818ec98066ba379572e58a96b42a2076dabd0c67 not found: ID does not exist" Oct 05 21:15:44 crc kubenswrapper[4754]: I1005 21:15:44.972908 4754 scope.go:117] "RemoveContainer" containerID="8cbf678dfb968e1a5f68bc5b282346cef6ff3c4725e9f27e572e9658e525228e" Oct 05 21:15:44 crc kubenswrapper[4754]: I1005 21:15:44.993229 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/746c64fc-7a41-440f-894f-79af6edcc9ee-config-data" (OuterVolumeSpecName: "config-data") pod "746c64fc-7a41-440f-894f-79af6edcc9ee" (UID: "746c64fc-7a41-440f-894f-79af6edcc9ee"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:15:44 crc kubenswrapper[4754]: I1005 21:15:44.997129 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/746c64fc-7a41-440f-894f-79af6edcc9ee-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "746c64fc-7a41-440f-894f-79af6edcc9ee" (UID: "746c64fc-7a41-440f-894f-79af6edcc9ee"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:15:45 crc kubenswrapper[4754]: I1005 21:15:45.052821 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 05 21:15:45 crc kubenswrapper[4754]: I1005 21:15:45.053825 4754 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/746c64fc-7a41-440f-894f-79af6edcc9ee-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 21:15:45 crc kubenswrapper[4754]: I1005 21:15:45.053849 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d27z7\" (UniqueName: \"kubernetes.io/projected/746c64fc-7a41-440f-894f-79af6edcc9ee-kube-api-access-d27z7\") on node \"crc\" DevicePath \"\"" Oct 05 21:15:45 crc kubenswrapper[4754]: I1005 21:15:45.053860 4754 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/746c64fc-7a41-440f-894f-79af6edcc9ee-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 21:15:45 crc kubenswrapper[4754]: I1005 21:15:45.058424 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 05 21:15:45 crc kubenswrapper[4754]: I1005 21:15:45.059798 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 05 21:15:45 crc kubenswrapper[4754]: I1005 21:15:45.279055 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 05 21:15:45 crc kubenswrapper[4754]: I1005 21:15:45.320125 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 05 21:15:45 crc kubenswrapper[4754]: I1005 21:15:45.332038 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 05 21:15:45 crc kubenswrapper[4754]: E1005 21:15:45.332559 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="746c64fc-7a41-440f-894f-79af6edcc9ee" containerName="nova-cell1-novncproxy-novncproxy" Oct 05 21:15:45 crc kubenswrapper[4754]: I1005 21:15:45.332583 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="746c64fc-7a41-440f-894f-79af6edcc9ee" containerName="nova-cell1-novncproxy-novncproxy" Oct 05 21:15:45 crc kubenswrapper[4754]: I1005 21:15:45.332860 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="746c64fc-7a41-440f-894f-79af6edcc9ee" containerName="nova-cell1-novncproxy-novncproxy" Oct 05 21:15:45 crc kubenswrapper[4754]: I1005 21:15:45.333732 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 05 21:15:45 crc kubenswrapper[4754]: I1005 21:15:45.338206 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Oct 05 21:15:45 crc kubenswrapper[4754]: I1005 21:15:45.338405 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Oct 05 21:15:45 crc kubenswrapper[4754]: I1005 21:15:45.338576 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Oct 05 21:15:45 crc kubenswrapper[4754]: I1005 21:15:45.361663 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 05 21:15:45 crc kubenswrapper[4754]: I1005 21:15:45.463987 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eaf71ba6-38b3-4dd1-98e7-e3e5e5d9e428-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"eaf71ba6-38b3-4dd1-98e7-e3e5e5d9e428\") " pod="openstack/nova-cell1-novncproxy-0" Oct 05 21:15:45 crc kubenswrapper[4754]: I1005 21:15:45.464044 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/eaf71ba6-38b3-4dd1-98e7-e3e5e5d9e428-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"eaf71ba6-38b3-4dd1-98e7-e3e5e5d9e428\") " pod="openstack/nova-cell1-novncproxy-0" Oct 05 21:15:45 crc kubenswrapper[4754]: I1005 21:15:45.464092 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/eaf71ba6-38b3-4dd1-98e7-e3e5e5d9e428-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"eaf71ba6-38b3-4dd1-98e7-e3e5e5d9e428\") " pod="openstack/nova-cell1-novncproxy-0" Oct 05 21:15:45 crc kubenswrapper[4754]: I1005 21:15:45.464295 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eaf71ba6-38b3-4dd1-98e7-e3e5e5d9e428-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"eaf71ba6-38b3-4dd1-98e7-e3e5e5d9e428\") " pod="openstack/nova-cell1-novncproxy-0" Oct 05 21:15:45 crc kubenswrapper[4754]: I1005 21:15:45.464481 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bcmdm\" (UniqueName: \"kubernetes.io/projected/eaf71ba6-38b3-4dd1-98e7-e3e5e5d9e428-kube-api-access-bcmdm\") pod \"nova-cell1-novncproxy-0\" (UID: \"eaf71ba6-38b3-4dd1-98e7-e3e5e5d9e428\") " pod="openstack/nova-cell1-novncproxy-0" Oct 05 21:15:45 crc kubenswrapper[4754]: I1005 21:15:45.566600 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bcmdm\" (UniqueName: \"kubernetes.io/projected/eaf71ba6-38b3-4dd1-98e7-e3e5e5d9e428-kube-api-access-bcmdm\") pod \"nova-cell1-novncproxy-0\" (UID: \"eaf71ba6-38b3-4dd1-98e7-e3e5e5d9e428\") " pod="openstack/nova-cell1-novncproxy-0" Oct 05 21:15:45 crc kubenswrapper[4754]: I1005 21:15:45.566978 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eaf71ba6-38b3-4dd1-98e7-e3e5e5d9e428-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"eaf71ba6-38b3-4dd1-98e7-e3e5e5d9e428\") " pod="openstack/nova-cell1-novncproxy-0" Oct 05 21:15:45 crc kubenswrapper[4754]: I1005 21:15:45.567011 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/eaf71ba6-38b3-4dd1-98e7-e3e5e5d9e428-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"eaf71ba6-38b3-4dd1-98e7-e3e5e5d9e428\") " pod="openstack/nova-cell1-novncproxy-0" Oct 05 21:15:45 crc kubenswrapper[4754]: I1005 21:15:45.567051 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/eaf71ba6-38b3-4dd1-98e7-e3e5e5d9e428-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"eaf71ba6-38b3-4dd1-98e7-e3e5e5d9e428\") " pod="openstack/nova-cell1-novncproxy-0" Oct 05 21:15:45 crc kubenswrapper[4754]: I1005 21:15:45.567075 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eaf71ba6-38b3-4dd1-98e7-e3e5e5d9e428-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"eaf71ba6-38b3-4dd1-98e7-e3e5e5d9e428\") " pod="openstack/nova-cell1-novncproxy-0" Oct 05 21:15:45 crc kubenswrapper[4754]: I1005 21:15:45.571562 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eaf71ba6-38b3-4dd1-98e7-e3e5e5d9e428-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"eaf71ba6-38b3-4dd1-98e7-e3e5e5d9e428\") " pod="openstack/nova-cell1-novncproxy-0" Oct 05 21:15:45 crc kubenswrapper[4754]: I1005 21:15:45.571716 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/eaf71ba6-38b3-4dd1-98e7-e3e5e5d9e428-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"eaf71ba6-38b3-4dd1-98e7-e3e5e5d9e428\") " pod="openstack/nova-cell1-novncproxy-0" Oct 05 21:15:45 crc kubenswrapper[4754]: I1005 21:15:45.572240 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/eaf71ba6-38b3-4dd1-98e7-e3e5e5d9e428-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"eaf71ba6-38b3-4dd1-98e7-e3e5e5d9e428\") " pod="openstack/nova-cell1-novncproxy-0" Oct 05 21:15:45 crc kubenswrapper[4754]: I1005 21:15:45.575828 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eaf71ba6-38b3-4dd1-98e7-e3e5e5d9e428-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"eaf71ba6-38b3-4dd1-98e7-e3e5e5d9e428\") " pod="openstack/nova-cell1-novncproxy-0" Oct 05 21:15:45 crc kubenswrapper[4754]: I1005 21:15:45.583296 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bcmdm\" (UniqueName: \"kubernetes.io/projected/eaf71ba6-38b3-4dd1-98e7-e3e5e5d9e428-kube-api-access-bcmdm\") pod \"nova-cell1-novncproxy-0\" (UID: \"eaf71ba6-38b3-4dd1-98e7-e3e5e5d9e428\") " pod="openstack/nova-cell1-novncproxy-0" Oct 05 21:15:45 crc kubenswrapper[4754]: I1005 21:15:45.604440 4754 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-68c6b67864-9msm5" podUID="27cd6be5-f5eb-4158-bdd1-37f4df3d530d" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.148:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.148:8443: connect: connection refused" Oct 05 21:15:45 crc kubenswrapper[4754]: I1005 21:15:45.681634 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 05 21:15:46 crc kubenswrapper[4754]: I1005 21:15:46.005987 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 05 21:15:46 crc kubenswrapper[4754]: I1005 21:15:46.011983 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 05 21:15:46 crc kubenswrapper[4754]: I1005 21:15:46.851189 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="746c64fc-7a41-440f-894f-79af6edcc9ee" path="/var/lib/kubelet/pods/746c64fc-7a41-440f-894f-79af6edcc9ee/volumes" Oct 05 21:15:47 crc kubenswrapper[4754]: I1005 21:15:47.020350 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"eaf71ba6-38b3-4dd1-98e7-e3e5e5d9e428","Type":"ContainerStarted","Data":"4281166fcf775329279ee6d46c62609dc8e8437b191fefee716d9ecb152f2cbf"} Oct 05 21:15:47 crc kubenswrapper[4754]: I1005 21:15:47.020402 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"eaf71ba6-38b3-4dd1-98e7-e3e5e5d9e428","Type":"ContainerStarted","Data":"f8901a47f4813a2f366032e046f8f77eef2e9bc6ba94d6fbcfba03374b29156d"} Oct 05 21:15:47 crc kubenswrapper[4754]: I1005 21:15:47.048720 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.048693726 podStartE2EDuration="2.048693726s" podCreationTimestamp="2025-10-05 21:15:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 21:15:47.045816989 +0000 UTC m=+1270.949935699" watchObservedRunningTime="2025-10-05 21:15:47.048693726 +0000 UTC m=+1270.952812426" Oct 05 21:15:48 crc kubenswrapper[4754]: I1005 21:15:48.180041 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 05 21:15:48 crc kubenswrapper[4754]: I1005 21:15:48.181144 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 05 21:15:48 crc kubenswrapper[4754]: I1005 21:15:48.188778 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 05 21:15:48 crc kubenswrapper[4754]: I1005 21:15:48.188872 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 05 21:15:49 crc kubenswrapper[4754]: I1005 21:15:49.062961 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 05 21:15:49 crc kubenswrapper[4754]: I1005 21:15:49.069732 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 05 21:15:49 crc kubenswrapper[4754]: I1005 21:15:49.306223 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-jlwmm"] Oct 05 21:15:49 crc kubenswrapper[4754]: I1005 21:15:49.308626 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59cf4bdb65-jlwmm" Oct 05 21:15:49 crc kubenswrapper[4754]: I1005 21:15:49.336036 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-jlwmm"] Oct 05 21:15:49 crc kubenswrapper[4754]: I1005 21:15:49.471597 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zb949\" (UniqueName: \"kubernetes.io/projected/ceb7a5cb-e7f2-4162-bd2b-7518c3ad4d52-kube-api-access-zb949\") pod \"dnsmasq-dns-59cf4bdb65-jlwmm\" (UID: \"ceb7a5cb-e7f2-4162-bd2b-7518c3ad4d52\") " pod="openstack/dnsmasq-dns-59cf4bdb65-jlwmm" Oct 05 21:15:49 crc kubenswrapper[4754]: I1005 21:15:49.471650 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ceb7a5cb-e7f2-4162-bd2b-7518c3ad4d52-ovsdbserver-sb\") pod \"dnsmasq-dns-59cf4bdb65-jlwmm\" (UID: \"ceb7a5cb-e7f2-4162-bd2b-7518c3ad4d52\") " pod="openstack/dnsmasq-dns-59cf4bdb65-jlwmm" Oct 05 21:15:49 crc kubenswrapper[4754]: I1005 21:15:49.471810 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ceb7a5cb-e7f2-4162-bd2b-7518c3ad4d52-ovsdbserver-nb\") pod \"dnsmasq-dns-59cf4bdb65-jlwmm\" (UID: \"ceb7a5cb-e7f2-4162-bd2b-7518c3ad4d52\") " pod="openstack/dnsmasq-dns-59cf4bdb65-jlwmm" Oct 05 21:15:49 crc kubenswrapper[4754]: I1005 21:15:49.471960 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ceb7a5cb-e7f2-4162-bd2b-7518c3ad4d52-config\") pod \"dnsmasq-dns-59cf4bdb65-jlwmm\" (UID: \"ceb7a5cb-e7f2-4162-bd2b-7518c3ad4d52\") " pod="openstack/dnsmasq-dns-59cf4bdb65-jlwmm" Oct 05 21:15:49 crc kubenswrapper[4754]: I1005 21:15:49.472027 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ceb7a5cb-e7f2-4162-bd2b-7518c3ad4d52-dns-swift-storage-0\") pod \"dnsmasq-dns-59cf4bdb65-jlwmm\" (UID: \"ceb7a5cb-e7f2-4162-bd2b-7518c3ad4d52\") " pod="openstack/dnsmasq-dns-59cf4bdb65-jlwmm" Oct 05 21:15:49 crc kubenswrapper[4754]: I1005 21:15:49.472216 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ceb7a5cb-e7f2-4162-bd2b-7518c3ad4d52-dns-svc\") pod \"dnsmasq-dns-59cf4bdb65-jlwmm\" (UID: \"ceb7a5cb-e7f2-4162-bd2b-7518c3ad4d52\") " pod="openstack/dnsmasq-dns-59cf4bdb65-jlwmm" Oct 05 21:15:49 crc kubenswrapper[4754]: I1005 21:15:49.574720 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zb949\" (UniqueName: \"kubernetes.io/projected/ceb7a5cb-e7f2-4162-bd2b-7518c3ad4d52-kube-api-access-zb949\") pod \"dnsmasq-dns-59cf4bdb65-jlwmm\" (UID: \"ceb7a5cb-e7f2-4162-bd2b-7518c3ad4d52\") " pod="openstack/dnsmasq-dns-59cf4bdb65-jlwmm" Oct 05 21:15:49 crc kubenswrapper[4754]: I1005 21:15:49.574777 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ceb7a5cb-e7f2-4162-bd2b-7518c3ad4d52-ovsdbserver-sb\") pod \"dnsmasq-dns-59cf4bdb65-jlwmm\" (UID: \"ceb7a5cb-e7f2-4162-bd2b-7518c3ad4d52\") " pod="openstack/dnsmasq-dns-59cf4bdb65-jlwmm" Oct 05 21:15:49 crc kubenswrapper[4754]: I1005 21:15:49.574817 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ceb7a5cb-e7f2-4162-bd2b-7518c3ad4d52-ovsdbserver-nb\") pod \"dnsmasq-dns-59cf4bdb65-jlwmm\" (UID: \"ceb7a5cb-e7f2-4162-bd2b-7518c3ad4d52\") " pod="openstack/dnsmasq-dns-59cf4bdb65-jlwmm" Oct 05 21:15:49 crc kubenswrapper[4754]: I1005 21:15:49.574861 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ceb7a5cb-e7f2-4162-bd2b-7518c3ad4d52-config\") pod \"dnsmasq-dns-59cf4bdb65-jlwmm\" (UID: \"ceb7a5cb-e7f2-4162-bd2b-7518c3ad4d52\") " pod="openstack/dnsmasq-dns-59cf4bdb65-jlwmm" Oct 05 21:15:49 crc kubenswrapper[4754]: I1005 21:15:49.574911 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ceb7a5cb-e7f2-4162-bd2b-7518c3ad4d52-dns-swift-storage-0\") pod \"dnsmasq-dns-59cf4bdb65-jlwmm\" (UID: \"ceb7a5cb-e7f2-4162-bd2b-7518c3ad4d52\") " pod="openstack/dnsmasq-dns-59cf4bdb65-jlwmm" Oct 05 21:15:49 crc kubenswrapper[4754]: I1005 21:15:49.575998 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ceb7a5cb-e7f2-4162-bd2b-7518c3ad4d52-config\") pod \"dnsmasq-dns-59cf4bdb65-jlwmm\" (UID: \"ceb7a5cb-e7f2-4162-bd2b-7518c3ad4d52\") " pod="openstack/dnsmasq-dns-59cf4bdb65-jlwmm" Oct 05 21:15:49 crc kubenswrapper[4754]: I1005 21:15:49.576015 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ceb7a5cb-e7f2-4162-bd2b-7518c3ad4d52-ovsdbserver-nb\") pod \"dnsmasq-dns-59cf4bdb65-jlwmm\" (UID: \"ceb7a5cb-e7f2-4162-bd2b-7518c3ad4d52\") " pod="openstack/dnsmasq-dns-59cf4bdb65-jlwmm" Oct 05 21:15:49 crc kubenswrapper[4754]: I1005 21:15:49.576082 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ceb7a5cb-e7f2-4162-bd2b-7518c3ad4d52-dns-swift-storage-0\") pod \"dnsmasq-dns-59cf4bdb65-jlwmm\" (UID: \"ceb7a5cb-e7f2-4162-bd2b-7518c3ad4d52\") " pod="openstack/dnsmasq-dns-59cf4bdb65-jlwmm" Oct 05 21:15:49 crc kubenswrapper[4754]: I1005 21:15:49.576128 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ceb7a5cb-e7f2-4162-bd2b-7518c3ad4d52-dns-svc\") pod \"dnsmasq-dns-59cf4bdb65-jlwmm\" (UID: \"ceb7a5cb-e7f2-4162-bd2b-7518c3ad4d52\") " pod="openstack/dnsmasq-dns-59cf4bdb65-jlwmm" Oct 05 21:15:49 crc kubenswrapper[4754]: I1005 21:15:49.576256 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ceb7a5cb-e7f2-4162-bd2b-7518c3ad4d52-ovsdbserver-sb\") pod \"dnsmasq-dns-59cf4bdb65-jlwmm\" (UID: \"ceb7a5cb-e7f2-4162-bd2b-7518c3ad4d52\") " pod="openstack/dnsmasq-dns-59cf4bdb65-jlwmm" Oct 05 21:15:49 crc kubenswrapper[4754]: I1005 21:15:49.576751 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ceb7a5cb-e7f2-4162-bd2b-7518c3ad4d52-dns-svc\") pod \"dnsmasq-dns-59cf4bdb65-jlwmm\" (UID: \"ceb7a5cb-e7f2-4162-bd2b-7518c3ad4d52\") " pod="openstack/dnsmasq-dns-59cf4bdb65-jlwmm" Oct 05 21:15:49 crc kubenswrapper[4754]: I1005 21:15:49.603336 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zb949\" (UniqueName: \"kubernetes.io/projected/ceb7a5cb-e7f2-4162-bd2b-7518c3ad4d52-kube-api-access-zb949\") pod \"dnsmasq-dns-59cf4bdb65-jlwmm\" (UID: \"ceb7a5cb-e7f2-4162-bd2b-7518c3ad4d52\") " pod="openstack/dnsmasq-dns-59cf4bdb65-jlwmm" Oct 05 21:15:49 crc kubenswrapper[4754]: I1005 21:15:49.636803 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59cf4bdb65-jlwmm" Oct 05 21:15:50 crc kubenswrapper[4754]: I1005 21:15:50.159218 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-jlwmm"] Oct 05 21:15:50 crc kubenswrapper[4754]: I1005 21:15:50.682797 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Oct 05 21:15:51 crc kubenswrapper[4754]: I1005 21:15:51.079244 4754 generic.go:334] "Generic (PLEG): container finished" podID="ceb7a5cb-e7f2-4162-bd2b-7518c3ad4d52" containerID="f871f68953da0289a85d10ec6082b36f55672e3b4a72fc715962828c8d78ade1" exitCode=0 Oct 05 21:15:51 crc kubenswrapper[4754]: I1005 21:15:51.080758 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cf4bdb65-jlwmm" event={"ID":"ceb7a5cb-e7f2-4162-bd2b-7518c3ad4d52","Type":"ContainerDied","Data":"f871f68953da0289a85d10ec6082b36f55672e3b4a72fc715962828c8d78ade1"} Oct 05 21:15:51 crc kubenswrapper[4754]: I1005 21:15:51.080789 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cf4bdb65-jlwmm" event={"ID":"ceb7a5cb-e7f2-4162-bd2b-7518c3ad4d52","Type":"ContainerStarted","Data":"04bce3bfe7f2909d600d2f876cfed3d07c0847721902ce559cdc63be292a0a1c"} Oct 05 21:15:52 crc kubenswrapper[4754]: I1005 21:15:52.096665 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cf4bdb65-jlwmm" event={"ID":"ceb7a5cb-e7f2-4162-bd2b-7518c3ad4d52","Type":"ContainerStarted","Data":"047bf50f6e906707a0a4e6953b750c33bfcddcbcf227b31d1f3ccb0f26a6997c"} Oct 05 21:15:52 crc kubenswrapper[4754]: I1005 21:15:52.097022 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-59cf4bdb65-jlwmm" Oct 05 21:15:52 crc kubenswrapper[4754]: I1005 21:15:52.136780 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-59cf4bdb65-jlwmm" podStartSLOduration=3.1354734300000002 podStartE2EDuration="3.13547343s" podCreationTimestamp="2025-10-05 21:15:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 21:15:52.126512159 +0000 UTC m=+1276.030630869" watchObservedRunningTime="2025-10-05 21:15:52.13547343 +0000 UTC m=+1276.039592150" Oct 05 21:15:52 crc kubenswrapper[4754]: I1005 21:15:52.293507 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 05 21:15:52 crc kubenswrapper[4754]: I1005 21:15:52.293877 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="6ee3b3b2-07e1-4bab-9372-893eef07f250" containerName="ceilometer-central-agent" containerID="cri-o://3380e04f976d4766b1f52c3232e01620a9cfc06705ef8c5c6004c0e41c7f6640" gracePeriod=30 Oct 05 21:15:52 crc kubenswrapper[4754]: I1005 21:15:52.294038 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="6ee3b3b2-07e1-4bab-9372-893eef07f250" containerName="proxy-httpd" containerID="cri-o://9ed026943be03e0a1a0532c59c6308a6e590ffc346f67cf516ee3fa829679dab" gracePeriod=30 Oct 05 21:15:52 crc kubenswrapper[4754]: I1005 21:15:52.294092 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="6ee3b3b2-07e1-4bab-9372-893eef07f250" containerName="sg-core" containerID="cri-o://58202d1eec91468517c6840252471678191af733282dd96c7961df59850a5bc7" gracePeriod=30 Oct 05 21:15:52 crc kubenswrapper[4754]: I1005 21:15:52.294133 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="6ee3b3b2-07e1-4bab-9372-893eef07f250" containerName="ceilometer-notification-agent" containerID="cri-o://c1ce65ecb8ae3e72e860298633e5f6d77759ea8ef5efa531ec34ad976df4e63a" gracePeriod=30 Oct 05 21:15:52 crc kubenswrapper[4754]: I1005 21:15:52.442044 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 05 21:15:52 crc kubenswrapper[4754]: I1005 21:15:52.442327 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="74c27a73-083c-4713-9345-4dc0ebf9a00e" containerName="nova-api-api" containerID="cri-o://414aa4ae6d587a775d944f425af1a445b4ad194a78e1c8fcf619395ec0531910" gracePeriod=30 Oct 05 21:15:52 crc kubenswrapper[4754]: I1005 21:15:52.442503 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="74c27a73-083c-4713-9345-4dc0ebf9a00e" containerName="nova-api-log" containerID="cri-o://535d1c814f0ab307a5a2a202d80fde5006ea9450607082f6ec36dccab9eec72f" gracePeriod=30 Oct 05 21:15:53 crc kubenswrapper[4754]: I1005 21:15:53.111267 4754 generic.go:334] "Generic (PLEG): container finished" podID="6ee3b3b2-07e1-4bab-9372-893eef07f250" containerID="9ed026943be03e0a1a0532c59c6308a6e590ffc346f67cf516ee3fa829679dab" exitCode=0 Oct 05 21:15:53 crc kubenswrapper[4754]: I1005 21:15:53.111747 4754 generic.go:334] "Generic (PLEG): container finished" podID="6ee3b3b2-07e1-4bab-9372-893eef07f250" containerID="58202d1eec91468517c6840252471678191af733282dd96c7961df59850a5bc7" exitCode=2 Oct 05 21:15:53 crc kubenswrapper[4754]: I1005 21:15:53.111759 4754 generic.go:334] "Generic (PLEG): container finished" podID="6ee3b3b2-07e1-4bab-9372-893eef07f250" containerID="3380e04f976d4766b1f52c3232e01620a9cfc06705ef8c5c6004c0e41c7f6640" exitCode=0 Oct 05 21:15:53 crc kubenswrapper[4754]: I1005 21:15:53.111335 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6ee3b3b2-07e1-4bab-9372-893eef07f250","Type":"ContainerDied","Data":"9ed026943be03e0a1a0532c59c6308a6e590ffc346f67cf516ee3fa829679dab"} Oct 05 21:15:53 crc kubenswrapper[4754]: I1005 21:15:53.111834 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6ee3b3b2-07e1-4bab-9372-893eef07f250","Type":"ContainerDied","Data":"58202d1eec91468517c6840252471678191af733282dd96c7961df59850a5bc7"} Oct 05 21:15:53 crc kubenswrapper[4754]: I1005 21:15:53.111850 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6ee3b3b2-07e1-4bab-9372-893eef07f250","Type":"ContainerDied","Data":"3380e04f976d4766b1f52c3232e01620a9cfc06705ef8c5c6004c0e41c7f6640"} Oct 05 21:15:53 crc kubenswrapper[4754]: I1005 21:15:53.113447 4754 generic.go:334] "Generic (PLEG): container finished" podID="74c27a73-083c-4713-9345-4dc0ebf9a00e" containerID="535d1c814f0ab307a5a2a202d80fde5006ea9450607082f6ec36dccab9eec72f" exitCode=143 Oct 05 21:15:53 crc kubenswrapper[4754]: I1005 21:15:53.113537 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"74c27a73-083c-4713-9345-4dc0ebf9a00e","Type":"ContainerDied","Data":"535d1c814f0ab307a5a2a202d80fde5006ea9450607082f6ec36dccab9eec72f"} Oct 05 21:15:53 crc kubenswrapper[4754]: I1005 21:15:53.808482 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 05 21:15:53 crc kubenswrapper[4754]: I1005 21:15:53.975828 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6ee3b3b2-07e1-4bab-9372-893eef07f250-log-httpd\") pod \"6ee3b3b2-07e1-4bab-9372-893eef07f250\" (UID: \"6ee3b3b2-07e1-4bab-9372-893eef07f250\") " Oct 05 21:15:53 crc kubenswrapper[4754]: I1005 21:15:53.975896 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6ee3b3b2-07e1-4bab-9372-893eef07f250-scripts\") pod \"6ee3b3b2-07e1-4bab-9372-893eef07f250\" (UID: \"6ee3b3b2-07e1-4bab-9372-893eef07f250\") " Oct 05 21:15:53 crc kubenswrapper[4754]: I1005 21:15:53.975961 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6ee3b3b2-07e1-4bab-9372-893eef07f250-run-httpd\") pod \"6ee3b3b2-07e1-4bab-9372-893eef07f250\" (UID: \"6ee3b3b2-07e1-4bab-9372-893eef07f250\") " Oct 05 21:15:53 crc kubenswrapper[4754]: I1005 21:15:53.975993 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ee3b3b2-07e1-4bab-9372-893eef07f250-combined-ca-bundle\") pod \"6ee3b3b2-07e1-4bab-9372-893eef07f250\" (UID: \"6ee3b3b2-07e1-4bab-9372-893eef07f250\") " Oct 05 21:15:53 crc kubenswrapper[4754]: I1005 21:15:53.976121 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6ee3b3b2-07e1-4bab-9372-893eef07f250-config-data\") pod \"6ee3b3b2-07e1-4bab-9372-893eef07f250\" (UID: \"6ee3b3b2-07e1-4bab-9372-893eef07f250\") " Oct 05 21:15:53 crc kubenswrapper[4754]: I1005 21:15:53.976193 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-94gmn\" (UniqueName: \"kubernetes.io/projected/6ee3b3b2-07e1-4bab-9372-893eef07f250-kube-api-access-94gmn\") pod \"6ee3b3b2-07e1-4bab-9372-893eef07f250\" (UID: \"6ee3b3b2-07e1-4bab-9372-893eef07f250\") " Oct 05 21:15:53 crc kubenswrapper[4754]: I1005 21:15:53.976227 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/6ee3b3b2-07e1-4bab-9372-893eef07f250-ceilometer-tls-certs\") pod \"6ee3b3b2-07e1-4bab-9372-893eef07f250\" (UID: \"6ee3b3b2-07e1-4bab-9372-893eef07f250\") " Oct 05 21:15:53 crc kubenswrapper[4754]: I1005 21:15:53.976251 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6ee3b3b2-07e1-4bab-9372-893eef07f250-sg-core-conf-yaml\") pod \"6ee3b3b2-07e1-4bab-9372-893eef07f250\" (UID: \"6ee3b3b2-07e1-4bab-9372-893eef07f250\") " Oct 05 21:15:53 crc kubenswrapper[4754]: I1005 21:15:53.976512 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6ee3b3b2-07e1-4bab-9372-893eef07f250-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "6ee3b3b2-07e1-4bab-9372-893eef07f250" (UID: "6ee3b3b2-07e1-4bab-9372-893eef07f250"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 21:15:53 crc kubenswrapper[4754]: I1005 21:15:53.977113 4754 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6ee3b3b2-07e1-4bab-9372-893eef07f250-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 05 21:15:53 crc kubenswrapper[4754]: I1005 21:15:53.977100 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6ee3b3b2-07e1-4bab-9372-893eef07f250-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "6ee3b3b2-07e1-4bab-9372-893eef07f250" (UID: "6ee3b3b2-07e1-4bab-9372-893eef07f250"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 21:15:53 crc kubenswrapper[4754]: I1005 21:15:53.983179 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ee3b3b2-07e1-4bab-9372-893eef07f250-scripts" (OuterVolumeSpecName: "scripts") pod "6ee3b3b2-07e1-4bab-9372-893eef07f250" (UID: "6ee3b3b2-07e1-4bab-9372-893eef07f250"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:15:53 crc kubenswrapper[4754]: I1005 21:15:53.983530 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ee3b3b2-07e1-4bab-9372-893eef07f250-kube-api-access-94gmn" (OuterVolumeSpecName: "kube-api-access-94gmn") pod "6ee3b3b2-07e1-4bab-9372-893eef07f250" (UID: "6ee3b3b2-07e1-4bab-9372-893eef07f250"). InnerVolumeSpecName "kube-api-access-94gmn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:15:54 crc kubenswrapper[4754]: I1005 21:15:54.031820 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ee3b3b2-07e1-4bab-9372-893eef07f250-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "6ee3b3b2-07e1-4bab-9372-893eef07f250" (UID: "6ee3b3b2-07e1-4bab-9372-893eef07f250"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:15:54 crc kubenswrapper[4754]: I1005 21:15:54.055167 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ee3b3b2-07e1-4bab-9372-893eef07f250-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "6ee3b3b2-07e1-4bab-9372-893eef07f250" (UID: "6ee3b3b2-07e1-4bab-9372-893eef07f250"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:15:54 crc kubenswrapper[4754]: I1005 21:15:54.081668 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-94gmn\" (UniqueName: \"kubernetes.io/projected/6ee3b3b2-07e1-4bab-9372-893eef07f250-kube-api-access-94gmn\") on node \"crc\" DevicePath \"\"" Oct 05 21:15:54 crc kubenswrapper[4754]: I1005 21:15:54.081704 4754 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/6ee3b3b2-07e1-4bab-9372-893eef07f250-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 05 21:15:54 crc kubenswrapper[4754]: I1005 21:15:54.081716 4754 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6ee3b3b2-07e1-4bab-9372-893eef07f250-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 05 21:15:54 crc kubenswrapper[4754]: I1005 21:15:54.081727 4754 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6ee3b3b2-07e1-4bab-9372-893eef07f250-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 05 21:15:54 crc kubenswrapper[4754]: I1005 21:15:54.081737 4754 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6ee3b3b2-07e1-4bab-9372-893eef07f250-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 21:15:54 crc kubenswrapper[4754]: I1005 21:15:54.124086 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ee3b3b2-07e1-4bab-9372-893eef07f250-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6ee3b3b2-07e1-4bab-9372-893eef07f250" (UID: "6ee3b3b2-07e1-4bab-9372-893eef07f250"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:15:54 crc kubenswrapper[4754]: I1005 21:15:54.129785 4754 generic.go:334] "Generic (PLEG): container finished" podID="6ee3b3b2-07e1-4bab-9372-893eef07f250" containerID="c1ce65ecb8ae3e72e860298633e5f6d77759ea8ef5efa531ec34ad976df4e63a" exitCode=0 Oct 05 21:15:54 crc kubenswrapper[4754]: I1005 21:15:54.129837 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6ee3b3b2-07e1-4bab-9372-893eef07f250","Type":"ContainerDied","Data":"c1ce65ecb8ae3e72e860298633e5f6d77759ea8ef5efa531ec34ad976df4e63a"} Oct 05 21:15:54 crc kubenswrapper[4754]: I1005 21:15:54.129867 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6ee3b3b2-07e1-4bab-9372-893eef07f250","Type":"ContainerDied","Data":"2066b48fffbebb577e4352eb48240b9ebfafe1926a62b2692b3f363961ebbf3d"} Oct 05 21:15:54 crc kubenswrapper[4754]: I1005 21:15:54.129886 4754 scope.go:117] "RemoveContainer" containerID="9ed026943be03e0a1a0532c59c6308a6e590ffc346f67cf516ee3fa829679dab" Oct 05 21:15:54 crc kubenswrapper[4754]: I1005 21:15:54.130063 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 05 21:15:54 crc kubenswrapper[4754]: I1005 21:15:54.154503 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ee3b3b2-07e1-4bab-9372-893eef07f250-config-data" (OuterVolumeSpecName: "config-data") pod "6ee3b3b2-07e1-4bab-9372-893eef07f250" (UID: "6ee3b3b2-07e1-4bab-9372-893eef07f250"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:15:54 crc kubenswrapper[4754]: I1005 21:15:54.173578 4754 scope.go:117] "RemoveContainer" containerID="58202d1eec91468517c6840252471678191af733282dd96c7961df59850a5bc7" Oct 05 21:15:54 crc kubenswrapper[4754]: I1005 21:15:54.183938 4754 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ee3b3b2-07e1-4bab-9372-893eef07f250-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 21:15:54 crc kubenswrapper[4754]: I1005 21:15:54.183978 4754 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6ee3b3b2-07e1-4bab-9372-893eef07f250-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 21:15:54 crc kubenswrapper[4754]: I1005 21:15:54.196144 4754 scope.go:117] "RemoveContainer" containerID="c1ce65ecb8ae3e72e860298633e5f6d77759ea8ef5efa531ec34ad976df4e63a" Oct 05 21:15:54 crc kubenswrapper[4754]: I1005 21:15:54.229403 4754 scope.go:117] "RemoveContainer" containerID="3380e04f976d4766b1f52c3232e01620a9cfc06705ef8c5c6004c0e41c7f6640" Oct 05 21:15:54 crc kubenswrapper[4754]: I1005 21:15:54.255599 4754 scope.go:117] "RemoveContainer" containerID="9ed026943be03e0a1a0532c59c6308a6e590ffc346f67cf516ee3fa829679dab" Oct 05 21:15:54 crc kubenswrapper[4754]: E1005 21:15:54.256094 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9ed026943be03e0a1a0532c59c6308a6e590ffc346f67cf516ee3fa829679dab\": container with ID starting with 9ed026943be03e0a1a0532c59c6308a6e590ffc346f67cf516ee3fa829679dab not found: ID does not exist" containerID="9ed026943be03e0a1a0532c59c6308a6e590ffc346f67cf516ee3fa829679dab" Oct 05 21:15:54 crc kubenswrapper[4754]: I1005 21:15:54.256149 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9ed026943be03e0a1a0532c59c6308a6e590ffc346f67cf516ee3fa829679dab"} err="failed to get container status \"9ed026943be03e0a1a0532c59c6308a6e590ffc346f67cf516ee3fa829679dab\": rpc error: code = NotFound desc = could not find container \"9ed026943be03e0a1a0532c59c6308a6e590ffc346f67cf516ee3fa829679dab\": container with ID starting with 9ed026943be03e0a1a0532c59c6308a6e590ffc346f67cf516ee3fa829679dab not found: ID does not exist" Oct 05 21:15:54 crc kubenswrapper[4754]: I1005 21:15:54.256185 4754 scope.go:117] "RemoveContainer" containerID="58202d1eec91468517c6840252471678191af733282dd96c7961df59850a5bc7" Oct 05 21:15:54 crc kubenswrapper[4754]: E1005 21:15:54.258171 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"58202d1eec91468517c6840252471678191af733282dd96c7961df59850a5bc7\": container with ID starting with 58202d1eec91468517c6840252471678191af733282dd96c7961df59850a5bc7 not found: ID does not exist" containerID="58202d1eec91468517c6840252471678191af733282dd96c7961df59850a5bc7" Oct 05 21:15:54 crc kubenswrapper[4754]: I1005 21:15:54.258213 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"58202d1eec91468517c6840252471678191af733282dd96c7961df59850a5bc7"} err="failed to get container status \"58202d1eec91468517c6840252471678191af733282dd96c7961df59850a5bc7\": rpc error: code = NotFound desc = could not find container \"58202d1eec91468517c6840252471678191af733282dd96c7961df59850a5bc7\": container with ID starting with 58202d1eec91468517c6840252471678191af733282dd96c7961df59850a5bc7 not found: ID does not exist" Oct 05 21:15:54 crc kubenswrapper[4754]: I1005 21:15:54.258240 4754 scope.go:117] "RemoveContainer" containerID="c1ce65ecb8ae3e72e860298633e5f6d77759ea8ef5efa531ec34ad976df4e63a" Oct 05 21:15:54 crc kubenswrapper[4754]: E1005 21:15:54.258805 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c1ce65ecb8ae3e72e860298633e5f6d77759ea8ef5efa531ec34ad976df4e63a\": container with ID starting with c1ce65ecb8ae3e72e860298633e5f6d77759ea8ef5efa531ec34ad976df4e63a not found: ID does not exist" containerID="c1ce65ecb8ae3e72e860298633e5f6d77759ea8ef5efa531ec34ad976df4e63a" Oct 05 21:15:54 crc kubenswrapper[4754]: I1005 21:15:54.258876 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c1ce65ecb8ae3e72e860298633e5f6d77759ea8ef5efa531ec34ad976df4e63a"} err="failed to get container status \"c1ce65ecb8ae3e72e860298633e5f6d77759ea8ef5efa531ec34ad976df4e63a\": rpc error: code = NotFound desc = could not find container \"c1ce65ecb8ae3e72e860298633e5f6d77759ea8ef5efa531ec34ad976df4e63a\": container with ID starting with c1ce65ecb8ae3e72e860298633e5f6d77759ea8ef5efa531ec34ad976df4e63a not found: ID does not exist" Oct 05 21:15:54 crc kubenswrapper[4754]: I1005 21:15:54.258916 4754 scope.go:117] "RemoveContainer" containerID="3380e04f976d4766b1f52c3232e01620a9cfc06705ef8c5c6004c0e41c7f6640" Oct 05 21:15:54 crc kubenswrapper[4754]: E1005 21:15:54.259393 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3380e04f976d4766b1f52c3232e01620a9cfc06705ef8c5c6004c0e41c7f6640\": container with ID starting with 3380e04f976d4766b1f52c3232e01620a9cfc06705ef8c5c6004c0e41c7f6640 not found: ID does not exist" containerID="3380e04f976d4766b1f52c3232e01620a9cfc06705ef8c5c6004c0e41c7f6640" Oct 05 21:15:54 crc kubenswrapper[4754]: I1005 21:15:54.259421 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3380e04f976d4766b1f52c3232e01620a9cfc06705ef8c5c6004c0e41c7f6640"} err="failed to get container status \"3380e04f976d4766b1f52c3232e01620a9cfc06705ef8c5c6004c0e41c7f6640\": rpc error: code = NotFound desc = could not find container \"3380e04f976d4766b1f52c3232e01620a9cfc06705ef8c5c6004c0e41c7f6640\": container with ID starting with 3380e04f976d4766b1f52c3232e01620a9cfc06705ef8c5c6004c0e41c7f6640 not found: ID does not exist" Oct 05 21:15:54 crc kubenswrapper[4754]: I1005 21:15:54.464966 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 05 21:15:54 crc kubenswrapper[4754]: I1005 21:15:54.476174 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 05 21:15:54 crc kubenswrapper[4754]: I1005 21:15:54.495028 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 05 21:15:54 crc kubenswrapper[4754]: E1005 21:15:54.495898 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ee3b3b2-07e1-4bab-9372-893eef07f250" containerName="sg-core" Oct 05 21:15:54 crc kubenswrapper[4754]: I1005 21:15:54.495973 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ee3b3b2-07e1-4bab-9372-893eef07f250" containerName="sg-core" Oct 05 21:15:54 crc kubenswrapper[4754]: E1005 21:15:54.496055 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ee3b3b2-07e1-4bab-9372-893eef07f250" containerName="ceilometer-central-agent" Oct 05 21:15:54 crc kubenswrapper[4754]: I1005 21:15:54.496106 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ee3b3b2-07e1-4bab-9372-893eef07f250" containerName="ceilometer-central-agent" Oct 05 21:15:54 crc kubenswrapper[4754]: E1005 21:15:54.496192 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ee3b3b2-07e1-4bab-9372-893eef07f250" containerName="proxy-httpd" Oct 05 21:15:54 crc kubenswrapper[4754]: I1005 21:15:54.496245 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ee3b3b2-07e1-4bab-9372-893eef07f250" containerName="proxy-httpd" Oct 05 21:15:54 crc kubenswrapper[4754]: E1005 21:15:54.496314 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ee3b3b2-07e1-4bab-9372-893eef07f250" containerName="ceilometer-notification-agent" Oct 05 21:15:54 crc kubenswrapper[4754]: I1005 21:15:54.496590 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ee3b3b2-07e1-4bab-9372-893eef07f250" containerName="ceilometer-notification-agent" Oct 05 21:15:54 crc kubenswrapper[4754]: I1005 21:15:54.496943 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ee3b3b2-07e1-4bab-9372-893eef07f250" containerName="sg-core" Oct 05 21:15:54 crc kubenswrapper[4754]: I1005 21:15:54.497025 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ee3b3b2-07e1-4bab-9372-893eef07f250" containerName="ceilometer-central-agent" Oct 05 21:15:54 crc kubenswrapper[4754]: I1005 21:15:54.497090 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ee3b3b2-07e1-4bab-9372-893eef07f250" containerName="ceilometer-notification-agent" Oct 05 21:15:54 crc kubenswrapper[4754]: I1005 21:15:54.497167 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ee3b3b2-07e1-4bab-9372-893eef07f250" containerName="proxy-httpd" Oct 05 21:15:54 crc kubenswrapper[4754]: I1005 21:15:54.500249 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 05 21:15:54 crc kubenswrapper[4754]: I1005 21:15:54.502890 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 05 21:15:54 crc kubenswrapper[4754]: I1005 21:15:54.502941 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 05 21:15:54 crc kubenswrapper[4754]: I1005 21:15:54.506142 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 05 21:15:54 crc kubenswrapper[4754]: I1005 21:15:54.545057 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 05 21:15:54 crc kubenswrapper[4754]: I1005 21:15:54.592390 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b5010900-7b57-48e9-9621-73940c2aefa1-log-httpd\") pod \"ceilometer-0\" (UID: \"b5010900-7b57-48e9-9621-73940c2aefa1\") " pod="openstack/ceilometer-0" Oct 05 21:15:54 crc kubenswrapper[4754]: I1005 21:15:54.592444 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b5010900-7b57-48e9-9621-73940c2aefa1-config-data\") pod \"ceilometer-0\" (UID: \"b5010900-7b57-48e9-9621-73940c2aefa1\") " pod="openstack/ceilometer-0" Oct 05 21:15:54 crc kubenswrapper[4754]: I1005 21:15:54.592467 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b5010900-7b57-48e9-9621-73940c2aefa1-scripts\") pod \"ceilometer-0\" (UID: \"b5010900-7b57-48e9-9621-73940c2aefa1\") " pod="openstack/ceilometer-0" Oct 05 21:15:54 crc kubenswrapper[4754]: I1005 21:15:54.592573 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vspks\" (UniqueName: \"kubernetes.io/projected/b5010900-7b57-48e9-9621-73940c2aefa1-kube-api-access-vspks\") pod \"ceilometer-0\" (UID: \"b5010900-7b57-48e9-9621-73940c2aefa1\") " pod="openstack/ceilometer-0" Oct 05 21:15:54 crc kubenswrapper[4754]: I1005 21:15:54.592627 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b5010900-7b57-48e9-9621-73940c2aefa1-run-httpd\") pod \"ceilometer-0\" (UID: \"b5010900-7b57-48e9-9621-73940c2aefa1\") " pod="openstack/ceilometer-0" Oct 05 21:15:54 crc kubenswrapper[4754]: I1005 21:15:54.592679 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b5010900-7b57-48e9-9621-73940c2aefa1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b5010900-7b57-48e9-9621-73940c2aefa1\") " pod="openstack/ceilometer-0" Oct 05 21:15:54 crc kubenswrapper[4754]: I1005 21:15:54.592712 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5010900-7b57-48e9-9621-73940c2aefa1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b5010900-7b57-48e9-9621-73940c2aefa1\") " pod="openstack/ceilometer-0" Oct 05 21:15:54 crc kubenswrapper[4754]: I1005 21:15:54.592756 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/b5010900-7b57-48e9-9621-73940c2aefa1-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"b5010900-7b57-48e9-9621-73940c2aefa1\") " pod="openstack/ceilometer-0" Oct 05 21:15:54 crc kubenswrapper[4754]: I1005 21:15:54.696458 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vspks\" (UniqueName: \"kubernetes.io/projected/b5010900-7b57-48e9-9621-73940c2aefa1-kube-api-access-vspks\") pod \"ceilometer-0\" (UID: \"b5010900-7b57-48e9-9621-73940c2aefa1\") " pod="openstack/ceilometer-0" Oct 05 21:15:54 crc kubenswrapper[4754]: I1005 21:15:54.696666 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b5010900-7b57-48e9-9621-73940c2aefa1-run-httpd\") pod \"ceilometer-0\" (UID: \"b5010900-7b57-48e9-9621-73940c2aefa1\") " pod="openstack/ceilometer-0" Oct 05 21:15:54 crc kubenswrapper[4754]: I1005 21:15:54.696798 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b5010900-7b57-48e9-9621-73940c2aefa1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b5010900-7b57-48e9-9621-73940c2aefa1\") " pod="openstack/ceilometer-0" Oct 05 21:15:54 crc kubenswrapper[4754]: I1005 21:15:54.696865 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5010900-7b57-48e9-9621-73940c2aefa1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b5010900-7b57-48e9-9621-73940c2aefa1\") " pod="openstack/ceilometer-0" Oct 05 21:15:54 crc kubenswrapper[4754]: I1005 21:15:54.696954 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/b5010900-7b57-48e9-9621-73940c2aefa1-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"b5010900-7b57-48e9-9621-73940c2aefa1\") " pod="openstack/ceilometer-0" Oct 05 21:15:54 crc kubenswrapper[4754]: I1005 21:15:54.697037 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b5010900-7b57-48e9-9621-73940c2aefa1-log-httpd\") pod \"ceilometer-0\" (UID: \"b5010900-7b57-48e9-9621-73940c2aefa1\") " pod="openstack/ceilometer-0" Oct 05 21:15:54 crc kubenswrapper[4754]: I1005 21:15:54.697111 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b5010900-7b57-48e9-9621-73940c2aefa1-config-data\") pod \"ceilometer-0\" (UID: \"b5010900-7b57-48e9-9621-73940c2aefa1\") " pod="openstack/ceilometer-0" Oct 05 21:15:54 crc kubenswrapper[4754]: I1005 21:15:54.697148 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b5010900-7b57-48e9-9621-73940c2aefa1-scripts\") pod \"ceilometer-0\" (UID: \"b5010900-7b57-48e9-9621-73940c2aefa1\") " pod="openstack/ceilometer-0" Oct 05 21:15:54 crc kubenswrapper[4754]: I1005 21:15:54.697482 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b5010900-7b57-48e9-9621-73940c2aefa1-log-httpd\") pod \"ceilometer-0\" (UID: \"b5010900-7b57-48e9-9621-73940c2aefa1\") " pod="openstack/ceilometer-0" Oct 05 21:15:54 crc kubenswrapper[4754]: I1005 21:15:54.697729 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b5010900-7b57-48e9-9621-73940c2aefa1-run-httpd\") pod \"ceilometer-0\" (UID: \"b5010900-7b57-48e9-9621-73940c2aefa1\") " pod="openstack/ceilometer-0" Oct 05 21:15:54 crc kubenswrapper[4754]: I1005 21:15:54.701006 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/b5010900-7b57-48e9-9621-73940c2aefa1-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"b5010900-7b57-48e9-9621-73940c2aefa1\") " pod="openstack/ceilometer-0" Oct 05 21:15:54 crc kubenswrapper[4754]: I1005 21:15:54.703717 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b5010900-7b57-48e9-9621-73940c2aefa1-config-data\") pod \"ceilometer-0\" (UID: \"b5010900-7b57-48e9-9621-73940c2aefa1\") " pod="openstack/ceilometer-0" Oct 05 21:15:54 crc kubenswrapper[4754]: I1005 21:15:54.704589 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b5010900-7b57-48e9-9621-73940c2aefa1-scripts\") pod \"ceilometer-0\" (UID: \"b5010900-7b57-48e9-9621-73940c2aefa1\") " pod="openstack/ceilometer-0" Oct 05 21:15:54 crc kubenswrapper[4754]: I1005 21:15:54.705106 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5010900-7b57-48e9-9621-73940c2aefa1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b5010900-7b57-48e9-9621-73940c2aefa1\") " pod="openstack/ceilometer-0" Oct 05 21:15:54 crc kubenswrapper[4754]: I1005 21:15:54.706127 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b5010900-7b57-48e9-9621-73940c2aefa1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b5010900-7b57-48e9-9621-73940c2aefa1\") " pod="openstack/ceilometer-0" Oct 05 21:15:54 crc kubenswrapper[4754]: I1005 21:15:54.732203 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vspks\" (UniqueName: \"kubernetes.io/projected/b5010900-7b57-48e9-9621-73940c2aefa1-kube-api-access-vspks\") pod \"ceilometer-0\" (UID: \"b5010900-7b57-48e9-9621-73940c2aefa1\") " pod="openstack/ceilometer-0" Oct 05 21:15:54 crc kubenswrapper[4754]: I1005 21:15:54.815624 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 05 21:15:54 crc kubenswrapper[4754]: I1005 21:15:54.854584 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ee3b3b2-07e1-4bab-9372-893eef07f250" path="/var/lib/kubelet/pods/6ee3b3b2-07e1-4bab-9372-893eef07f250/volumes" Oct 05 21:15:55 crc kubenswrapper[4754]: I1005 21:15:55.356521 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 05 21:15:55 crc kubenswrapper[4754]: I1005 21:15:55.604686 4754 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-68c6b67864-9msm5" podUID="27cd6be5-f5eb-4158-bdd1-37f4df3d530d" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.148:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.148:8443: connect: connection refused" Oct 05 21:15:55 crc kubenswrapper[4754]: I1005 21:15:55.682525 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Oct 05 21:15:55 crc kubenswrapper[4754]: I1005 21:15:55.706342 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Oct 05 21:15:56 crc kubenswrapper[4754]: I1005 21:15:56.114541 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 05 21:15:56 crc kubenswrapper[4754]: I1005 21:15:56.167033 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b5010900-7b57-48e9-9621-73940c2aefa1","Type":"ContainerStarted","Data":"dc4131652f2fb7f1d0921cfd802388d4687129320ddcbc0464e2324cf89b4aad"} Oct 05 21:15:56 crc kubenswrapper[4754]: I1005 21:15:56.173936 4754 generic.go:334] "Generic (PLEG): container finished" podID="74c27a73-083c-4713-9345-4dc0ebf9a00e" containerID="414aa4ae6d587a775d944f425af1a445b4ad194a78e1c8fcf619395ec0531910" exitCode=0 Oct 05 21:15:56 crc kubenswrapper[4754]: I1005 21:15:56.174030 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"74c27a73-083c-4713-9345-4dc0ebf9a00e","Type":"ContainerDied","Data":"414aa4ae6d587a775d944f425af1a445b4ad194a78e1c8fcf619395ec0531910"} Oct 05 21:15:56 crc kubenswrapper[4754]: I1005 21:15:56.174041 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 05 21:15:56 crc kubenswrapper[4754]: I1005 21:15:56.174087 4754 scope.go:117] "RemoveContainer" containerID="414aa4ae6d587a775d944f425af1a445b4ad194a78e1c8fcf619395ec0531910" Oct 05 21:15:56 crc kubenswrapper[4754]: I1005 21:15:56.174068 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"74c27a73-083c-4713-9345-4dc0ebf9a00e","Type":"ContainerDied","Data":"8cad5f9883aab31b721c9ac8c8226809cfd817ac8b46963d70883f1d9f472f47"} Oct 05 21:15:56 crc kubenswrapper[4754]: I1005 21:15:56.204875 4754 scope.go:117] "RemoveContainer" containerID="535d1c814f0ab307a5a2a202d80fde5006ea9450607082f6ec36dccab9eec72f" Oct 05 21:15:56 crc kubenswrapper[4754]: I1005 21:15:56.206636 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Oct 05 21:15:56 crc kubenswrapper[4754]: I1005 21:15:56.233636 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9fz7t\" (UniqueName: \"kubernetes.io/projected/74c27a73-083c-4713-9345-4dc0ebf9a00e-kube-api-access-9fz7t\") pod \"74c27a73-083c-4713-9345-4dc0ebf9a00e\" (UID: \"74c27a73-083c-4713-9345-4dc0ebf9a00e\") " Oct 05 21:15:56 crc kubenswrapper[4754]: I1005 21:15:56.233867 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74c27a73-083c-4713-9345-4dc0ebf9a00e-combined-ca-bundle\") pod \"74c27a73-083c-4713-9345-4dc0ebf9a00e\" (UID: \"74c27a73-083c-4713-9345-4dc0ebf9a00e\") " Oct 05 21:15:56 crc kubenswrapper[4754]: I1005 21:15:56.233926 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/74c27a73-083c-4713-9345-4dc0ebf9a00e-logs\") pod \"74c27a73-083c-4713-9345-4dc0ebf9a00e\" (UID: \"74c27a73-083c-4713-9345-4dc0ebf9a00e\") " Oct 05 21:15:56 crc kubenswrapper[4754]: I1005 21:15:56.234013 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74c27a73-083c-4713-9345-4dc0ebf9a00e-config-data\") pod \"74c27a73-083c-4713-9345-4dc0ebf9a00e\" (UID: \"74c27a73-083c-4713-9345-4dc0ebf9a00e\") " Oct 05 21:15:56 crc kubenswrapper[4754]: I1005 21:15:56.239298 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/74c27a73-083c-4713-9345-4dc0ebf9a00e-logs" (OuterVolumeSpecName: "logs") pod "74c27a73-083c-4713-9345-4dc0ebf9a00e" (UID: "74c27a73-083c-4713-9345-4dc0ebf9a00e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 21:15:56 crc kubenswrapper[4754]: I1005 21:15:56.248975 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/74c27a73-083c-4713-9345-4dc0ebf9a00e-kube-api-access-9fz7t" (OuterVolumeSpecName: "kube-api-access-9fz7t") pod "74c27a73-083c-4713-9345-4dc0ebf9a00e" (UID: "74c27a73-083c-4713-9345-4dc0ebf9a00e"). InnerVolumeSpecName "kube-api-access-9fz7t". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:15:56 crc kubenswrapper[4754]: I1005 21:15:56.261317 4754 scope.go:117] "RemoveContainer" containerID="414aa4ae6d587a775d944f425af1a445b4ad194a78e1c8fcf619395ec0531910" Oct 05 21:15:56 crc kubenswrapper[4754]: E1005 21:15:56.264716 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"414aa4ae6d587a775d944f425af1a445b4ad194a78e1c8fcf619395ec0531910\": container with ID starting with 414aa4ae6d587a775d944f425af1a445b4ad194a78e1c8fcf619395ec0531910 not found: ID does not exist" containerID="414aa4ae6d587a775d944f425af1a445b4ad194a78e1c8fcf619395ec0531910" Oct 05 21:15:56 crc kubenswrapper[4754]: I1005 21:15:56.264771 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"414aa4ae6d587a775d944f425af1a445b4ad194a78e1c8fcf619395ec0531910"} err="failed to get container status \"414aa4ae6d587a775d944f425af1a445b4ad194a78e1c8fcf619395ec0531910\": rpc error: code = NotFound desc = could not find container \"414aa4ae6d587a775d944f425af1a445b4ad194a78e1c8fcf619395ec0531910\": container with ID starting with 414aa4ae6d587a775d944f425af1a445b4ad194a78e1c8fcf619395ec0531910 not found: ID does not exist" Oct 05 21:15:56 crc kubenswrapper[4754]: I1005 21:15:56.264801 4754 scope.go:117] "RemoveContainer" containerID="535d1c814f0ab307a5a2a202d80fde5006ea9450607082f6ec36dccab9eec72f" Oct 05 21:15:56 crc kubenswrapper[4754]: E1005 21:15:56.266255 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"535d1c814f0ab307a5a2a202d80fde5006ea9450607082f6ec36dccab9eec72f\": container with ID starting with 535d1c814f0ab307a5a2a202d80fde5006ea9450607082f6ec36dccab9eec72f not found: ID does not exist" containerID="535d1c814f0ab307a5a2a202d80fde5006ea9450607082f6ec36dccab9eec72f" Oct 05 21:15:56 crc kubenswrapper[4754]: I1005 21:15:56.266274 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"535d1c814f0ab307a5a2a202d80fde5006ea9450607082f6ec36dccab9eec72f"} err="failed to get container status \"535d1c814f0ab307a5a2a202d80fde5006ea9450607082f6ec36dccab9eec72f\": rpc error: code = NotFound desc = could not find container \"535d1c814f0ab307a5a2a202d80fde5006ea9450607082f6ec36dccab9eec72f\": container with ID starting with 535d1c814f0ab307a5a2a202d80fde5006ea9450607082f6ec36dccab9eec72f not found: ID does not exist" Oct 05 21:15:56 crc kubenswrapper[4754]: I1005 21:15:56.285904 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74c27a73-083c-4713-9345-4dc0ebf9a00e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "74c27a73-083c-4713-9345-4dc0ebf9a00e" (UID: "74c27a73-083c-4713-9345-4dc0ebf9a00e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:15:56 crc kubenswrapper[4754]: I1005 21:15:56.298732 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74c27a73-083c-4713-9345-4dc0ebf9a00e-config-data" (OuterVolumeSpecName: "config-data") pod "74c27a73-083c-4713-9345-4dc0ebf9a00e" (UID: "74c27a73-083c-4713-9345-4dc0ebf9a00e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:15:56 crc kubenswrapper[4754]: I1005 21:15:56.336767 4754 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74c27a73-083c-4713-9345-4dc0ebf9a00e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 21:15:56 crc kubenswrapper[4754]: I1005 21:15:56.336796 4754 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/74c27a73-083c-4713-9345-4dc0ebf9a00e-logs\") on node \"crc\" DevicePath \"\"" Oct 05 21:15:56 crc kubenswrapper[4754]: I1005 21:15:56.336805 4754 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74c27a73-083c-4713-9345-4dc0ebf9a00e-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 21:15:56 crc kubenswrapper[4754]: I1005 21:15:56.336829 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9fz7t\" (UniqueName: \"kubernetes.io/projected/74c27a73-083c-4713-9345-4dc0ebf9a00e-kube-api-access-9fz7t\") on node \"crc\" DevicePath \"\"" Oct 05 21:15:56 crc kubenswrapper[4754]: I1005 21:15:56.449791 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-stkc2"] Oct 05 21:15:56 crc kubenswrapper[4754]: E1005 21:15:56.450676 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74c27a73-083c-4713-9345-4dc0ebf9a00e" containerName="nova-api-api" Oct 05 21:15:56 crc kubenswrapper[4754]: I1005 21:15:56.450699 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="74c27a73-083c-4713-9345-4dc0ebf9a00e" containerName="nova-api-api" Oct 05 21:15:56 crc kubenswrapper[4754]: E1005 21:15:56.450759 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74c27a73-083c-4713-9345-4dc0ebf9a00e" containerName="nova-api-log" Oct 05 21:15:56 crc kubenswrapper[4754]: I1005 21:15:56.450769 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="74c27a73-083c-4713-9345-4dc0ebf9a00e" containerName="nova-api-log" Oct 05 21:15:56 crc kubenswrapper[4754]: I1005 21:15:56.451173 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="74c27a73-083c-4713-9345-4dc0ebf9a00e" containerName="nova-api-log" Oct 05 21:15:56 crc kubenswrapper[4754]: I1005 21:15:56.451209 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="74c27a73-083c-4713-9345-4dc0ebf9a00e" containerName="nova-api-api" Oct 05 21:15:56 crc kubenswrapper[4754]: I1005 21:15:56.452234 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-stkc2" Oct 05 21:15:56 crc kubenswrapper[4754]: I1005 21:15:56.489873 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Oct 05 21:15:56 crc kubenswrapper[4754]: I1005 21:15:56.502152 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-stkc2"] Oct 05 21:15:56 crc kubenswrapper[4754]: I1005 21:15:56.502809 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Oct 05 21:15:56 crc kubenswrapper[4754]: I1005 21:15:56.582567 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zqxxx\" (UniqueName: \"kubernetes.io/projected/ba3c8f4e-033c-48d4-b37e-7b2542be3bd1-kube-api-access-zqxxx\") pod \"nova-cell1-cell-mapping-stkc2\" (UID: \"ba3c8f4e-033c-48d4-b37e-7b2542be3bd1\") " pod="openstack/nova-cell1-cell-mapping-stkc2" Oct 05 21:15:56 crc kubenswrapper[4754]: I1005 21:15:56.582719 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba3c8f4e-033c-48d4-b37e-7b2542be3bd1-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-stkc2\" (UID: \"ba3c8f4e-033c-48d4-b37e-7b2542be3bd1\") " pod="openstack/nova-cell1-cell-mapping-stkc2" Oct 05 21:15:56 crc kubenswrapper[4754]: I1005 21:15:56.582997 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ba3c8f4e-033c-48d4-b37e-7b2542be3bd1-config-data\") pod \"nova-cell1-cell-mapping-stkc2\" (UID: \"ba3c8f4e-033c-48d4-b37e-7b2542be3bd1\") " pod="openstack/nova-cell1-cell-mapping-stkc2" Oct 05 21:15:56 crc kubenswrapper[4754]: I1005 21:15:56.583367 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ba3c8f4e-033c-48d4-b37e-7b2542be3bd1-scripts\") pod \"nova-cell1-cell-mapping-stkc2\" (UID: \"ba3c8f4e-033c-48d4-b37e-7b2542be3bd1\") " pod="openstack/nova-cell1-cell-mapping-stkc2" Oct 05 21:15:56 crc kubenswrapper[4754]: I1005 21:15:56.602689 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 05 21:15:56 crc kubenswrapper[4754]: I1005 21:15:56.626983 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 05 21:15:56 crc kubenswrapper[4754]: I1005 21:15:56.635639 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 05 21:15:56 crc kubenswrapper[4754]: I1005 21:15:56.637740 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 05 21:15:56 crc kubenswrapper[4754]: I1005 21:15:56.640474 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 05 21:15:56 crc kubenswrapper[4754]: I1005 21:15:56.643835 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Oct 05 21:15:56 crc kubenswrapper[4754]: I1005 21:15:56.645211 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 05 21:15:56 crc kubenswrapper[4754]: I1005 21:15:56.645315 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Oct 05 21:15:56 crc kubenswrapper[4754]: I1005 21:15:56.686037 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ba3c8f4e-033c-48d4-b37e-7b2542be3bd1-config-data\") pod \"nova-cell1-cell-mapping-stkc2\" (UID: \"ba3c8f4e-033c-48d4-b37e-7b2542be3bd1\") " pod="openstack/nova-cell1-cell-mapping-stkc2" Oct 05 21:15:56 crc kubenswrapper[4754]: I1005 21:15:56.711802 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ba3c8f4e-033c-48d4-b37e-7b2542be3bd1-config-data\") pod \"nova-cell1-cell-mapping-stkc2\" (UID: \"ba3c8f4e-033c-48d4-b37e-7b2542be3bd1\") " pod="openstack/nova-cell1-cell-mapping-stkc2" Oct 05 21:15:56 crc kubenswrapper[4754]: I1005 21:15:56.718781 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ba3c8f4e-033c-48d4-b37e-7b2542be3bd1-scripts\") pod \"nova-cell1-cell-mapping-stkc2\" (UID: \"ba3c8f4e-033c-48d4-b37e-7b2542be3bd1\") " pod="openstack/nova-cell1-cell-mapping-stkc2" Oct 05 21:15:56 crc kubenswrapper[4754]: I1005 21:15:56.718933 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zqxxx\" (UniqueName: \"kubernetes.io/projected/ba3c8f4e-033c-48d4-b37e-7b2542be3bd1-kube-api-access-zqxxx\") pod \"nova-cell1-cell-mapping-stkc2\" (UID: \"ba3c8f4e-033c-48d4-b37e-7b2542be3bd1\") " pod="openstack/nova-cell1-cell-mapping-stkc2" Oct 05 21:15:56 crc kubenswrapper[4754]: I1005 21:15:56.718988 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba3c8f4e-033c-48d4-b37e-7b2542be3bd1-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-stkc2\" (UID: \"ba3c8f4e-033c-48d4-b37e-7b2542be3bd1\") " pod="openstack/nova-cell1-cell-mapping-stkc2" Oct 05 21:15:56 crc kubenswrapper[4754]: I1005 21:15:56.729806 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba3c8f4e-033c-48d4-b37e-7b2542be3bd1-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-stkc2\" (UID: \"ba3c8f4e-033c-48d4-b37e-7b2542be3bd1\") " pod="openstack/nova-cell1-cell-mapping-stkc2" Oct 05 21:15:56 crc kubenswrapper[4754]: I1005 21:15:56.736955 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ba3c8f4e-033c-48d4-b37e-7b2542be3bd1-scripts\") pod \"nova-cell1-cell-mapping-stkc2\" (UID: \"ba3c8f4e-033c-48d4-b37e-7b2542be3bd1\") " pod="openstack/nova-cell1-cell-mapping-stkc2" Oct 05 21:15:56 crc kubenswrapper[4754]: I1005 21:15:56.762047 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zqxxx\" (UniqueName: \"kubernetes.io/projected/ba3c8f4e-033c-48d4-b37e-7b2542be3bd1-kube-api-access-zqxxx\") pod \"nova-cell1-cell-mapping-stkc2\" (UID: \"ba3c8f4e-033c-48d4-b37e-7b2542be3bd1\") " pod="openstack/nova-cell1-cell-mapping-stkc2" Oct 05 21:15:56 crc kubenswrapper[4754]: I1005 21:15:56.821352 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95b27888-bd9d-4ebf-a0d8-9f6653d3fec5-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"95b27888-bd9d-4ebf-a0d8-9f6653d3fec5\") " pod="openstack/nova-api-0" Oct 05 21:15:56 crc kubenswrapper[4754]: I1005 21:15:56.821465 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95b27888-bd9d-4ebf-a0d8-9f6653d3fec5-config-data\") pod \"nova-api-0\" (UID: \"95b27888-bd9d-4ebf-a0d8-9f6653d3fec5\") " pod="openstack/nova-api-0" Oct 05 21:15:56 crc kubenswrapper[4754]: I1005 21:15:56.821508 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sz5lh\" (UniqueName: \"kubernetes.io/projected/95b27888-bd9d-4ebf-a0d8-9f6653d3fec5-kube-api-access-sz5lh\") pod \"nova-api-0\" (UID: \"95b27888-bd9d-4ebf-a0d8-9f6653d3fec5\") " pod="openstack/nova-api-0" Oct 05 21:15:56 crc kubenswrapper[4754]: I1005 21:15:56.821528 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/95b27888-bd9d-4ebf-a0d8-9f6653d3fec5-logs\") pod \"nova-api-0\" (UID: \"95b27888-bd9d-4ebf-a0d8-9f6653d3fec5\") " pod="openstack/nova-api-0" Oct 05 21:15:56 crc kubenswrapper[4754]: I1005 21:15:56.821561 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/95b27888-bd9d-4ebf-a0d8-9f6653d3fec5-public-tls-certs\") pod \"nova-api-0\" (UID: \"95b27888-bd9d-4ebf-a0d8-9f6653d3fec5\") " pod="openstack/nova-api-0" Oct 05 21:15:56 crc kubenswrapper[4754]: I1005 21:15:56.821599 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/95b27888-bd9d-4ebf-a0d8-9f6653d3fec5-internal-tls-certs\") pod \"nova-api-0\" (UID: \"95b27888-bd9d-4ebf-a0d8-9f6653d3fec5\") " pod="openstack/nova-api-0" Oct 05 21:15:56 crc kubenswrapper[4754]: I1005 21:15:56.860261 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-stkc2" Oct 05 21:15:56 crc kubenswrapper[4754]: I1005 21:15:56.879110 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="74c27a73-083c-4713-9345-4dc0ebf9a00e" path="/var/lib/kubelet/pods/74c27a73-083c-4713-9345-4dc0ebf9a00e/volumes" Oct 05 21:15:56 crc kubenswrapper[4754]: I1005 21:15:56.924920 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95b27888-bd9d-4ebf-a0d8-9f6653d3fec5-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"95b27888-bd9d-4ebf-a0d8-9f6653d3fec5\") " pod="openstack/nova-api-0" Oct 05 21:15:56 crc kubenswrapper[4754]: I1005 21:15:56.925037 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95b27888-bd9d-4ebf-a0d8-9f6653d3fec5-config-data\") pod \"nova-api-0\" (UID: \"95b27888-bd9d-4ebf-a0d8-9f6653d3fec5\") " pod="openstack/nova-api-0" Oct 05 21:15:56 crc kubenswrapper[4754]: I1005 21:15:56.925070 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sz5lh\" (UniqueName: \"kubernetes.io/projected/95b27888-bd9d-4ebf-a0d8-9f6653d3fec5-kube-api-access-sz5lh\") pod \"nova-api-0\" (UID: \"95b27888-bd9d-4ebf-a0d8-9f6653d3fec5\") " pod="openstack/nova-api-0" Oct 05 21:15:56 crc kubenswrapper[4754]: I1005 21:15:56.925095 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/95b27888-bd9d-4ebf-a0d8-9f6653d3fec5-logs\") pod \"nova-api-0\" (UID: \"95b27888-bd9d-4ebf-a0d8-9f6653d3fec5\") " pod="openstack/nova-api-0" Oct 05 21:15:56 crc kubenswrapper[4754]: I1005 21:15:56.925126 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/95b27888-bd9d-4ebf-a0d8-9f6653d3fec5-public-tls-certs\") pod \"nova-api-0\" (UID: \"95b27888-bd9d-4ebf-a0d8-9f6653d3fec5\") " pod="openstack/nova-api-0" Oct 05 21:15:56 crc kubenswrapper[4754]: I1005 21:15:56.925165 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/95b27888-bd9d-4ebf-a0d8-9f6653d3fec5-internal-tls-certs\") pod \"nova-api-0\" (UID: \"95b27888-bd9d-4ebf-a0d8-9f6653d3fec5\") " pod="openstack/nova-api-0" Oct 05 21:15:56 crc kubenswrapper[4754]: I1005 21:15:56.931209 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/95b27888-bd9d-4ebf-a0d8-9f6653d3fec5-logs\") pod \"nova-api-0\" (UID: \"95b27888-bd9d-4ebf-a0d8-9f6653d3fec5\") " pod="openstack/nova-api-0" Oct 05 21:15:56 crc kubenswrapper[4754]: I1005 21:15:56.939365 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95b27888-bd9d-4ebf-a0d8-9f6653d3fec5-config-data\") pod \"nova-api-0\" (UID: \"95b27888-bd9d-4ebf-a0d8-9f6653d3fec5\") " pod="openstack/nova-api-0" Oct 05 21:15:56 crc kubenswrapper[4754]: I1005 21:15:56.940511 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/95b27888-bd9d-4ebf-a0d8-9f6653d3fec5-internal-tls-certs\") pod \"nova-api-0\" (UID: \"95b27888-bd9d-4ebf-a0d8-9f6653d3fec5\") " pod="openstack/nova-api-0" Oct 05 21:15:56 crc kubenswrapper[4754]: I1005 21:15:56.948597 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95b27888-bd9d-4ebf-a0d8-9f6653d3fec5-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"95b27888-bd9d-4ebf-a0d8-9f6653d3fec5\") " pod="openstack/nova-api-0" Oct 05 21:15:56 crc kubenswrapper[4754]: I1005 21:15:56.951460 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/95b27888-bd9d-4ebf-a0d8-9f6653d3fec5-public-tls-certs\") pod \"nova-api-0\" (UID: \"95b27888-bd9d-4ebf-a0d8-9f6653d3fec5\") " pod="openstack/nova-api-0" Oct 05 21:15:56 crc kubenswrapper[4754]: I1005 21:15:56.955334 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sz5lh\" (UniqueName: \"kubernetes.io/projected/95b27888-bd9d-4ebf-a0d8-9f6653d3fec5-kube-api-access-sz5lh\") pod \"nova-api-0\" (UID: \"95b27888-bd9d-4ebf-a0d8-9f6653d3fec5\") " pod="openstack/nova-api-0" Oct 05 21:15:56 crc kubenswrapper[4754]: I1005 21:15:56.976394 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 05 21:15:57 crc kubenswrapper[4754]: I1005 21:15:57.206173 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b5010900-7b57-48e9-9621-73940c2aefa1","Type":"ContainerStarted","Data":"543f0a968ebefe22d4821c1200bfd8b4bd8c32fbbe0d0ab4c821e665da3851c8"} Oct 05 21:15:57 crc kubenswrapper[4754]: W1005 21:15:57.438184 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podba3c8f4e_033c_48d4_b37e_7b2542be3bd1.slice/crio-a22929e023b20c2dda1c8b0e060862a3270349d306f68279cf3c92fa0b2c0474 WatchSource:0}: Error finding container a22929e023b20c2dda1c8b0e060862a3270349d306f68279cf3c92fa0b2c0474: Status 404 returned error can't find the container with id a22929e023b20c2dda1c8b0e060862a3270349d306f68279cf3c92fa0b2c0474 Oct 05 21:15:57 crc kubenswrapper[4754]: I1005 21:15:57.438697 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-stkc2"] Oct 05 21:15:57 crc kubenswrapper[4754]: I1005 21:15:57.578300 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 05 21:15:58 crc kubenswrapper[4754]: I1005 21:15:58.243856 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"95b27888-bd9d-4ebf-a0d8-9f6653d3fec5","Type":"ContainerStarted","Data":"4ac7d7f41e4b7a9a3f01256e0411d11ff35d07006942978c7ee3635980fc5fc8"} Oct 05 21:15:58 crc kubenswrapper[4754]: I1005 21:15:58.244416 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"95b27888-bd9d-4ebf-a0d8-9f6653d3fec5","Type":"ContainerStarted","Data":"a1fdfabbf98552e60eeeae588f22f843362cec9781110d97076049277412098a"} Oct 05 21:15:58 crc kubenswrapper[4754]: I1005 21:15:58.244435 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"95b27888-bd9d-4ebf-a0d8-9f6653d3fec5","Type":"ContainerStarted","Data":"bd1014b234168d04dc783a70f78735175b155178927ffb3da0b353a36d3aa1a8"} Oct 05 21:15:58 crc kubenswrapper[4754]: I1005 21:15:58.254129 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b5010900-7b57-48e9-9621-73940c2aefa1","Type":"ContainerStarted","Data":"7d029e5245274343fd0b8f297dd50dc6df3da1fb621c1298550087f3043de42f"} Oct 05 21:15:58 crc kubenswrapper[4754]: I1005 21:15:58.254202 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b5010900-7b57-48e9-9621-73940c2aefa1","Type":"ContainerStarted","Data":"bd3a8ec19d4154ba447c0f140103c7ce97ad799503b0ef320c4e290d5716403b"} Oct 05 21:15:58 crc kubenswrapper[4754]: I1005 21:15:58.257093 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-stkc2" event={"ID":"ba3c8f4e-033c-48d4-b37e-7b2542be3bd1","Type":"ContainerStarted","Data":"aff0a26aedb047df62a0a5b78a35b03c9c74f10020d36a1af66b2a6b57901861"} Oct 05 21:15:58 crc kubenswrapper[4754]: I1005 21:15:58.257146 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-stkc2" event={"ID":"ba3c8f4e-033c-48d4-b37e-7b2542be3bd1","Type":"ContainerStarted","Data":"a22929e023b20c2dda1c8b0e060862a3270349d306f68279cf3c92fa0b2c0474"} Oct 05 21:15:58 crc kubenswrapper[4754]: I1005 21:15:58.269388 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.269369758 podStartE2EDuration="2.269369758s" podCreationTimestamp="2025-10-05 21:15:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 21:15:58.264813056 +0000 UTC m=+1282.168931766" watchObservedRunningTime="2025-10-05 21:15:58.269369758 +0000 UTC m=+1282.173488468" Oct 05 21:15:58 crc kubenswrapper[4754]: I1005 21:15:58.283749 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-stkc2" podStartSLOduration=2.283728814 podStartE2EDuration="2.283728814s" podCreationTimestamp="2025-10-05 21:15:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 21:15:58.281847383 +0000 UTC m=+1282.185966093" watchObservedRunningTime="2025-10-05 21:15:58.283728814 +0000 UTC m=+1282.187847524" Oct 05 21:15:59 crc kubenswrapper[4754]: I1005 21:15:59.638772 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-59cf4bdb65-jlwmm" Oct 05 21:15:59 crc kubenswrapper[4754]: I1005 21:15:59.748126 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-886x9"] Oct 05 21:15:59 crc kubenswrapper[4754]: I1005 21:15:59.748343 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-845d6d6f59-886x9" podUID="1bb3793f-57ce-4f10-9452-30ff52baabda" containerName="dnsmasq-dns" containerID="cri-o://b744b5d6b2440cfdb316893cc10ee6f27f818aef3a954222567332e928539f0f" gracePeriod=10 Oct 05 21:16:00 crc kubenswrapper[4754]: I1005 21:16:00.270187 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-845d6d6f59-886x9" Oct 05 21:16:00 crc kubenswrapper[4754]: I1005 21:16:00.277062 4754 generic.go:334] "Generic (PLEG): container finished" podID="1bb3793f-57ce-4f10-9452-30ff52baabda" containerID="b744b5d6b2440cfdb316893cc10ee6f27f818aef3a954222567332e928539f0f" exitCode=0 Oct 05 21:16:00 crc kubenswrapper[4754]: I1005 21:16:00.277106 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-886x9" event={"ID":"1bb3793f-57ce-4f10-9452-30ff52baabda","Type":"ContainerDied","Data":"b744b5d6b2440cfdb316893cc10ee6f27f818aef3a954222567332e928539f0f"} Oct 05 21:16:00 crc kubenswrapper[4754]: I1005 21:16:00.277132 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-845d6d6f59-886x9" Oct 05 21:16:00 crc kubenswrapper[4754]: I1005 21:16:00.277395 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-886x9" event={"ID":"1bb3793f-57ce-4f10-9452-30ff52baabda","Type":"ContainerDied","Data":"81a35e8f3d10eab42b989d814dcb355c9d07d4154711205819ed25e2f470bc43"} Oct 05 21:16:00 crc kubenswrapper[4754]: I1005 21:16:00.277426 4754 scope.go:117] "RemoveContainer" containerID="b744b5d6b2440cfdb316893cc10ee6f27f818aef3a954222567332e928539f0f" Oct 05 21:16:00 crc kubenswrapper[4754]: I1005 21:16:00.281289 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b5010900-7b57-48e9-9621-73940c2aefa1","Type":"ContainerStarted","Data":"2e48ca2f6232d552ac4f511aa6f0ab11095079acf2f3068b6612588344fc3489"} Oct 05 21:16:00 crc kubenswrapper[4754]: I1005 21:16:00.281695 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 05 21:16:00 crc kubenswrapper[4754]: I1005 21:16:00.310942 4754 scope.go:117] "RemoveContainer" containerID="0b410ed9181022b6ebd1849808b46acb9dd06c414a2cfbeabb1c6c56fabbf760" Oct 05 21:16:00 crc kubenswrapper[4754]: I1005 21:16:00.334784 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.256569078 podStartE2EDuration="6.334758161s" podCreationTimestamp="2025-10-05 21:15:54 +0000 UTC" firstStartedPulling="2025-10-05 21:15:55.374513427 +0000 UTC m=+1279.278632157" lastFinishedPulling="2025-10-05 21:15:59.4527025 +0000 UTC m=+1283.356821240" observedRunningTime="2025-10-05 21:16:00.333373934 +0000 UTC m=+1284.237492644" watchObservedRunningTime="2025-10-05 21:16:00.334758161 +0000 UTC m=+1284.238876871" Oct 05 21:16:00 crc kubenswrapper[4754]: I1005 21:16:00.353091 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1bb3793f-57ce-4f10-9452-30ff52baabda-dns-swift-storage-0\") pod \"1bb3793f-57ce-4f10-9452-30ff52baabda\" (UID: \"1bb3793f-57ce-4f10-9452-30ff52baabda\") " Oct 05 21:16:00 crc kubenswrapper[4754]: I1005 21:16:00.353396 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1bb3793f-57ce-4f10-9452-30ff52baabda-ovsdbserver-sb\") pod \"1bb3793f-57ce-4f10-9452-30ff52baabda\" (UID: \"1bb3793f-57ce-4f10-9452-30ff52baabda\") " Oct 05 21:16:00 crc kubenswrapper[4754]: I1005 21:16:00.353457 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bb3793f-57ce-4f10-9452-30ff52baabda-config\") pod \"1bb3793f-57ce-4f10-9452-30ff52baabda\" (UID: \"1bb3793f-57ce-4f10-9452-30ff52baabda\") " Oct 05 21:16:00 crc kubenswrapper[4754]: I1005 21:16:00.353536 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-74fxg\" (UniqueName: \"kubernetes.io/projected/1bb3793f-57ce-4f10-9452-30ff52baabda-kube-api-access-74fxg\") pod \"1bb3793f-57ce-4f10-9452-30ff52baabda\" (UID: \"1bb3793f-57ce-4f10-9452-30ff52baabda\") " Oct 05 21:16:00 crc kubenswrapper[4754]: I1005 21:16:00.353615 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1bb3793f-57ce-4f10-9452-30ff52baabda-dns-svc\") pod \"1bb3793f-57ce-4f10-9452-30ff52baabda\" (UID: \"1bb3793f-57ce-4f10-9452-30ff52baabda\") " Oct 05 21:16:00 crc kubenswrapper[4754]: I1005 21:16:00.353643 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1bb3793f-57ce-4f10-9452-30ff52baabda-ovsdbserver-nb\") pod \"1bb3793f-57ce-4f10-9452-30ff52baabda\" (UID: \"1bb3793f-57ce-4f10-9452-30ff52baabda\") " Oct 05 21:16:00 crc kubenswrapper[4754]: I1005 21:16:00.365930 4754 scope.go:117] "RemoveContainer" containerID="b744b5d6b2440cfdb316893cc10ee6f27f818aef3a954222567332e928539f0f" Oct 05 21:16:00 crc kubenswrapper[4754]: I1005 21:16:00.367447 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bb3793f-57ce-4f10-9452-30ff52baabda-kube-api-access-74fxg" (OuterVolumeSpecName: "kube-api-access-74fxg") pod "1bb3793f-57ce-4f10-9452-30ff52baabda" (UID: "1bb3793f-57ce-4f10-9452-30ff52baabda"). InnerVolumeSpecName "kube-api-access-74fxg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:16:00 crc kubenswrapper[4754]: E1005 21:16:00.380322 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b744b5d6b2440cfdb316893cc10ee6f27f818aef3a954222567332e928539f0f\": container with ID starting with b744b5d6b2440cfdb316893cc10ee6f27f818aef3a954222567332e928539f0f not found: ID does not exist" containerID="b744b5d6b2440cfdb316893cc10ee6f27f818aef3a954222567332e928539f0f" Oct 05 21:16:00 crc kubenswrapper[4754]: I1005 21:16:00.380399 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b744b5d6b2440cfdb316893cc10ee6f27f818aef3a954222567332e928539f0f"} err="failed to get container status \"b744b5d6b2440cfdb316893cc10ee6f27f818aef3a954222567332e928539f0f\": rpc error: code = NotFound desc = could not find container \"b744b5d6b2440cfdb316893cc10ee6f27f818aef3a954222567332e928539f0f\": container with ID starting with b744b5d6b2440cfdb316893cc10ee6f27f818aef3a954222567332e928539f0f not found: ID does not exist" Oct 05 21:16:00 crc kubenswrapper[4754]: I1005 21:16:00.380430 4754 scope.go:117] "RemoveContainer" containerID="0b410ed9181022b6ebd1849808b46acb9dd06c414a2cfbeabb1c6c56fabbf760" Oct 05 21:16:00 crc kubenswrapper[4754]: E1005 21:16:00.390437 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0b410ed9181022b6ebd1849808b46acb9dd06c414a2cfbeabb1c6c56fabbf760\": container with ID starting with 0b410ed9181022b6ebd1849808b46acb9dd06c414a2cfbeabb1c6c56fabbf760 not found: ID does not exist" containerID="0b410ed9181022b6ebd1849808b46acb9dd06c414a2cfbeabb1c6c56fabbf760" Oct 05 21:16:00 crc kubenswrapper[4754]: I1005 21:16:00.390562 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0b410ed9181022b6ebd1849808b46acb9dd06c414a2cfbeabb1c6c56fabbf760"} err="failed to get container status \"0b410ed9181022b6ebd1849808b46acb9dd06c414a2cfbeabb1c6c56fabbf760\": rpc error: code = NotFound desc = could not find container \"0b410ed9181022b6ebd1849808b46acb9dd06c414a2cfbeabb1c6c56fabbf760\": container with ID starting with 0b410ed9181022b6ebd1849808b46acb9dd06c414a2cfbeabb1c6c56fabbf760 not found: ID does not exist" Oct 05 21:16:00 crc kubenswrapper[4754]: I1005 21:16:00.442119 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bb3793f-57ce-4f10-9452-30ff52baabda-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "1bb3793f-57ce-4f10-9452-30ff52baabda" (UID: "1bb3793f-57ce-4f10-9452-30ff52baabda"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 21:16:00 crc kubenswrapper[4754]: I1005 21:16:00.444041 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bb3793f-57ce-4f10-9452-30ff52baabda-config" (OuterVolumeSpecName: "config") pod "1bb3793f-57ce-4f10-9452-30ff52baabda" (UID: "1bb3793f-57ce-4f10-9452-30ff52baabda"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 21:16:00 crc kubenswrapper[4754]: I1005 21:16:00.455962 4754 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bb3793f-57ce-4f10-9452-30ff52baabda-config\") on node \"crc\" DevicePath \"\"" Oct 05 21:16:00 crc kubenswrapper[4754]: I1005 21:16:00.455991 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-74fxg\" (UniqueName: \"kubernetes.io/projected/1bb3793f-57ce-4f10-9452-30ff52baabda-kube-api-access-74fxg\") on node \"crc\" DevicePath \"\"" Oct 05 21:16:00 crc kubenswrapper[4754]: I1005 21:16:00.456002 4754 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1bb3793f-57ce-4f10-9452-30ff52baabda-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 05 21:16:00 crc kubenswrapper[4754]: I1005 21:16:00.456878 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bb3793f-57ce-4f10-9452-30ff52baabda-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "1bb3793f-57ce-4f10-9452-30ff52baabda" (UID: "1bb3793f-57ce-4f10-9452-30ff52baabda"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 21:16:00 crc kubenswrapper[4754]: I1005 21:16:00.460620 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bb3793f-57ce-4f10-9452-30ff52baabda-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "1bb3793f-57ce-4f10-9452-30ff52baabda" (UID: "1bb3793f-57ce-4f10-9452-30ff52baabda"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 21:16:00 crc kubenswrapper[4754]: I1005 21:16:00.486916 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bb3793f-57ce-4f10-9452-30ff52baabda-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "1bb3793f-57ce-4f10-9452-30ff52baabda" (UID: "1bb3793f-57ce-4f10-9452-30ff52baabda"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 21:16:00 crc kubenswrapper[4754]: I1005 21:16:00.557777 4754 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1bb3793f-57ce-4f10-9452-30ff52baabda-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 05 21:16:00 crc kubenswrapper[4754]: I1005 21:16:00.558137 4754 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1bb3793f-57ce-4f10-9452-30ff52baabda-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 05 21:16:00 crc kubenswrapper[4754]: I1005 21:16:00.558206 4754 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1bb3793f-57ce-4f10-9452-30ff52baabda-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 05 21:16:00 crc kubenswrapper[4754]: I1005 21:16:00.609004 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-886x9"] Oct 05 21:16:00 crc kubenswrapper[4754]: I1005 21:16:00.615736 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-886x9"] Oct 05 21:16:00 crc kubenswrapper[4754]: I1005 21:16:00.848483 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bb3793f-57ce-4f10-9452-30ff52baabda" path="/var/lib/kubelet/pods/1bb3793f-57ce-4f10-9452-30ff52baabda/volumes" Oct 05 21:16:03 crc kubenswrapper[4754]: I1005 21:16:03.312181 4754 generic.go:334] "Generic (PLEG): container finished" podID="ba3c8f4e-033c-48d4-b37e-7b2542be3bd1" containerID="aff0a26aedb047df62a0a5b78a35b03c9c74f10020d36a1af66b2a6b57901861" exitCode=0 Oct 05 21:16:03 crc kubenswrapper[4754]: I1005 21:16:03.313198 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-stkc2" event={"ID":"ba3c8f4e-033c-48d4-b37e-7b2542be3bd1","Type":"ContainerDied","Data":"aff0a26aedb047df62a0a5b78a35b03c9c74f10020d36a1af66b2a6b57901861"} Oct 05 21:16:04 crc kubenswrapper[4754]: I1005 21:16:04.805286 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-stkc2" Oct 05 21:16:04 crc kubenswrapper[4754]: I1005 21:16:04.953601 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba3c8f4e-033c-48d4-b37e-7b2542be3bd1-combined-ca-bundle\") pod \"ba3c8f4e-033c-48d4-b37e-7b2542be3bd1\" (UID: \"ba3c8f4e-033c-48d4-b37e-7b2542be3bd1\") " Oct 05 21:16:04 crc kubenswrapper[4754]: I1005 21:16:04.954419 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zqxxx\" (UniqueName: \"kubernetes.io/projected/ba3c8f4e-033c-48d4-b37e-7b2542be3bd1-kube-api-access-zqxxx\") pod \"ba3c8f4e-033c-48d4-b37e-7b2542be3bd1\" (UID: \"ba3c8f4e-033c-48d4-b37e-7b2542be3bd1\") " Oct 05 21:16:04 crc kubenswrapper[4754]: I1005 21:16:04.954472 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ba3c8f4e-033c-48d4-b37e-7b2542be3bd1-scripts\") pod \"ba3c8f4e-033c-48d4-b37e-7b2542be3bd1\" (UID: \"ba3c8f4e-033c-48d4-b37e-7b2542be3bd1\") " Oct 05 21:16:04 crc kubenswrapper[4754]: I1005 21:16:04.954592 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ba3c8f4e-033c-48d4-b37e-7b2542be3bd1-config-data\") pod \"ba3c8f4e-033c-48d4-b37e-7b2542be3bd1\" (UID: \"ba3c8f4e-033c-48d4-b37e-7b2542be3bd1\") " Oct 05 21:16:04 crc kubenswrapper[4754]: I1005 21:16:04.966830 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ba3c8f4e-033c-48d4-b37e-7b2542be3bd1-kube-api-access-zqxxx" (OuterVolumeSpecName: "kube-api-access-zqxxx") pod "ba3c8f4e-033c-48d4-b37e-7b2542be3bd1" (UID: "ba3c8f4e-033c-48d4-b37e-7b2542be3bd1"). InnerVolumeSpecName "kube-api-access-zqxxx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:16:04 crc kubenswrapper[4754]: I1005 21:16:04.967028 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ba3c8f4e-033c-48d4-b37e-7b2542be3bd1-scripts" (OuterVolumeSpecName: "scripts") pod "ba3c8f4e-033c-48d4-b37e-7b2542be3bd1" (UID: "ba3c8f4e-033c-48d4-b37e-7b2542be3bd1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:16:04 crc kubenswrapper[4754]: I1005 21:16:04.995752 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ba3c8f4e-033c-48d4-b37e-7b2542be3bd1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ba3c8f4e-033c-48d4-b37e-7b2542be3bd1" (UID: "ba3c8f4e-033c-48d4-b37e-7b2542be3bd1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:16:04 crc kubenswrapper[4754]: I1005 21:16:04.995739 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ba3c8f4e-033c-48d4-b37e-7b2542be3bd1-config-data" (OuterVolumeSpecName: "config-data") pod "ba3c8f4e-033c-48d4-b37e-7b2542be3bd1" (UID: "ba3c8f4e-033c-48d4-b37e-7b2542be3bd1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:16:05 crc kubenswrapper[4754]: I1005 21:16:05.057669 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zqxxx\" (UniqueName: \"kubernetes.io/projected/ba3c8f4e-033c-48d4-b37e-7b2542be3bd1-kube-api-access-zqxxx\") on node \"crc\" DevicePath \"\"" Oct 05 21:16:05 crc kubenswrapper[4754]: I1005 21:16:05.057715 4754 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ba3c8f4e-033c-48d4-b37e-7b2542be3bd1-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 21:16:05 crc kubenswrapper[4754]: I1005 21:16:05.057735 4754 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ba3c8f4e-033c-48d4-b37e-7b2542be3bd1-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 21:16:05 crc kubenswrapper[4754]: I1005 21:16:05.057753 4754 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba3c8f4e-033c-48d4-b37e-7b2542be3bd1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 21:16:05 crc kubenswrapper[4754]: I1005 21:16:05.372859 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-stkc2" event={"ID":"ba3c8f4e-033c-48d4-b37e-7b2542be3bd1","Type":"ContainerDied","Data":"a22929e023b20c2dda1c8b0e060862a3270349d306f68279cf3c92fa0b2c0474"} Oct 05 21:16:05 crc kubenswrapper[4754]: I1005 21:16:05.372921 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a22929e023b20c2dda1c8b0e060862a3270349d306f68279cf3c92fa0b2c0474" Oct 05 21:16:05 crc kubenswrapper[4754]: I1005 21:16:05.373033 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-stkc2" Oct 05 21:16:05 crc kubenswrapper[4754]: I1005 21:16:05.546609 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 05 21:16:05 crc kubenswrapper[4754]: I1005 21:16:05.546895 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="95b27888-bd9d-4ebf-a0d8-9f6653d3fec5" containerName="nova-api-log" containerID="cri-o://a1fdfabbf98552e60eeeae588f22f843362cec9781110d97076049277412098a" gracePeriod=30 Oct 05 21:16:05 crc kubenswrapper[4754]: I1005 21:16:05.547065 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="95b27888-bd9d-4ebf-a0d8-9f6653d3fec5" containerName="nova-api-api" containerID="cri-o://4ac7d7f41e4b7a9a3f01256e0411d11ff35d07006942978c7ee3635980fc5fc8" gracePeriod=30 Oct 05 21:16:05 crc kubenswrapper[4754]: I1005 21:16:05.568010 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 05 21:16:05 crc kubenswrapper[4754]: I1005 21:16:05.568324 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="4b820b69-1578-4f07-9819-246603cdd777" containerName="nova-scheduler-scheduler" containerID="cri-o://06e6bd3415112bee3d9524ee46f595f8d3cca1a787b692ef83d0b0bc782a9aae" gracePeriod=30 Oct 05 21:16:05 crc kubenswrapper[4754]: I1005 21:16:05.604301 4754 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-68c6b67864-9msm5" podUID="27cd6be5-f5eb-4158-bdd1-37f4df3d530d" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.148:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.148:8443: connect: connection refused" Oct 05 21:16:05 crc kubenswrapper[4754]: I1005 21:16:05.604424 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-68c6b67864-9msm5" Oct 05 21:16:05 crc kubenswrapper[4754]: I1005 21:16:05.645937 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 05 21:16:05 crc kubenswrapper[4754]: I1005 21:16:05.646204 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="c45e75d9-997b-4b2b-a923-8ee9739fcbbc" containerName="nova-metadata-log" containerID="cri-o://c6039287394c53d920bccc08c1a5fc75ad1e47d196696fb83bcc68a3e88c13b5" gracePeriod=30 Oct 05 21:16:05 crc kubenswrapper[4754]: I1005 21:16:05.646325 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="c45e75d9-997b-4b2b-a923-8ee9739fcbbc" containerName="nova-metadata-metadata" containerID="cri-o://40f7e9db9c4cfc1a4743fb8a28bcd85bf3946e7cfabbb175345526f616f3cd1d" gracePeriod=30 Oct 05 21:16:06 crc kubenswrapper[4754]: I1005 21:16:06.147361 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 05 21:16:06 crc kubenswrapper[4754]: I1005 21:16:06.292901 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95b27888-bd9d-4ebf-a0d8-9f6653d3fec5-combined-ca-bundle\") pod \"95b27888-bd9d-4ebf-a0d8-9f6653d3fec5\" (UID: \"95b27888-bd9d-4ebf-a0d8-9f6653d3fec5\") " Oct 05 21:16:06 crc kubenswrapper[4754]: I1005 21:16:06.293684 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95b27888-bd9d-4ebf-a0d8-9f6653d3fec5-config-data\") pod \"95b27888-bd9d-4ebf-a0d8-9f6653d3fec5\" (UID: \"95b27888-bd9d-4ebf-a0d8-9f6653d3fec5\") " Oct 05 21:16:06 crc kubenswrapper[4754]: I1005 21:16:06.294217 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/95b27888-bd9d-4ebf-a0d8-9f6653d3fec5-public-tls-certs\") pod \"95b27888-bd9d-4ebf-a0d8-9f6653d3fec5\" (UID: \"95b27888-bd9d-4ebf-a0d8-9f6653d3fec5\") " Oct 05 21:16:06 crc kubenswrapper[4754]: I1005 21:16:06.294389 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sz5lh\" (UniqueName: \"kubernetes.io/projected/95b27888-bd9d-4ebf-a0d8-9f6653d3fec5-kube-api-access-sz5lh\") pod \"95b27888-bd9d-4ebf-a0d8-9f6653d3fec5\" (UID: \"95b27888-bd9d-4ebf-a0d8-9f6653d3fec5\") " Oct 05 21:16:06 crc kubenswrapper[4754]: I1005 21:16:06.294643 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/95b27888-bd9d-4ebf-a0d8-9f6653d3fec5-internal-tls-certs\") pod \"95b27888-bd9d-4ebf-a0d8-9f6653d3fec5\" (UID: \"95b27888-bd9d-4ebf-a0d8-9f6653d3fec5\") " Oct 05 21:16:06 crc kubenswrapper[4754]: I1005 21:16:06.295048 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/95b27888-bd9d-4ebf-a0d8-9f6653d3fec5-logs\") pod \"95b27888-bd9d-4ebf-a0d8-9f6653d3fec5\" (UID: \"95b27888-bd9d-4ebf-a0d8-9f6653d3fec5\") " Oct 05 21:16:06 crc kubenswrapper[4754]: I1005 21:16:06.295873 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/95b27888-bd9d-4ebf-a0d8-9f6653d3fec5-logs" (OuterVolumeSpecName: "logs") pod "95b27888-bd9d-4ebf-a0d8-9f6653d3fec5" (UID: "95b27888-bd9d-4ebf-a0d8-9f6653d3fec5"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 21:16:06 crc kubenswrapper[4754]: I1005 21:16:06.320719 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/95b27888-bd9d-4ebf-a0d8-9f6653d3fec5-kube-api-access-sz5lh" (OuterVolumeSpecName: "kube-api-access-sz5lh") pod "95b27888-bd9d-4ebf-a0d8-9f6653d3fec5" (UID: "95b27888-bd9d-4ebf-a0d8-9f6653d3fec5"). InnerVolumeSpecName "kube-api-access-sz5lh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:16:06 crc kubenswrapper[4754]: I1005 21:16:06.330581 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/95b27888-bd9d-4ebf-a0d8-9f6653d3fec5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "95b27888-bd9d-4ebf-a0d8-9f6653d3fec5" (UID: "95b27888-bd9d-4ebf-a0d8-9f6653d3fec5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:16:06 crc kubenswrapper[4754]: I1005 21:16:06.330938 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/95b27888-bd9d-4ebf-a0d8-9f6653d3fec5-config-data" (OuterVolumeSpecName: "config-data") pod "95b27888-bd9d-4ebf-a0d8-9f6653d3fec5" (UID: "95b27888-bd9d-4ebf-a0d8-9f6653d3fec5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:16:06 crc kubenswrapper[4754]: I1005 21:16:06.361425 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/95b27888-bd9d-4ebf-a0d8-9f6653d3fec5-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "95b27888-bd9d-4ebf-a0d8-9f6653d3fec5" (UID: "95b27888-bd9d-4ebf-a0d8-9f6653d3fec5"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:16:06 crc kubenswrapper[4754]: I1005 21:16:06.388072 4754 generic.go:334] "Generic (PLEG): container finished" podID="95b27888-bd9d-4ebf-a0d8-9f6653d3fec5" containerID="4ac7d7f41e4b7a9a3f01256e0411d11ff35d07006942978c7ee3635980fc5fc8" exitCode=0 Oct 05 21:16:06 crc kubenswrapper[4754]: I1005 21:16:06.388106 4754 generic.go:334] "Generic (PLEG): container finished" podID="95b27888-bd9d-4ebf-a0d8-9f6653d3fec5" containerID="a1fdfabbf98552e60eeeae588f22f843362cec9781110d97076049277412098a" exitCode=143 Oct 05 21:16:06 crc kubenswrapper[4754]: I1005 21:16:06.388151 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 05 21:16:06 crc kubenswrapper[4754]: I1005 21:16:06.388161 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"95b27888-bd9d-4ebf-a0d8-9f6653d3fec5","Type":"ContainerDied","Data":"4ac7d7f41e4b7a9a3f01256e0411d11ff35d07006942978c7ee3635980fc5fc8"} Oct 05 21:16:06 crc kubenswrapper[4754]: I1005 21:16:06.388197 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"95b27888-bd9d-4ebf-a0d8-9f6653d3fec5","Type":"ContainerDied","Data":"a1fdfabbf98552e60eeeae588f22f843362cec9781110d97076049277412098a"} Oct 05 21:16:06 crc kubenswrapper[4754]: I1005 21:16:06.388208 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"95b27888-bd9d-4ebf-a0d8-9f6653d3fec5","Type":"ContainerDied","Data":"bd1014b234168d04dc783a70f78735175b155178927ffb3da0b353a36d3aa1a8"} Oct 05 21:16:06 crc kubenswrapper[4754]: I1005 21:16:06.388223 4754 scope.go:117] "RemoveContainer" containerID="4ac7d7f41e4b7a9a3f01256e0411d11ff35d07006942978c7ee3635980fc5fc8" Oct 05 21:16:06 crc kubenswrapper[4754]: I1005 21:16:06.394830 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/95b27888-bd9d-4ebf-a0d8-9f6653d3fec5-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "95b27888-bd9d-4ebf-a0d8-9f6653d3fec5" (UID: "95b27888-bd9d-4ebf-a0d8-9f6653d3fec5"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:16:06 crc kubenswrapper[4754]: I1005 21:16:06.398636 4754 generic.go:334] "Generic (PLEG): container finished" podID="c45e75d9-997b-4b2b-a923-8ee9739fcbbc" containerID="c6039287394c53d920bccc08c1a5fc75ad1e47d196696fb83bcc68a3e88c13b5" exitCode=143 Oct 05 21:16:06 crc kubenswrapper[4754]: I1005 21:16:06.398671 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c45e75d9-997b-4b2b-a923-8ee9739fcbbc","Type":"ContainerDied","Data":"c6039287394c53d920bccc08c1a5fc75ad1e47d196696fb83bcc68a3e88c13b5"} Oct 05 21:16:06 crc kubenswrapper[4754]: I1005 21:16:06.408718 4754 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95b27888-bd9d-4ebf-a0d8-9f6653d3fec5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 21:16:06 crc kubenswrapper[4754]: I1005 21:16:06.408757 4754 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95b27888-bd9d-4ebf-a0d8-9f6653d3fec5-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 21:16:06 crc kubenswrapper[4754]: I1005 21:16:06.408767 4754 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/95b27888-bd9d-4ebf-a0d8-9f6653d3fec5-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 05 21:16:06 crc kubenswrapper[4754]: I1005 21:16:06.408778 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sz5lh\" (UniqueName: \"kubernetes.io/projected/95b27888-bd9d-4ebf-a0d8-9f6653d3fec5-kube-api-access-sz5lh\") on node \"crc\" DevicePath \"\"" Oct 05 21:16:06 crc kubenswrapper[4754]: I1005 21:16:06.408793 4754 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/95b27888-bd9d-4ebf-a0d8-9f6653d3fec5-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 05 21:16:06 crc kubenswrapper[4754]: I1005 21:16:06.408801 4754 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/95b27888-bd9d-4ebf-a0d8-9f6653d3fec5-logs\") on node \"crc\" DevicePath \"\"" Oct 05 21:16:06 crc kubenswrapper[4754]: I1005 21:16:06.447611 4754 scope.go:117] "RemoveContainer" containerID="a1fdfabbf98552e60eeeae588f22f843362cec9781110d97076049277412098a" Oct 05 21:16:06 crc kubenswrapper[4754]: I1005 21:16:06.465275 4754 scope.go:117] "RemoveContainer" containerID="4ac7d7f41e4b7a9a3f01256e0411d11ff35d07006942978c7ee3635980fc5fc8" Oct 05 21:16:06 crc kubenswrapper[4754]: E1005 21:16:06.465737 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4ac7d7f41e4b7a9a3f01256e0411d11ff35d07006942978c7ee3635980fc5fc8\": container with ID starting with 4ac7d7f41e4b7a9a3f01256e0411d11ff35d07006942978c7ee3635980fc5fc8 not found: ID does not exist" containerID="4ac7d7f41e4b7a9a3f01256e0411d11ff35d07006942978c7ee3635980fc5fc8" Oct 05 21:16:06 crc kubenswrapper[4754]: I1005 21:16:06.465784 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4ac7d7f41e4b7a9a3f01256e0411d11ff35d07006942978c7ee3635980fc5fc8"} err="failed to get container status \"4ac7d7f41e4b7a9a3f01256e0411d11ff35d07006942978c7ee3635980fc5fc8\": rpc error: code = NotFound desc = could not find container \"4ac7d7f41e4b7a9a3f01256e0411d11ff35d07006942978c7ee3635980fc5fc8\": container with ID starting with 4ac7d7f41e4b7a9a3f01256e0411d11ff35d07006942978c7ee3635980fc5fc8 not found: ID does not exist" Oct 05 21:16:06 crc kubenswrapper[4754]: I1005 21:16:06.465998 4754 scope.go:117] "RemoveContainer" containerID="a1fdfabbf98552e60eeeae588f22f843362cec9781110d97076049277412098a" Oct 05 21:16:06 crc kubenswrapper[4754]: E1005 21:16:06.466419 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a1fdfabbf98552e60eeeae588f22f843362cec9781110d97076049277412098a\": container with ID starting with a1fdfabbf98552e60eeeae588f22f843362cec9781110d97076049277412098a not found: ID does not exist" containerID="a1fdfabbf98552e60eeeae588f22f843362cec9781110d97076049277412098a" Oct 05 21:16:06 crc kubenswrapper[4754]: I1005 21:16:06.466457 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a1fdfabbf98552e60eeeae588f22f843362cec9781110d97076049277412098a"} err="failed to get container status \"a1fdfabbf98552e60eeeae588f22f843362cec9781110d97076049277412098a\": rpc error: code = NotFound desc = could not find container \"a1fdfabbf98552e60eeeae588f22f843362cec9781110d97076049277412098a\": container with ID starting with a1fdfabbf98552e60eeeae588f22f843362cec9781110d97076049277412098a not found: ID does not exist" Oct 05 21:16:06 crc kubenswrapper[4754]: I1005 21:16:06.466480 4754 scope.go:117] "RemoveContainer" containerID="4ac7d7f41e4b7a9a3f01256e0411d11ff35d07006942978c7ee3635980fc5fc8" Oct 05 21:16:06 crc kubenswrapper[4754]: I1005 21:16:06.466823 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4ac7d7f41e4b7a9a3f01256e0411d11ff35d07006942978c7ee3635980fc5fc8"} err="failed to get container status \"4ac7d7f41e4b7a9a3f01256e0411d11ff35d07006942978c7ee3635980fc5fc8\": rpc error: code = NotFound desc = could not find container \"4ac7d7f41e4b7a9a3f01256e0411d11ff35d07006942978c7ee3635980fc5fc8\": container with ID starting with 4ac7d7f41e4b7a9a3f01256e0411d11ff35d07006942978c7ee3635980fc5fc8 not found: ID does not exist" Oct 05 21:16:06 crc kubenswrapper[4754]: I1005 21:16:06.466850 4754 scope.go:117] "RemoveContainer" containerID="a1fdfabbf98552e60eeeae588f22f843362cec9781110d97076049277412098a" Oct 05 21:16:06 crc kubenswrapper[4754]: I1005 21:16:06.467158 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a1fdfabbf98552e60eeeae588f22f843362cec9781110d97076049277412098a"} err="failed to get container status \"a1fdfabbf98552e60eeeae588f22f843362cec9781110d97076049277412098a\": rpc error: code = NotFound desc = could not find container \"a1fdfabbf98552e60eeeae588f22f843362cec9781110d97076049277412098a\": container with ID starting with a1fdfabbf98552e60eeeae588f22f843362cec9781110d97076049277412098a not found: ID does not exist" Oct 05 21:16:06 crc kubenswrapper[4754]: I1005 21:16:06.722516 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 05 21:16:06 crc kubenswrapper[4754]: I1005 21:16:06.731609 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 05 21:16:06 crc kubenswrapper[4754]: I1005 21:16:06.746511 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 05 21:16:06 crc kubenswrapper[4754]: E1005 21:16:06.746892 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95b27888-bd9d-4ebf-a0d8-9f6653d3fec5" containerName="nova-api-log" Oct 05 21:16:06 crc kubenswrapper[4754]: I1005 21:16:06.746908 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="95b27888-bd9d-4ebf-a0d8-9f6653d3fec5" containerName="nova-api-log" Oct 05 21:16:06 crc kubenswrapper[4754]: E1005 21:16:06.746924 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bb3793f-57ce-4f10-9452-30ff52baabda" containerName="init" Oct 05 21:16:06 crc kubenswrapper[4754]: I1005 21:16:06.746933 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bb3793f-57ce-4f10-9452-30ff52baabda" containerName="init" Oct 05 21:16:06 crc kubenswrapper[4754]: E1005 21:16:06.746945 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bb3793f-57ce-4f10-9452-30ff52baabda" containerName="dnsmasq-dns" Oct 05 21:16:06 crc kubenswrapper[4754]: I1005 21:16:06.746952 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bb3793f-57ce-4f10-9452-30ff52baabda" containerName="dnsmasq-dns" Oct 05 21:16:06 crc kubenswrapper[4754]: E1005 21:16:06.746964 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba3c8f4e-033c-48d4-b37e-7b2542be3bd1" containerName="nova-manage" Oct 05 21:16:06 crc kubenswrapper[4754]: I1005 21:16:06.746971 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba3c8f4e-033c-48d4-b37e-7b2542be3bd1" containerName="nova-manage" Oct 05 21:16:06 crc kubenswrapper[4754]: E1005 21:16:06.746994 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95b27888-bd9d-4ebf-a0d8-9f6653d3fec5" containerName="nova-api-api" Oct 05 21:16:06 crc kubenswrapper[4754]: I1005 21:16:06.747000 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="95b27888-bd9d-4ebf-a0d8-9f6653d3fec5" containerName="nova-api-api" Oct 05 21:16:06 crc kubenswrapper[4754]: I1005 21:16:06.747179 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="95b27888-bd9d-4ebf-a0d8-9f6653d3fec5" containerName="nova-api-api" Oct 05 21:16:06 crc kubenswrapper[4754]: I1005 21:16:06.747192 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="ba3c8f4e-033c-48d4-b37e-7b2542be3bd1" containerName="nova-manage" Oct 05 21:16:06 crc kubenswrapper[4754]: I1005 21:16:06.747205 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="95b27888-bd9d-4ebf-a0d8-9f6653d3fec5" containerName="nova-api-log" Oct 05 21:16:06 crc kubenswrapper[4754]: I1005 21:16:06.747222 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="1bb3793f-57ce-4f10-9452-30ff52baabda" containerName="dnsmasq-dns" Oct 05 21:16:06 crc kubenswrapper[4754]: I1005 21:16:06.748178 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 05 21:16:06 crc kubenswrapper[4754]: I1005 21:16:06.751861 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Oct 05 21:16:06 crc kubenswrapper[4754]: I1005 21:16:06.752272 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 05 21:16:06 crc kubenswrapper[4754]: I1005 21:16:06.766714 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Oct 05 21:16:06 crc kubenswrapper[4754]: I1005 21:16:06.784853 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 05 21:16:06 crc kubenswrapper[4754]: I1005 21:16:06.820292 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/527adc21-0816-4fba-b7a1-22b3db2e1d73-public-tls-certs\") pod \"nova-api-0\" (UID: \"527adc21-0816-4fba-b7a1-22b3db2e1d73\") " pod="openstack/nova-api-0" Oct 05 21:16:06 crc kubenswrapper[4754]: I1005 21:16:06.820375 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/527adc21-0816-4fba-b7a1-22b3db2e1d73-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"527adc21-0816-4fba-b7a1-22b3db2e1d73\") " pod="openstack/nova-api-0" Oct 05 21:16:06 crc kubenswrapper[4754]: I1005 21:16:06.820442 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ljfhl\" (UniqueName: \"kubernetes.io/projected/527adc21-0816-4fba-b7a1-22b3db2e1d73-kube-api-access-ljfhl\") pod \"nova-api-0\" (UID: \"527adc21-0816-4fba-b7a1-22b3db2e1d73\") " pod="openstack/nova-api-0" Oct 05 21:16:06 crc kubenswrapper[4754]: I1005 21:16:06.820464 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/527adc21-0816-4fba-b7a1-22b3db2e1d73-logs\") pod \"nova-api-0\" (UID: \"527adc21-0816-4fba-b7a1-22b3db2e1d73\") " pod="openstack/nova-api-0" Oct 05 21:16:06 crc kubenswrapper[4754]: I1005 21:16:06.820502 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/527adc21-0816-4fba-b7a1-22b3db2e1d73-config-data\") pod \"nova-api-0\" (UID: \"527adc21-0816-4fba-b7a1-22b3db2e1d73\") " pod="openstack/nova-api-0" Oct 05 21:16:06 crc kubenswrapper[4754]: I1005 21:16:06.820550 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/527adc21-0816-4fba-b7a1-22b3db2e1d73-internal-tls-certs\") pod \"nova-api-0\" (UID: \"527adc21-0816-4fba-b7a1-22b3db2e1d73\") " pod="openstack/nova-api-0" Oct 05 21:16:06 crc kubenswrapper[4754]: I1005 21:16:06.847362 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="95b27888-bd9d-4ebf-a0d8-9f6653d3fec5" path="/var/lib/kubelet/pods/95b27888-bd9d-4ebf-a0d8-9f6653d3fec5/volumes" Oct 05 21:16:06 crc kubenswrapper[4754]: I1005 21:16:06.922017 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/527adc21-0816-4fba-b7a1-22b3db2e1d73-public-tls-certs\") pod \"nova-api-0\" (UID: \"527adc21-0816-4fba-b7a1-22b3db2e1d73\") " pod="openstack/nova-api-0" Oct 05 21:16:06 crc kubenswrapper[4754]: I1005 21:16:06.922074 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/527adc21-0816-4fba-b7a1-22b3db2e1d73-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"527adc21-0816-4fba-b7a1-22b3db2e1d73\") " pod="openstack/nova-api-0" Oct 05 21:16:06 crc kubenswrapper[4754]: I1005 21:16:06.922117 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ljfhl\" (UniqueName: \"kubernetes.io/projected/527adc21-0816-4fba-b7a1-22b3db2e1d73-kube-api-access-ljfhl\") pod \"nova-api-0\" (UID: \"527adc21-0816-4fba-b7a1-22b3db2e1d73\") " pod="openstack/nova-api-0" Oct 05 21:16:06 crc kubenswrapper[4754]: I1005 21:16:06.922136 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/527adc21-0816-4fba-b7a1-22b3db2e1d73-logs\") pod \"nova-api-0\" (UID: \"527adc21-0816-4fba-b7a1-22b3db2e1d73\") " pod="openstack/nova-api-0" Oct 05 21:16:06 crc kubenswrapper[4754]: I1005 21:16:06.922162 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/527adc21-0816-4fba-b7a1-22b3db2e1d73-config-data\") pod \"nova-api-0\" (UID: \"527adc21-0816-4fba-b7a1-22b3db2e1d73\") " pod="openstack/nova-api-0" Oct 05 21:16:06 crc kubenswrapper[4754]: I1005 21:16:06.922200 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/527adc21-0816-4fba-b7a1-22b3db2e1d73-internal-tls-certs\") pod \"nova-api-0\" (UID: \"527adc21-0816-4fba-b7a1-22b3db2e1d73\") " pod="openstack/nova-api-0" Oct 05 21:16:06 crc kubenswrapper[4754]: I1005 21:16:06.923942 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/527adc21-0816-4fba-b7a1-22b3db2e1d73-logs\") pod \"nova-api-0\" (UID: \"527adc21-0816-4fba-b7a1-22b3db2e1d73\") " pod="openstack/nova-api-0" Oct 05 21:16:06 crc kubenswrapper[4754]: I1005 21:16:06.927259 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/527adc21-0816-4fba-b7a1-22b3db2e1d73-config-data\") pod \"nova-api-0\" (UID: \"527adc21-0816-4fba-b7a1-22b3db2e1d73\") " pod="openstack/nova-api-0" Oct 05 21:16:06 crc kubenswrapper[4754]: I1005 21:16:06.928896 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/527adc21-0816-4fba-b7a1-22b3db2e1d73-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"527adc21-0816-4fba-b7a1-22b3db2e1d73\") " pod="openstack/nova-api-0" Oct 05 21:16:06 crc kubenswrapper[4754]: I1005 21:16:06.929806 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/527adc21-0816-4fba-b7a1-22b3db2e1d73-internal-tls-certs\") pod \"nova-api-0\" (UID: \"527adc21-0816-4fba-b7a1-22b3db2e1d73\") " pod="openstack/nova-api-0" Oct 05 21:16:06 crc kubenswrapper[4754]: I1005 21:16:06.932445 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/527adc21-0816-4fba-b7a1-22b3db2e1d73-public-tls-certs\") pod \"nova-api-0\" (UID: \"527adc21-0816-4fba-b7a1-22b3db2e1d73\") " pod="openstack/nova-api-0" Oct 05 21:16:06 crc kubenswrapper[4754]: I1005 21:16:06.946026 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ljfhl\" (UniqueName: \"kubernetes.io/projected/527adc21-0816-4fba-b7a1-22b3db2e1d73-kube-api-access-ljfhl\") pod \"nova-api-0\" (UID: \"527adc21-0816-4fba-b7a1-22b3db2e1d73\") " pod="openstack/nova-api-0" Oct 05 21:16:07 crc kubenswrapper[4754]: I1005 21:16:07.064618 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 05 21:16:07 crc kubenswrapper[4754]: I1005 21:16:07.153558 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 05 21:16:07 crc kubenswrapper[4754]: I1005 21:16:07.227211 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b820b69-1578-4f07-9819-246603cdd777-combined-ca-bundle\") pod \"4b820b69-1578-4f07-9819-246603cdd777\" (UID: \"4b820b69-1578-4f07-9819-246603cdd777\") " Oct 05 21:16:07 crc kubenswrapper[4754]: I1005 21:16:07.227367 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-22p2b\" (UniqueName: \"kubernetes.io/projected/4b820b69-1578-4f07-9819-246603cdd777-kube-api-access-22p2b\") pod \"4b820b69-1578-4f07-9819-246603cdd777\" (UID: \"4b820b69-1578-4f07-9819-246603cdd777\") " Oct 05 21:16:07 crc kubenswrapper[4754]: I1005 21:16:07.227448 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b820b69-1578-4f07-9819-246603cdd777-config-data\") pod \"4b820b69-1578-4f07-9819-246603cdd777\" (UID: \"4b820b69-1578-4f07-9819-246603cdd777\") " Oct 05 21:16:07 crc kubenswrapper[4754]: I1005 21:16:07.231794 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4b820b69-1578-4f07-9819-246603cdd777-kube-api-access-22p2b" (OuterVolumeSpecName: "kube-api-access-22p2b") pod "4b820b69-1578-4f07-9819-246603cdd777" (UID: "4b820b69-1578-4f07-9819-246603cdd777"). InnerVolumeSpecName "kube-api-access-22p2b". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:16:07 crc kubenswrapper[4754]: I1005 21:16:07.254880 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b820b69-1578-4f07-9819-246603cdd777-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4b820b69-1578-4f07-9819-246603cdd777" (UID: "4b820b69-1578-4f07-9819-246603cdd777"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:16:07 crc kubenswrapper[4754]: I1005 21:16:07.255176 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b820b69-1578-4f07-9819-246603cdd777-config-data" (OuterVolumeSpecName: "config-data") pod "4b820b69-1578-4f07-9819-246603cdd777" (UID: "4b820b69-1578-4f07-9819-246603cdd777"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:16:07 crc kubenswrapper[4754]: I1005 21:16:07.329797 4754 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b820b69-1578-4f07-9819-246603cdd777-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 21:16:07 crc kubenswrapper[4754]: I1005 21:16:07.329836 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-22p2b\" (UniqueName: \"kubernetes.io/projected/4b820b69-1578-4f07-9819-246603cdd777-kube-api-access-22p2b\") on node \"crc\" DevicePath \"\"" Oct 05 21:16:07 crc kubenswrapper[4754]: I1005 21:16:07.329846 4754 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b820b69-1578-4f07-9819-246603cdd777-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 21:16:07 crc kubenswrapper[4754]: I1005 21:16:07.410706 4754 generic.go:334] "Generic (PLEG): container finished" podID="4b820b69-1578-4f07-9819-246603cdd777" containerID="06e6bd3415112bee3d9524ee46f595f8d3cca1a787b692ef83d0b0bc782a9aae" exitCode=0 Oct 05 21:16:07 crc kubenswrapper[4754]: I1005 21:16:07.410765 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 05 21:16:07 crc kubenswrapper[4754]: I1005 21:16:07.410814 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"4b820b69-1578-4f07-9819-246603cdd777","Type":"ContainerDied","Data":"06e6bd3415112bee3d9524ee46f595f8d3cca1a787b692ef83d0b0bc782a9aae"} Oct 05 21:16:07 crc kubenswrapper[4754]: I1005 21:16:07.410895 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"4b820b69-1578-4f07-9819-246603cdd777","Type":"ContainerDied","Data":"f3949d2edb0f65cf87106622eed86f928b5d8ed4e9919654a5be798926a8e8c8"} Oct 05 21:16:07 crc kubenswrapper[4754]: I1005 21:16:07.410917 4754 scope.go:117] "RemoveContainer" containerID="06e6bd3415112bee3d9524ee46f595f8d3cca1a787b692ef83d0b0bc782a9aae" Oct 05 21:16:07 crc kubenswrapper[4754]: I1005 21:16:07.432243 4754 scope.go:117] "RemoveContainer" containerID="06e6bd3415112bee3d9524ee46f595f8d3cca1a787b692ef83d0b0bc782a9aae" Oct 05 21:16:07 crc kubenswrapper[4754]: E1005 21:16:07.432684 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"06e6bd3415112bee3d9524ee46f595f8d3cca1a787b692ef83d0b0bc782a9aae\": container with ID starting with 06e6bd3415112bee3d9524ee46f595f8d3cca1a787b692ef83d0b0bc782a9aae not found: ID does not exist" containerID="06e6bd3415112bee3d9524ee46f595f8d3cca1a787b692ef83d0b0bc782a9aae" Oct 05 21:16:07 crc kubenswrapper[4754]: I1005 21:16:07.432722 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"06e6bd3415112bee3d9524ee46f595f8d3cca1a787b692ef83d0b0bc782a9aae"} err="failed to get container status \"06e6bd3415112bee3d9524ee46f595f8d3cca1a787b692ef83d0b0bc782a9aae\": rpc error: code = NotFound desc = could not find container \"06e6bd3415112bee3d9524ee46f595f8d3cca1a787b692ef83d0b0bc782a9aae\": container with ID starting with 06e6bd3415112bee3d9524ee46f595f8d3cca1a787b692ef83d0b0bc782a9aae not found: ID does not exist" Oct 05 21:16:07 crc kubenswrapper[4754]: I1005 21:16:07.449978 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 05 21:16:07 crc kubenswrapper[4754]: I1005 21:16:07.458072 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 05 21:16:07 crc kubenswrapper[4754]: I1005 21:16:07.476024 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 05 21:16:07 crc kubenswrapper[4754]: E1005 21:16:07.476473 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b820b69-1578-4f07-9819-246603cdd777" containerName="nova-scheduler-scheduler" Oct 05 21:16:07 crc kubenswrapper[4754]: I1005 21:16:07.476503 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b820b69-1578-4f07-9819-246603cdd777" containerName="nova-scheduler-scheduler" Oct 05 21:16:07 crc kubenswrapper[4754]: I1005 21:16:07.476742 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b820b69-1578-4f07-9819-246603cdd777" containerName="nova-scheduler-scheduler" Oct 05 21:16:07 crc kubenswrapper[4754]: I1005 21:16:07.477449 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 05 21:16:07 crc kubenswrapper[4754]: I1005 21:16:07.479889 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 05 21:16:07 crc kubenswrapper[4754]: I1005 21:16:07.484014 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 05 21:16:07 crc kubenswrapper[4754]: I1005 21:16:07.535707 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3e33c8f1-8161-4919-98e0-971f46b33b05-config-data\") pod \"nova-scheduler-0\" (UID: \"3e33c8f1-8161-4919-98e0-971f46b33b05\") " pod="openstack/nova-scheduler-0" Oct 05 21:16:07 crc kubenswrapper[4754]: I1005 21:16:07.535756 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m6t2m\" (UniqueName: \"kubernetes.io/projected/3e33c8f1-8161-4919-98e0-971f46b33b05-kube-api-access-m6t2m\") pod \"nova-scheduler-0\" (UID: \"3e33c8f1-8161-4919-98e0-971f46b33b05\") " pod="openstack/nova-scheduler-0" Oct 05 21:16:07 crc kubenswrapper[4754]: I1005 21:16:07.535920 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e33c8f1-8161-4919-98e0-971f46b33b05-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"3e33c8f1-8161-4919-98e0-971f46b33b05\") " pod="openstack/nova-scheduler-0" Oct 05 21:16:07 crc kubenswrapper[4754]: I1005 21:16:07.540824 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 05 21:16:07 crc kubenswrapper[4754]: I1005 21:16:07.638997 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3e33c8f1-8161-4919-98e0-971f46b33b05-config-data\") pod \"nova-scheduler-0\" (UID: \"3e33c8f1-8161-4919-98e0-971f46b33b05\") " pod="openstack/nova-scheduler-0" Oct 05 21:16:07 crc kubenswrapper[4754]: I1005 21:16:07.639064 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m6t2m\" (UniqueName: \"kubernetes.io/projected/3e33c8f1-8161-4919-98e0-971f46b33b05-kube-api-access-m6t2m\") pod \"nova-scheduler-0\" (UID: \"3e33c8f1-8161-4919-98e0-971f46b33b05\") " pod="openstack/nova-scheduler-0" Oct 05 21:16:07 crc kubenswrapper[4754]: I1005 21:16:07.639100 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e33c8f1-8161-4919-98e0-971f46b33b05-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"3e33c8f1-8161-4919-98e0-971f46b33b05\") " pod="openstack/nova-scheduler-0" Oct 05 21:16:07 crc kubenswrapper[4754]: I1005 21:16:07.643853 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e33c8f1-8161-4919-98e0-971f46b33b05-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"3e33c8f1-8161-4919-98e0-971f46b33b05\") " pod="openstack/nova-scheduler-0" Oct 05 21:16:07 crc kubenswrapper[4754]: I1005 21:16:07.647449 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3e33c8f1-8161-4919-98e0-971f46b33b05-config-data\") pod \"nova-scheduler-0\" (UID: \"3e33c8f1-8161-4919-98e0-971f46b33b05\") " pod="openstack/nova-scheduler-0" Oct 05 21:16:07 crc kubenswrapper[4754]: I1005 21:16:07.658858 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m6t2m\" (UniqueName: \"kubernetes.io/projected/3e33c8f1-8161-4919-98e0-971f46b33b05-kube-api-access-m6t2m\") pod \"nova-scheduler-0\" (UID: \"3e33c8f1-8161-4919-98e0-971f46b33b05\") " pod="openstack/nova-scheduler-0" Oct 05 21:16:07 crc kubenswrapper[4754]: I1005 21:16:07.801057 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 05 21:16:08 crc kubenswrapper[4754]: I1005 21:16:08.425477 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"527adc21-0816-4fba-b7a1-22b3db2e1d73","Type":"ContainerStarted","Data":"b0518f36578d8084b699748a7b355855b7a1e5a7398aa468ac6be094a6ef4dce"} Oct 05 21:16:08 crc kubenswrapper[4754]: I1005 21:16:08.426023 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"527adc21-0816-4fba-b7a1-22b3db2e1d73","Type":"ContainerStarted","Data":"982f680fb87ea74baf4f36aad019d7450d8cb1da18442f01d30e07cbe39b4103"} Oct 05 21:16:08 crc kubenswrapper[4754]: I1005 21:16:08.426033 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"527adc21-0816-4fba-b7a1-22b3db2e1d73","Type":"ContainerStarted","Data":"3d4afd10571f2724b67ffb97d34a97dd7d08c9b8f0c6e2afa7d47c7544ab761f"} Oct 05 21:16:08 crc kubenswrapper[4754]: I1005 21:16:08.453726 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.453704266 podStartE2EDuration="2.453704266s" podCreationTimestamp="2025-10-05 21:16:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 21:16:08.444613061 +0000 UTC m=+1292.348731771" watchObservedRunningTime="2025-10-05 21:16:08.453704266 +0000 UTC m=+1292.357822976" Oct 05 21:16:08 crc kubenswrapper[4754]: I1005 21:16:08.853220 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4b820b69-1578-4f07-9819-246603cdd777" path="/var/lib/kubelet/pods/4b820b69-1578-4f07-9819-246603cdd777/volumes" Oct 05 21:16:09 crc kubenswrapper[4754]: I1005 21:16:09.060389 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 05 21:16:09 crc kubenswrapper[4754]: I1005 21:16:09.262338 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 05 21:16:09 crc kubenswrapper[4754]: I1005 21:16:09.284669 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/c45e75d9-997b-4b2b-a923-8ee9739fcbbc-nova-metadata-tls-certs\") pod \"c45e75d9-997b-4b2b-a923-8ee9739fcbbc\" (UID: \"c45e75d9-997b-4b2b-a923-8ee9739fcbbc\") " Oct 05 21:16:09 crc kubenswrapper[4754]: I1005 21:16:09.284775 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c45e75d9-997b-4b2b-a923-8ee9739fcbbc-combined-ca-bundle\") pod \"c45e75d9-997b-4b2b-a923-8ee9739fcbbc\" (UID: \"c45e75d9-997b-4b2b-a923-8ee9739fcbbc\") " Oct 05 21:16:09 crc kubenswrapper[4754]: I1005 21:16:09.284866 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c45e75d9-997b-4b2b-a923-8ee9739fcbbc-logs\") pod \"c45e75d9-997b-4b2b-a923-8ee9739fcbbc\" (UID: \"c45e75d9-997b-4b2b-a923-8ee9739fcbbc\") " Oct 05 21:16:09 crc kubenswrapper[4754]: I1005 21:16:09.284965 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s45ll\" (UniqueName: \"kubernetes.io/projected/c45e75d9-997b-4b2b-a923-8ee9739fcbbc-kube-api-access-s45ll\") pod \"c45e75d9-997b-4b2b-a923-8ee9739fcbbc\" (UID: \"c45e75d9-997b-4b2b-a923-8ee9739fcbbc\") " Oct 05 21:16:09 crc kubenswrapper[4754]: I1005 21:16:09.285006 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c45e75d9-997b-4b2b-a923-8ee9739fcbbc-config-data\") pod \"c45e75d9-997b-4b2b-a923-8ee9739fcbbc\" (UID: \"c45e75d9-997b-4b2b-a923-8ee9739fcbbc\") " Oct 05 21:16:09 crc kubenswrapper[4754]: I1005 21:16:09.290139 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c45e75d9-997b-4b2b-a923-8ee9739fcbbc-logs" (OuterVolumeSpecName: "logs") pod "c45e75d9-997b-4b2b-a923-8ee9739fcbbc" (UID: "c45e75d9-997b-4b2b-a923-8ee9739fcbbc"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 21:16:09 crc kubenswrapper[4754]: I1005 21:16:09.294403 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c45e75d9-997b-4b2b-a923-8ee9739fcbbc-kube-api-access-s45ll" (OuterVolumeSpecName: "kube-api-access-s45ll") pod "c45e75d9-997b-4b2b-a923-8ee9739fcbbc" (UID: "c45e75d9-997b-4b2b-a923-8ee9739fcbbc"). InnerVolumeSpecName "kube-api-access-s45ll". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:16:09 crc kubenswrapper[4754]: I1005 21:16:09.333789 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c45e75d9-997b-4b2b-a923-8ee9739fcbbc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c45e75d9-997b-4b2b-a923-8ee9739fcbbc" (UID: "c45e75d9-997b-4b2b-a923-8ee9739fcbbc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:16:09 crc kubenswrapper[4754]: I1005 21:16:09.351672 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c45e75d9-997b-4b2b-a923-8ee9739fcbbc-config-data" (OuterVolumeSpecName: "config-data") pod "c45e75d9-997b-4b2b-a923-8ee9739fcbbc" (UID: "c45e75d9-997b-4b2b-a923-8ee9739fcbbc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:16:09 crc kubenswrapper[4754]: I1005 21:16:09.382365 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c45e75d9-997b-4b2b-a923-8ee9739fcbbc-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "c45e75d9-997b-4b2b-a923-8ee9739fcbbc" (UID: "c45e75d9-997b-4b2b-a923-8ee9739fcbbc"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:16:09 crc kubenswrapper[4754]: I1005 21:16:09.387447 4754 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c45e75d9-997b-4b2b-a923-8ee9739fcbbc-logs\") on node \"crc\" DevicePath \"\"" Oct 05 21:16:09 crc kubenswrapper[4754]: I1005 21:16:09.387485 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s45ll\" (UniqueName: \"kubernetes.io/projected/c45e75d9-997b-4b2b-a923-8ee9739fcbbc-kube-api-access-s45ll\") on node \"crc\" DevicePath \"\"" Oct 05 21:16:09 crc kubenswrapper[4754]: I1005 21:16:09.387561 4754 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c45e75d9-997b-4b2b-a923-8ee9739fcbbc-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 21:16:09 crc kubenswrapper[4754]: I1005 21:16:09.387574 4754 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/c45e75d9-997b-4b2b-a923-8ee9739fcbbc-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 05 21:16:09 crc kubenswrapper[4754]: I1005 21:16:09.387588 4754 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c45e75d9-997b-4b2b-a923-8ee9739fcbbc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 21:16:09 crc kubenswrapper[4754]: I1005 21:16:09.435136 4754 generic.go:334] "Generic (PLEG): container finished" podID="c45e75d9-997b-4b2b-a923-8ee9739fcbbc" containerID="40f7e9db9c4cfc1a4743fb8a28bcd85bf3946e7cfabbb175345526f616f3cd1d" exitCode=0 Oct 05 21:16:09 crc kubenswrapper[4754]: I1005 21:16:09.435214 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c45e75d9-997b-4b2b-a923-8ee9739fcbbc","Type":"ContainerDied","Data":"40f7e9db9c4cfc1a4743fb8a28bcd85bf3946e7cfabbb175345526f616f3cd1d"} Oct 05 21:16:09 crc kubenswrapper[4754]: I1005 21:16:09.435245 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c45e75d9-997b-4b2b-a923-8ee9739fcbbc","Type":"ContainerDied","Data":"54e5975bfd60a82b4331cc54db1b409d2c2bab2841856c7a404ebbc218129086"} Oct 05 21:16:09 crc kubenswrapper[4754]: I1005 21:16:09.435253 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 05 21:16:09 crc kubenswrapper[4754]: I1005 21:16:09.435262 4754 scope.go:117] "RemoveContainer" containerID="40f7e9db9c4cfc1a4743fb8a28bcd85bf3946e7cfabbb175345526f616f3cd1d" Oct 05 21:16:09 crc kubenswrapper[4754]: I1005 21:16:09.438629 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"3e33c8f1-8161-4919-98e0-971f46b33b05","Type":"ContainerStarted","Data":"3b5367d4af31eea274387a44ae135213f4a803e9ec77e9319e6409b1c3404749"} Oct 05 21:16:09 crc kubenswrapper[4754]: I1005 21:16:09.438675 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"3e33c8f1-8161-4919-98e0-971f46b33b05","Type":"ContainerStarted","Data":"297689c13c57a9e4bef75726a11184bfa6d25e6a7270e953e274a50994c7e063"} Oct 05 21:16:09 crc kubenswrapper[4754]: I1005 21:16:09.461670 4754 scope.go:117] "RemoveContainer" containerID="c6039287394c53d920bccc08c1a5fc75ad1e47d196696fb83bcc68a3e88c13b5" Oct 05 21:16:09 crc kubenswrapper[4754]: I1005 21:16:09.465737 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.465719578 podStartE2EDuration="2.465719578s" podCreationTimestamp="2025-10-05 21:16:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 21:16:09.457445255 +0000 UTC m=+1293.361563965" watchObservedRunningTime="2025-10-05 21:16:09.465719578 +0000 UTC m=+1293.369838288" Oct 05 21:16:09 crc kubenswrapper[4754]: I1005 21:16:09.479987 4754 scope.go:117] "RemoveContainer" containerID="40f7e9db9c4cfc1a4743fb8a28bcd85bf3946e7cfabbb175345526f616f3cd1d" Oct 05 21:16:09 crc kubenswrapper[4754]: E1005 21:16:09.480662 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"40f7e9db9c4cfc1a4743fb8a28bcd85bf3946e7cfabbb175345526f616f3cd1d\": container with ID starting with 40f7e9db9c4cfc1a4743fb8a28bcd85bf3946e7cfabbb175345526f616f3cd1d not found: ID does not exist" containerID="40f7e9db9c4cfc1a4743fb8a28bcd85bf3946e7cfabbb175345526f616f3cd1d" Oct 05 21:16:09 crc kubenswrapper[4754]: I1005 21:16:09.480709 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"40f7e9db9c4cfc1a4743fb8a28bcd85bf3946e7cfabbb175345526f616f3cd1d"} err="failed to get container status \"40f7e9db9c4cfc1a4743fb8a28bcd85bf3946e7cfabbb175345526f616f3cd1d\": rpc error: code = NotFound desc = could not find container \"40f7e9db9c4cfc1a4743fb8a28bcd85bf3946e7cfabbb175345526f616f3cd1d\": container with ID starting with 40f7e9db9c4cfc1a4743fb8a28bcd85bf3946e7cfabbb175345526f616f3cd1d not found: ID does not exist" Oct 05 21:16:09 crc kubenswrapper[4754]: I1005 21:16:09.480731 4754 scope.go:117] "RemoveContainer" containerID="c6039287394c53d920bccc08c1a5fc75ad1e47d196696fb83bcc68a3e88c13b5" Oct 05 21:16:09 crc kubenswrapper[4754]: E1005 21:16:09.481045 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c6039287394c53d920bccc08c1a5fc75ad1e47d196696fb83bcc68a3e88c13b5\": container with ID starting with c6039287394c53d920bccc08c1a5fc75ad1e47d196696fb83bcc68a3e88c13b5 not found: ID does not exist" containerID="c6039287394c53d920bccc08c1a5fc75ad1e47d196696fb83bcc68a3e88c13b5" Oct 05 21:16:09 crc kubenswrapper[4754]: I1005 21:16:09.481067 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c6039287394c53d920bccc08c1a5fc75ad1e47d196696fb83bcc68a3e88c13b5"} err="failed to get container status \"c6039287394c53d920bccc08c1a5fc75ad1e47d196696fb83bcc68a3e88c13b5\": rpc error: code = NotFound desc = could not find container \"c6039287394c53d920bccc08c1a5fc75ad1e47d196696fb83bcc68a3e88c13b5\": container with ID starting with c6039287394c53d920bccc08c1a5fc75ad1e47d196696fb83bcc68a3e88c13b5 not found: ID does not exist" Oct 05 21:16:09 crc kubenswrapper[4754]: I1005 21:16:09.484857 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 05 21:16:09 crc kubenswrapper[4754]: I1005 21:16:09.498186 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 05 21:16:09 crc kubenswrapper[4754]: I1005 21:16:09.519454 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 05 21:16:09 crc kubenswrapper[4754]: E1005 21:16:09.519858 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c45e75d9-997b-4b2b-a923-8ee9739fcbbc" containerName="nova-metadata-metadata" Oct 05 21:16:09 crc kubenswrapper[4754]: I1005 21:16:09.519877 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="c45e75d9-997b-4b2b-a923-8ee9739fcbbc" containerName="nova-metadata-metadata" Oct 05 21:16:09 crc kubenswrapper[4754]: E1005 21:16:09.519906 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c45e75d9-997b-4b2b-a923-8ee9739fcbbc" containerName="nova-metadata-log" Oct 05 21:16:09 crc kubenswrapper[4754]: I1005 21:16:09.519913 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="c45e75d9-997b-4b2b-a923-8ee9739fcbbc" containerName="nova-metadata-log" Oct 05 21:16:09 crc kubenswrapper[4754]: I1005 21:16:09.520081 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="c45e75d9-997b-4b2b-a923-8ee9739fcbbc" containerName="nova-metadata-log" Oct 05 21:16:09 crc kubenswrapper[4754]: I1005 21:16:09.520105 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="c45e75d9-997b-4b2b-a923-8ee9739fcbbc" containerName="nova-metadata-metadata" Oct 05 21:16:09 crc kubenswrapper[4754]: I1005 21:16:09.521093 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 05 21:16:09 crc kubenswrapper[4754]: I1005 21:16:09.523912 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 05 21:16:09 crc kubenswrapper[4754]: I1005 21:16:09.534918 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 05 21:16:09 crc kubenswrapper[4754]: I1005 21:16:09.536275 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 05 21:16:09 crc kubenswrapper[4754]: I1005 21:16:09.591273 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-74qc8\" (UniqueName: \"kubernetes.io/projected/bfc1abc2-68ba-4fc4-9dac-5e15ce2f52cd-kube-api-access-74qc8\") pod \"nova-metadata-0\" (UID: \"bfc1abc2-68ba-4fc4-9dac-5e15ce2f52cd\") " pod="openstack/nova-metadata-0" Oct 05 21:16:09 crc kubenswrapper[4754]: I1005 21:16:09.591350 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bfc1abc2-68ba-4fc4-9dac-5e15ce2f52cd-logs\") pod \"nova-metadata-0\" (UID: \"bfc1abc2-68ba-4fc4-9dac-5e15ce2f52cd\") " pod="openstack/nova-metadata-0" Oct 05 21:16:09 crc kubenswrapper[4754]: I1005 21:16:09.591373 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bfc1abc2-68ba-4fc4-9dac-5e15ce2f52cd-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"bfc1abc2-68ba-4fc4-9dac-5e15ce2f52cd\") " pod="openstack/nova-metadata-0" Oct 05 21:16:09 crc kubenswrapper[4754]: I1005 21:16:09.591411 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/bfc1abc2-68ba-4fc4-9dac-5e15ce2f52cd-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"bfc1abc2-68ba-4fc4-9dac-5e15ce2f52cd\") " pod="openstack/nova-metadata-0" Oct 05 21:16:09 crc kubenswrapper[4754]: I1005 21:16:09.591451 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bfc1abc2-68ba-4fc4-9dac-5e15ce2f52cd-config-data\") pod \"nova-metadata-0\" (UID: \"bfc1abc2-68ba-4fc4-9dac-5e15ce2f52cd\") " pod="openstack/nova-metadata-0" Oct 05 21:16:09 crc kubenswrapper[4754]: I1005 21:16:09.692925 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-74qc8\" (UniqueName: \"kubernetes.io/projected/bfc1abc2-68ba-4fc4-9dac-5e15ce2f52cd-kube-api-access-74qc8\") pod \"nova-metadata-0\" (UID: \"bfc1abc2-68ba-4fc4-9dac-5e15ce2f52cd\") " pod="openstack/nova-metadata-0" Oct 05 21:16:09 crc kubenswrapper[4754]: I1005 21:16:09.693011 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bfc1abc2-68ba-4fc4-9dac-5e15ce2f52cd-logs\") pod \"nova-metadata-0\" (UID: \"bfc1abc2-68ba-4fc4-9dac-5e15ce2f52cd\") " pod="openstack/nova-metadata-0" Oct 05 21:16:09 crc kubenswrapper[4754]: I1005 21:16:09.693036 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bfc1abc2-68ba-4fc4-9dac-5e15ce2f52cd-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"bfc1abc2-68ba-4fc4-9dac-5e15ce2f52cd\") " pod="openstack/nova-metadata-0" Oct 05 21:16:09 crc kubenswrapper[4754]: I1005 21:16:09.693071 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/bfc1abc2-68ba-4fc4-9dac-5e15ce2f52cd-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"bfc1abc2-68ba-4fc4-9dac-5e15ce2f52cd\") " pod="openstack/nova-metadata-0" Oct 05 21:16:09 crc kubenswrapper[4754]: I1005 21:16:09.693126 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bfc1abc2-68ba-4fc4-9dac-5e15ce2f52cd-config-data\") pod \"nova-metadata-0\" (UID: \"bfc1abc2-68ba-4fc4-9dac-5e15ce2f52cd\") " pod="openstack/nova-metadata-0" Oct 05 21:16:09 crc kubenswrapper[4754]: I1005 21:16:09.693517 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bfc1abc2-68ba-4fc4-9dac-5e15ce2f52cd-logs\") pod \"nova-metadata-0\" (UID: \"bfc1abc2-68ba-4fc4-9dac-5e15ce2f52cd\") " pod="openstack/nova-metadata-0" Oct 05 21:16:09 crc kubenswrapper[4754]: I1005 21:16:09.697264 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bfc1abc2-68ba-4fc4-9dac-5e15ce2f52cd-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"bfc1abc2-68ba-4fc4-9dac-5e15ce2f52cd\") " pod="openstack/nova-metadata-0" Oct 05 21:16:09 crc kubenswrapper[4754]: I1005 21:16:09.697316 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bfc1abc2-68ba-4fc4-9dac-5e15ce2f52cd-config-data\") pod \"nova-metadata-0\" (UID: \"bfc1abc2-68ba-4fc4-9dac-5e15ce2f52cd\") " pod="openstack/nova-metadata-0" Oct 05 21:16:09 crc kubenswrapper[4754]: I1005 21:16:09.697697 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/bfc1abc2-68ba-4fc4-9dac-5e15ce2f52cd-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"bfc1abc2-68ba-4fc4-9dac-5e15ce2f52cd\") " pod="openstack/nova-metadata-0" Oct 05 21:16:09 crc kubenswrapper[4754]: I1005 21:16:09.711165 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-74qc8\" (UniqueName: \"kubernetes.io/projected/bfc1abc2-68ba-4fc4-9dac-5e15ce2f52cd-kube-api-access-74qc8\") pod \"nova-metadata-0\" (UID: \"bfc1abc2-68ba-4fc4-9dac-5e15ce2f52cd\") " pod="openstack/nova-metadata-0" Oct 05 21:16:09 crc kubenswrapper[4754]: I1005 21:16:09.836215 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 05 21:16:10 crc kubenswrapper[4754]: I1005 21:16:10.345555 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 05 21:16:10 crc kubenswrapper[4754]: I1005 21:16:10.447432 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"bfc1abc2-68ba-4fc4-9dac-5e15ce2f52cd","Type":"ContainerStarted","Data":"5eb0ba13e57ccbb6ff3c1a78eb4be0348ac7d22852980aeac83eca41d55ff81f"} Oct 05 21:16:10 crc kubenswrapper[4754]: I1005 21:16:10.854018 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c45e75d9-997b-4b2b-a923-8ee9739fcbbc" path="/var/lib/kubelet/pods/c45e75d9-997b-4b2b-a923-8ee9739fcbbc/volumes" Oct 05 21:16:11 crc kubenswrapper[4754]: I1005 21:16:11.477427 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"bfc1abc2-68ba-4fc4-9dac-5e15ce2f52cd","Type":"ContainerStarted","Data":"8447b0fad78eca59cae65e3b6b1374b66d2a0341e0bbadfd88166465f298939c"} Oct 05 21:16:11 crc kubenswrapper[4754]: I1005 21:16:11.478010 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"bfc1abc2-68ba-4fc4-9dac-5e15ce2f52cd","Type":"ContainerStarted","Data":"ad793fe1bcaca0bf8cdbd6e6c26dfd8a5f68a3d7dce75622d6070a74cefff096"} Oct 05 21:16:11 crc kubenswrapper[4754]: I1005 21:16:11.481622 4754 generic.go:334] "Generic (PLEG): container finished" podID="27cd6be5-f5eb-4158-bdd1-37f4df3d530d" containerID="48ce6ae318ddb68e70258170c00936a21c53d9535eb9576fac3f4ad34120b991" exitCode=137 Oct 05 21:16:11 crc kubenswrapper[4754]: I1005 21:16:11.481683 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-68c6b67864-9msm5" event={"ID":"27cd6be5-f5eb-4158-bdd1-37f4df3d530d","Type":"ContainerDied","Data":"48ce6ae318ddb68e70258170c00936a21c53d9535eb9576fac3f4ad34120b991"} Oct 05 21:16:11 crc kubenswrapper[4754]: I1005 21:16:11.531111 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.531089058 podStartE2EDuration="2.531089058s" podCreationTimestamp="2025-10-05 21:16:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 21:16:11.525799749 +0000 UTC m=+1295.429918479" watchObservedRunningTime="2025-10-05 21:16:11.531089058 +0000 UTC m=+1295.435207768" Oct 05 21:16:11 crc kubenswrapper[4754]: I1005 21:16:11.612906 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-68c6b67864-9msm5" Oct 05 21:16:11 crc kubenswrapper[4754]: I1005 21:16:11.743220 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27cd6be5-f5eb-4158-bdd1-37f4df3d530d-combined-ca-bundle\") pod \"27cd6be5-f5eb-4158-bdd1-37f4df3d530d\" (UID: \"27cd6be5-f5eb-4158-bdd1-37f4df3d530d\") " Oct 05 21:16:11 crc kubenswrapper[4754]: I1005 21:16:11.743322 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/27cd6be5-f5eb-4158-bdd1-37f4df3d530d-horizon-tls-certs\") pod \"27cd6be5-f5eb-4158-bdd1-37f4df3d530d\" (UID: \"27cd6be5-f5eb-4158-bdd1-37f4df3d530d\") " Oct 05 21:16:11 crc kubenswrapper[4754]: I1005 21:16:11.743656 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/27cd6be5-f5eb-4158-bdd1-37f4df3d530d-config-data\") pod \"27cd6be5-f5eb-4158-bdd1-37f4df3d530d\" (UID: \"27cd6be5-f5eb-4158-bdd1-37f4df3d530d\") " Oct 05 21:16:11 crc kubenswrapper[4754]: I1005 21:16:11.743730 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/27cd6be5-f5eb-4158-bdd1-37f4df3d530d-logs\") pod \"27cd6be5-f5eb-4158-bdd1-37f4df3d530d\" (UID: \"27cd6be5-f5eb-4158-bdd1-37f4df3d530d\") " Oct 05 21:16:11 crc kubenswrapper[4754]: I1005 21:16:11.744516 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tmrsq\" (UniqueName: \"kubernetes.io/projected/27cd6be5-f5eb-4158-bdd1-37f4df3d530d-kube-api-access-tmrsq\") pod \"27cd6be5-f5eb-4158-bdd1-37f4df3d530d\" (UID: \"27cd6be5-f5eb-4158-bdd1-37f4df3d530d\") " Oct 05 21:16:11 crc kubenswrapper[4754]: I1005 21:16:11.744560 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/27cd6be5-f5eb-4158-bdd1-37f4df3d530d-scripts\") pod \"27cd6be5-f5eb-4158-bdd1-37f4df3d530d\" (UID: \"27cd6be5-f5eb-4158-bdd1-37f4df3d530d\") " Oct 05 21:16:11 crc kubenswrapper[4754]: I1005 21:16:11.744615 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/27cd6be5-f5eb-4158-bdd1-37f4df3d530d-horizon-secret-key\") pod \"27cd6be5-f5eb-4158-bdd1-37f4df3d530d\" (UID: \"27cd6be5-f5eb-4158-bdd1-37f4df3d530d\") " Oct 05 21:16:11 crc kubenswrapper[4754]: I1005 21:16:11.745138 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/27cd6be5-f5eb-4158-bdd1-37f4df3d530d-logs" (OuterVolumeSpecName: "logs") pod "27cd6be5-f5eb-4158-bdd1-37f4df3d530d" (UID: "27cd6be5-f5eb-4158-bdd1-37f4df3d530d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 21:16:11 crc kubenswrapper[4754]: I1005 21:16:11.745486 4754 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/27cd6be5-f5eb-4158-bdd1-37f4df3d530d-logs\") on node \"crc\" DevicePath \"\"" Oct 05 21:16:11 crc kubenswrapper[4754]: I1005 21:16:11.749301 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/27cd6be5-f5eb-4158-bdd1-37f4df3d530d-kube-api-access-tmrsq" (OuterVolumeSpecName: "kube-api-access-tmrsq") pod "27cd6be5-f5eb-4158-bdd1-37f4df3d530d" (UID: "27cd6be5-f5eb-4158-bdd1-37f4df3d530d"). InnerVolumeSpecName "kube-api-access-tmrsq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:16:11 crc kubenswrapper[4754]: I1005 21:16:11.772282 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27cd6be5-f5eb-4158-bdd1-37f4df3d530d-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "27cd6be5-f5eb-4158-bdd1-37f4df3d530d" (UID: "27cd6be5-f5eb-4158-bdd1-37f4df3d530d"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:16:11 crc kubenswrapper[4754]: I1005 21:16:11.774155 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/27cd6be5-f5eb-4158-bdd1-37f4df3d530d-config-data" (OuterVolumeSpecName: "config-data") pod "27cd6be5-f5eb-4158-bdd1-37f4df3d530d" (UID: "27cd6be5-f5eb-4158-bdd1-37f4df3d530d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 21:16:11 crc kubenswrapper[4754]: I1005 21:16:11.776810 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/27cd6be5-f5eb-4158-bdd1-37f4df3d530d-scripts" (OuterVolumeSpecName: "scripts") pod "27cd6be5-f5eb-4158-bdd1-37f4df3d530d" (UID: "27cd6be5-f5eb-4158-bdd1-37f4df3d530d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 21:16:11 crc kubenswrapper[4754]: I1005 21:16:11.784127 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27cd6be5-f5eb-4158-bdd1-37f4df3d530d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "27cd6be5-f5eb-4158-bdd1-37f4df3d530d" (UID: "27cd6be5-f5eb-4158-bdd1-37f4df3d530d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:16:11 crc kubenswrapper[4754]: I1005 21:16:11.806520 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27cd6be5-f5eb-4158-bdd1-37f4df3d530d-horizon-tls-certs" (OuterVolumeSpecName: "horizon-tls-certs") pod "27cd6be5-f5eb-4158-bdd1-37f4df3d530d" (UID: "27cd6be5-f5eb-4158-bdd1-37f4df3d530d"). InnerVolumeSpecName "horizon-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:16:11 crc kubenswrapper[4754]: I1005 21:16:11.847372 4754 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/27cd6be5-f5eb-4158-bdd1-37f4df3d530d-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 21:16:11 crc kubenswrapper[4754]: I1005 21:16:11.847732 4754 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/27cd6be5-f5eb-4158-bdd1-37f4df3d530d-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 05 21:16:11 crc kubenswrapper[4754]: I1005 21:16:11.847744 4754 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27cd6be5-f5eb-4158-bdd1-37f4df3d530d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 21:16:11 crc kubenswrapper[4754]: I1005 21:16:11.847754 4754 reconciler_common.go:293] "Volume detached for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/27cd6be5-f5eb-4158-bdd1-37f4df3d530d-horizon-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 05 21:16:11 crc kubenswrapper[4754]: I1005 21:16:11.847762 4754 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/27cd6be5-f5eb-4158-bdd1-37f4df3d530d-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 21:16:11 crc kubenswrapper[4754]: I1005 21:16:11.847772 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tmrsq\" (UniqueName: \"kubernetes.io/projected/27cd6be5-f5eb-4158-bdd1-37f4df3d530d-kube-api-access-tmrsq\") on node \"crc\" DevicePath \"\"" Oct 05 21:16:12 crc kubenswrapper[4754]: I1005 21:16:12.495668 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-68c6b67864-9msm5" Oct 05 21:16:12 crc kubenswrapper[4754]: I1005 21:16:12.505206 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-68c6b67864-9msm5" event={"ID":"27cd6be5-f5eb-4158-bdd1-37f4df3d530d","Type":"ContainerDied","Data":"de518e9a66f985fdfb6488044bfa833556110cf2d73cb9ff476ce82862781eae"} Oct 05 21:16:12 crc kubenswrapper[4754]: I1005 21:16:12.505700 4754 scope.go:117] "RemoveContainer" containerID="e8c8cc4de7167b842bac5c879a08a541081eaf3a777d19676eea7ed48c8eae17" Oct 05 21:16:12 crc kubenswrapper[4754]: I1005 21:16:12.541557 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-68c6b67864-9msm5"] Oct 05 21:16:12 crc kubenswrapper[4754]: I1005 21:16:12.548147 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-68c6b67864-9msm5"] Oct 05 21:16:12 crc kubenswrapper[4754]: I1005 21:16:12.744306 4754 scope.go:117] "RemoveContainer" containerID="48ce6ae318ddb68e70258170c00936a21c53d9535eb9576fac3f4ad34120b991" Oct 05 21:16:12 crc kubenswrapper[4754]: I1005 21:16:12.802140 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 05 21:16:12 crc kubenswrapper[4754]: I1005 21:16:12.858581 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="27cd6be5-f5eb-4158-bdd1-37f4df3d530d" path="/var/lib/kubelet/pods/27cd6be5-f5eb-4158-bdd1-37f4df3d530d/volumes" Oct 05 21:16:14 crc kubenswrapper[4754]: I1005 21:16:14.854616 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 05 21:16:14 crc kubenswrapper[4754]: I1005 21:16:14.855293 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 05 21:16:17 crc kubenswrapper[4754]: I1005 21:16:17.066556 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 05 21:16:17 crc kubenswrapper[4754]: I1005 21:16:17.067005 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 05 21:16:17 crc kubenswrapper[4754]: I1005 21:16:17.801653 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 05 21:16:17 crc kubenswrapper[4754]: I1005 21:16:17.852911 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 05 21:16:18 crc kubenswrapper[4754]: I1005 21:16:18.091753 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="527adc21-0816-4fba-b7a1-22b3db2e1d73" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.204:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 05 21:16:18 crc kubenswrapper[4754]: I1005 21:16:18.091797 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="527adc21-0816-4fba-b7a1-22b3db2e1d73" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.204:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 05 21:16:18 crc kubenswrapper[4754]: I1005 21:16:18.623638 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 05 21:16:19 crc kubenswrapper[4754]: I1005 21:16:19.836702 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 05 21:16:19 crc kubenswrapper[4754]: I1005 21:16:19.837116 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 05 21:16:20 crc kubenswrapper[4754]: I1005 21:16:20.855618 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="bfc1abc2-68ba-4fc4-9dac-5e15ce2f52cd" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.206:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 05 21:16:20 crc kubenswrapper[4754]: I1005 21:16:20.855657 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="bfc1abc2-68ba-4fc4-9dac-5e15ce2f52cd" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.206:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 05 21:16:24 crc kubenswrapper[4754]: I1005 21:16:24.827833 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 05 21:16:27 crc kubenswrapper[4754]: I1005 21:16:27.076961 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 05 21:16:27 crc kubenswrapper[4754]: I1005 21:16:27.078914 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 05 21:16:27 crc kubenswrapper[4754]: I1005 21:16:27.079095 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 05 21:16:27 crc kubenswrapper[4754]: I1005 21:16:27.097998 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 05 21:16:27 crc kubenswrapper[4754]: I1005 21:16:27.701522 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 05 21:16:27 crc kubenswrapper[4754]: I1005 21:16:27.708650 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 05 21:16:29 crc kubenswrapper[4754]: I1005 21:16:29.843838 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 05 21:16:29 crc kubenswrapper[4754]: I1005 21:16:29.847832 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 05 21:16:29 crc kubenswrapper[4754]: I1005 21:16:29.853556 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 05 21:16:30 crc kubenswrapper[4754]: I1005 21:16:30.743257 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 05 21:16:35 crc kubenswrapper[4754]: I1005 21:16:35.245238 4754 patch_prober.go:28] interesting pod/machine-config-daemon-b2h9k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 21:16:35 crc kubenswrapper[4754]: I1005 21:16:35.246158 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 21:16:39 crc kubenswrapper[4754]: I1005 21:16:39.058563 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 05 21:16:40 crc kubenswrapper[4754]: I1005 21:16:40.264025 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 05 21:16:44 crc kubenswrapper[4754]: I1005 21:16:44.465799 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="39aff8a0-2905-4ad3-ad2d-5bf6b5de3858" containerName="rabbitmq" containerID="cri-o://e4012712472c94cd11327163abb6c6f5c55cff5c6625d3242d418e93290fa5bf" gracePeriod=604795 Oct 05 21:16:45 crc kubenswrapper[4754]: I1005 21:16:45.157208 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="1d96e280-8c26-41ea-ab29-64ce5b9c6a2c" containerName="rabbitmq" containerID="cri-o://ea5291016f4053d149506d9f4e1a17c9d6281d48b900045f7a5bc743cc386a39" gracePeriod=604796 Oct 05 21:16:49 crc kubenswrapper[4754]: I1005 21:16:49.825026 4754 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="39aff8a0-2905-4ad3-ad2d-5bf6b5de3858" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.100:5671: connect: connection refused" Oct 05 21:16:50 crc kubenswrapper[4754]: I1005 21:16:50.374100 4754 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="1d96e280-8c26-41ea-ab29-64ce5b9c6a2c" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.101:5671: connect: connection refused" Oct 05 21:16:51 crc kubenswrapper[4754]: I1005 21:16:51.009888 4754 generic.go:334] "Generic (PLEG): container finished" podID="39aff8a0-2905-4ad3-ad2d-5bf6b5de3858" containerID="e4012712472c94cd11327163abb6c6f5c55cff5c6625d3242d418e93290fa5bf" exitCode=0 Oct 05 21:16:51 crc kubenswrapper[4754]: I1005 21:16:51.009938 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"39aff8a0-2905-4ad3-ad2d-5bf6b5de3858","Type":"ContainerDied","Data":"e4012712472c94cd11327163abb6c6f5c55cff5c6625d3242d418e93290fa5bf"} Oct 05 21:16:51 crc kubenswrapper[4754]: I1005 21:16:51.090165 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 05 21:16:51 crc kubenswrapper[4754]: I1005 21:16:51.210044 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/39aff8a0-2905-4ad3-ad2d-5bf6b5de3858-erlang-cookie-secret\") pod \"39aff8a0-2905-4ad3-ad2d-5bf6b5de3858\" (UID: \"39aff8a0-2905-4ad3-ad2d-5bf6b5de3858\") " Oct 05 21:16:51 crc kubenswrapper[4754]: I1005 21:16:51.210190 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/39aff8a0-2905-4ad3-ad2d-5bf6b5de3858-rabbitmq-erlang-cookie\") pod \"39aff8a0-2905-4ad3-ad2d-5bf6b5de3858\" (UID: \"39aff8a0-2905-4ad3-ad2d-5bf6b5de3858\") " Oct 05 21:16:51 crc kubenswrapper[4754]: I1005 21:16:51.210227 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/39aff8a0-2905-4ad3-ad2d-5bf6b5de3858-config-data\") pod \"39aff8a0-2905-4ad3-ad2d-5bf6b5de3858\" (UID: \"39aff8a0-2905-4ad3-ad2d-5bf6b5de3858\") " Oct 05 21:16:51 crc kubenswrapper[4754]: I1005 21:16:51.210252 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/39aff8a0-2905-4ad3-ad2d-5bf6b5de3858-rabbitmq-confd\") pod \"39aff8a0-2905-4ad3-ad2d-5bf6b5de3858\" (UID: \"39aff8a0-2905-4ad3-ad2d-5bf6b5de3858\") " Oct 05 21:16:51 crc kubenswrapper[4754]: I1005 21:16:51.210270 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/39aff8a0-2905-4ad3-ad2d-5bf6b5de3858-rabbitmq-plugins\") pod \"39aff8a0-2905-4ad3-ad2d-5bf6b5de3858\" (UID: \"39aff8a0-2905-4ad3-ad2d-5bf6b5de3858\") " Oct 05 21:16:51 crc kubenswrapper[4754]: I1005 21:16:51.210297 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/39aff8a0-2905-4ad3-ad2d-5bf6b5de3858-plugins-conf\") pod \"39aff8a0-2905-4ad3-ad2d-5bf6b5de3858\" (UID: \"39aff8a0-2905-4ad3-ad2d-5bf6b5de3858\") " Oct 05 21:16:51 crc kubenswrapper[4754]: I1005 21:16:51.210325 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/39aff8a0-2905-4ad3-ad2d-5bf6b5de3858-pod-info\") pod \"39aff8a0-2905-4ad3-ad2d-5bf6b5de3858\" (UID: \"39aff8a0-2905-4ad3-ad2d-5bf6b5de3858\") " Oct 05 21:16:51 crc kubenswrapper[4754]: I1005 21:16:51.210359 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"39aff8a0-2905-4ad3-ad2d-5bf6b5de3858\" (UID: \"39aff8a0-2905-4ad3-ad2d-5bf6b5de3858\") " Oct 05 21:16:51 crc kubenswrapper[4754]: I1005 21:16:51.210417 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/39aff8a0-2905-4ad3-ad2d-5bf6b5de3858-server-conf\") pod \"39aff8a0-2905-4ad3-ad2d-5bf6b5de3858\" (UID: \"39aff8a0-2905-4ad3-ad2d-5bf6b5de3858\") " Oct 05 21:16:51 crc kubenswrapper[4754]: I1005 21:16:51.210444 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkfvq\" (UniqueName: \"kubernetes.io/projected/39aff8a0-2905-4ad3-ad2d-5bf6b5de3858-kube-api-access-jkfvq\") pod \"39aff8a0-2905-4ad3-ad2d-5bf6b5de3858\" (UID: \"39aff8a0-2905-4ad3-ad2d-5bf6b5de3858\") " Oct 05 21:16:51 crc kubenswrapper[4754]: I1005 21:16:51.210508 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/39aff8a0-2905-4ad3-ad2d-5bf6b5de3858-rabbitmq-tls\") pod \"39aff8a0-2905-4ad3-ad2d-5bf6b5de3858\" (UID: \"39aff8a0-2905-4ad3-ad2d-5bf6b5de3858\") " Oct 05 21:16:51 crc kubenswrapper[4754]: I1005 21:16:51.212811 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/39aff8a0-2905-4ad3-ad2d-5bf6b5de3858-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "39aff8a0-2905-4ad3-ad2d-5bf6b5de3858" (UID: "39aff8a0-2905-4ad3-ad2d-5bf6b5de3858"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 21:16:51 crc kubenswrapper[4754]: I1005 21:16:51.213641 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/39aff8a0-2905-4ad3-ad2d-5bf6b5de3858-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "39aff8a0-2905-4ad3-ad2d-5bf6b5de3858" (UID: "39aff8a0-2905-4ad3-ad2d-5bf6b5de3858"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 21:16:51 crc kubenswrapper[4754]: I1005 21:16:51.214634 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/39aff8a0-2905-4ad3-ad2d-5bf6b5de3858-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "39aff8a0-2905-4ad3-ad2d-5bf6b5de3858" (UID: "39aff8a0-2905-4ad3-ad2d-5bf6b5de3858"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 21:16:51 crc kubenswrapper[4754]: I1005 21:16:51.223812 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/39aff8a0-2905-4ad3-ad2d-5bf6b5de3858-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "39aff8a0-2905-4ad3-ad2d-5bf6b5de3858" (UID: "39aff8a0-2905-4ad3-ad2d-5bf6b5de3858"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:16:51 crc kubenswrapper[4754]: I1005 21:16:51.236657 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/39aff8a0-2905-4ad3-ad2d-5bf6b5de3858-pod-info" (OuterVolumeSpecName: "pod-info") pod "39aff8a0-2905-4ad3-ad2d-5bf6b5de3858" (UID: "39aff8a0-2905-4ad3-ad2d-5bf6b5de3858"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 05 21:16:51 crc kubenswrapper[4754]: I1005 21:16:51.256177 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/39aff8a0-2905-4ad3-ad2d-5bf6b5de3858-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "39aff8a0-2905-4ad3-ad2d-5bf6b5de3858" (UID: "39aff8a0-2905-4ad3-ad2d-5bf6b5de3858"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:16:51 crc kubenswrapper[4754]: I1005 21:16:51.265202 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/39aff8a0-2905-4ad3-ad2d-5bf6b5de3858-config-data" (OuterVolumeSpecName: "config-data") pod "39aff8a0-2905-4ad3-ad2d-5bf6b5de3858" (UID: "39aff8a0-2905-4ad3-ad2d-5bf6b5de3858"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 21:16:51 crc kubenswrapper[4754]: I1005 21:16:51.266396 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/39aff8a0-2905-4ad3-ad2d-5bf6b5de3858-kube-api-access-jkfvq" (OuterVolumeSpecName: "kube-api-access-jkfvq") pod "39aff8a0-2905-4ad3-ad2d-5bf6b5de3858" (UID: "39aff8a0-2905-4ad3-ad2d-5bf6b5de3858"). InnerVolumeSpecName "kube-api-access-jkfvq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:16:51 crc kubenswrapper[4754]: I1005 21:16:51.277948 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage11-crc" (OuterVolumeSpecName: "persistence") pod "39aff8a0-2905-4ad3-ad2d-5bf6b5de3858" (UID: "39aff8a0-2905-4ad3-ad2d-5bf6b5de3858"). InnerVolumeSpecName "local-storage11-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 05 21:16:51 crc kubenswrapper[4754]: I1005 21:16:51.313273 4754 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/39aff8a0-2905-4ad3-ad2d-5bf6b5de3858-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Oct 05 21:16:51 crc kubenswrapper[4754]: I1005 21:16:51.313302 4754 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/39aff8a0-2905-4ad3-ad2d-5bf6b5de3858-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Oct 05 21:16:51 crc kubenswrapper[4754]: I1005 21:16:51.313313 4754 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/39aff8a0-2905-4ad3-ad2d-5bf6b5de3858-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 21:16:51 crc kubenswrapper[4754]: I1005 21:16:51.313342 4754 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/39aff8a0-2905-4ad3-ad2d-5bf6b5de3858-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Oct 05 21:16:51 crc kubenswrapper[4754]: I1005 21:16:51.313352 4754 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/39aff8a0-2905-4ad3-ad2d-5bf6b5de3858-plugins-conf\") on node \"crc\" DevicePath \"\"" Oct 05 21:16:51 crc kubenswrapper[4754]: I1005 21:16:51.313361 4754 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/39aff8a0-2905-4ad3-ad2d-5bf6b5de3858-pod-info\") on node \"crc\" DevicePath \"\"" Oct 05 21:16:51 crc kubenswrapper[4754]: I1005 21:16:51.313393 4754 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" " Oct 05 21:16:51 crc kubenswrapper[4754]: I1005 21:16:51.313402 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkfvq\" (UniqueName: \"kubernetes.io/projected/39aff8a0-2905-4ad3-ad2d-5bf6b5de3858-kube-api-access-jkfvq\") on node \"crc\" DevicePath \"\"" Oct 05 21:16:51 crc kubenswrapper[4754]: I1005 21:16:51.313428 4754 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/39aff8a0-2905-4ad3-ad2d-5bf6b5de3858-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Oct 05 21:16:51 crc kubenswrapper[4754]: I1005 21:16:51.319417 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/39aff8a0-2905-4ad3-ad2d-5bf6b5de3858-server-conf" (OuterVolumeSpecName: "server-conf") pod "39aff8a0-2905-4ad3-ad2d-5bf6b5de3858" (UID: "39aff8a0-2905-4ad3-ad2d-5bf6b5de3858"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 21:16:51 crc kubenswrapper[4754]: I1005 21:16:51.335152 4754 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage11-crc" (UniqueName: "kubernetes.io/local-volume/local-storage11-crc") on node "crc" Oct 05 21:16:51 crc kubenswrapper[4754]: I1005 21:16:51.389119 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/39aff8a0-2905-4ad3-ad2d-5bf6b5de3858-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "39aff8a0-2905-4ad3-ad2d-5bf6b5de3858" (UID: "39aff8a0-2905-4ad3-ad2d-5bf6b5de3858"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:16:51 crc kubenswrapper[4754]: I1005 21:16:51.415232 4754 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/39aff8a0-2905-4ad3-ad2d-5bf6b5de3858-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Oct 05 21:16:51 crc kubenswrapper[4754]: I1005 21:16:51.415300 4754 reconciler_common.go:293] "Volume detached for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" DevicePath \"\"" Oct 05 21:16:51 crc kubenswrapper[4754]: I1005 21:16:51.415313 4754 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/39aff8a0-2905-4ad3-ad2d-5bf6b5de3858-server-conf\") on node \"crc\" DevicePath \"\"" Oct 05 21:16:51 crc kubenswrapper[4754]: I1005 21:16:51.610896 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 05 21:16:51 crc kubenswrapper[4754]: I1005 21:16:51.721442 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/1d96e280-8c26-41ea-ab29-64ce5b9c6a2c-rabbitmq-tls\") pod \"1d96e280-8c26-41ea-ab29-64ce5b9c6a2c\" (UID: \"1d96e280-8c26-41ea-ab29-64ce5b9c6a2c\") " Oct 05 21:16:51 crc kubenswrapper[4754]: I1005 21:16:51.721527 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"1d96e280-8c26-41ea-ab29-64ce5b9c6a2c\" (UID: \"1d96e280-8c26-41ea-ab29-64ce5b9c6a2c\") " Oct 05 21:16:51 crc kubenswrapper[4754]: I1005 21:16:51.721556 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/1d96e280-8c26-41ea-ab29-64ce5b9c6a2c-erlang-cookie-secret\") pod \"1d96e280-8c26-41ea-ab29-64ce5b9c6a2c\" (UID: \"1d96e280-8c26-41ea-ab29-64ce5b9c6a2c\") " Oct 05 21:16:51 crc kubenswrapper[4754]: I1005 21:16:51.721742 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/1d96e280-8c26-41ea-ab29-64ce5b9c6a2c-rabbitmq-plugins\") pod \"1d96e280-8c26-41ea-ab29-64ce5b9c6a2c\" (UID: \"1d96e280-8c26-41ea-ab29-64ce5b9c6a2c\") " Oct 05 21:16:51 crc kubenswrapper[4754]: I1005 21:16:51.721777 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/1d96e280-8c26-41ea-ab29-64ce5b9c6a2c-server-conf\") pod \"1d96e280-8c26-41ea-ab29-64ce5b9c6a2c\" (UID: \"1d96e280-8c26-41ea-ab29-64ce5b9c6a2c\") " Oct 05 21:16:51 crc kubenswrapper[4754]: I1005 21:16:51.721817 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ltd4\" (UniqueName: \"kubernetes.io/projected/1d96e280-8c26-41ea-ab29-64ce5b9c6a2c-kube-api-access-6ltd4\") pod \"1d96e280-8c26-41ea-ab29-64ce5b9c6a2c\" (UID: \"1d96e280-8c26-41ea-ab29-64ce5b9c6a2c\") " Oct 05 21:16:51 crc kubenswrapper[4754]: I1005 21:16:51.721849 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/1d96e280-8c26-41ea-ab29-64ce5b9c6a2c-pod-info\") pod \"1d96e280-8c26-41ea-ab29-64ce5b9c6a2c\" (UID: \"1d96e280-8c26-41ea-ab29-64ce5b9c6a2c\") " Oct 05 21:16:51 crc kubenswrapper[4754]: I1005 21:16:51.721899 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/1d96e280-8c26-41ea-ab29-64ce5b9c6a2c-rabbitmq-erlang-cookie\") pod \"1d96e280-8c26-41ea-ab29-64ce5b9c6a2c\" (UID: \"1d96e280-8c26-41ea-ab29-64ce5b9c6a2c\") " Oct 05 21:16:51 crc kubenswrapper[4754]: I1005 21:16:51.721924 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/1d96e280-8c26-41ea-ab29-64ce5b9c6a2c-plugins-conf\") pod \"1d96e280-8c26-41ea-ab29-64ce5b9c6a2c\" (UID: \"1d96e280-8c26-41ea-ab29-64ce5b9c6a2c\") " Oct 05 21:16:51 crc kubenswrapper[4754]: I1005 21:16:51.721941 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1d96e280-8c26-41ea-ab29-64ce5b9c6a2c-config-data\") pod \"1d96e280-8c26-41ea-ab29-64ce5b9c6a2c\" (UID: \"1d96e280-8c26-41ea-ab29-64ce5b9c6a2c\") " Oct 05 21:16:51 crc kubenswrapper[4754]: I1005 21:16:51.721982 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/1d96e280-8c26-41ea-ab29-64ce5b9c6a2c-rabbitmq-confd\") pod \"1d96e280-8c26-41ea-ab29-64ce5b9c6a2c\" (UID: \"1d96e280-8c26-41ea-ab29-64ce5b9c6a2c\") " Oct 05 21:16:51 crc kubenswrapper[4754]: I1005 21:16:51.724095 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d96e280-8c26-41ea-ab29-64ce5b9c6a2c-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "1d96e280-8c26-41ea-ab29-64ce5b9c6a2c" (UID: "1d96e280-8c26-41ea-ab29-64ce5b9c6a2c"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 21:16:51 crc kubenswrapper[4754]: I1005 21:16:51.724222 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1d96e280-8c26-41ea-ab29-64ce5b9c6a2c-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "1d96e280-8c26-41ea-ab29-64ce5b9c6a2c" (UID: "1d96e280-8c26-41ea-ab29-64ce5b9c6a2c"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 21:16:51 crc kubenswrapper[4754]: I1005 21:16:51.726831 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d96e280-8c26-41ea-ab29-64ce5b9c6a2c-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "1d96e280-8c26-41ea-ab29-64ce5b9c6a2c" (UID: "1d96e280-8c26-41ea-ab29-64ce5b9c6a2c"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 21:16:51 crc kubenswrapper[4754]: I1005 21:16:51.728596 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/1d96e280-8c26-41ea-ab29-64ce5b9c6a2c-pod-info" (OuterVolumeSpecName: "pod-info") pod "1d96e280-8c26-41ea-ab29-64ce5b9c6a2c" (UID: "1d96e280-8c26-41ea-ab29-64ce5b9c6a2c"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 05 21:16:51 crc kubenswrapper[4754]: I1005 21:16:51.733986 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d96e280-8c26-41ea-ab29-64ce5b9c6a2c-kube-api-access-6ltd4" (OuterVolumeSpecName: "kube-api-access-6ltd4") pod "1d96e280-8c26-41ea-ab29-64ce5b9c6a2c" (UID: "1d96e280-8c26-41ea-ab29-64ce5b9c6a2c"). InnerVolumeSpecName "kube-api-access-6ltd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:16:51 crc kubenswrapper[4754]: I1005 21:16:51.734354 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "persistence") pod "1d96e280-8c26-41ea-ab29-64ce5b9c6a2c" (UID: "1d96e280-8c26-41ea-ab29-64ce5b9c6a2c"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 05 21:16:51 crc kubenswrapper[4754]: I1005 21:16:51.742514 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d96e280-8c26-41ea-ab29-64ce5b9c6a2c-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "1d96e280-8c26-41ea-ab29-64ce5b9c6a2c" (UID: "1d96e280-8c26-41ea-ab29-64ce5b9c6a2c"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:16:51 crc kubenswrapper[4754]: I1005 21:16:51.742686 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d96e280-8c26-41ea-ab29-64ce5b9c6a2c-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "1d96e280-8c26-41ea-ab29-64ce5b9c6a2c" (UID: "1d96e280-8c26-41ea-ab29-64ce5b9c6a2c"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:16:51 crc kubenswrapper[4754]: I1005 21:16:51.781609 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1d96e280-8c26-41ea-ab29-64ce5b9c6a2c-config-data" (OuterVolumeSpecName: "config-data") pod "1d96e280-8c26-41ea-ab29-64ce5b9c6a2c" (UID: "1d96e280-8c26-41ea-ab29-64ce5b9c6a2c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 21:16:51 crc kubenswrapper[4754]: I1005 21:16:51.824264 4754 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/1d96e280-8c26-41ea-ab29-64ce5b9c6a2c-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Oct 05 21:16:51 crc kubenswrapper[4754]: I1005 21:16:51.824293 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ltd4\" (UniqueName: \"kubernetes.io/projected/1d96e280-8c26-41ea-ab29-64ce5b9c6a2c-kube-api-access-6ltd4\") on node \"crc\" DevicePath \"\"" Oct 05 21:16:51 crc kubenswrapper[4754]: I1005 21:16:51.824306 4754 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/1d96e280-8c26-41ea-ab29-64ce5b9c6a2c-pod-info\") on node \"crc\" DevicePath \"\"" Oct 05 21:16:51 crc kubenswrapper[4754]: I1005 21:16:51.824314 4754 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/1d96e280-8c26-41ea-ab29-64ce5b9c6a2c-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Oct 05 21:16:51 crc kubenswrapper[4754]: I1005 21:16:51.824322 4754 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/1d96e280-8c26-41ea-ab29-64ce5b9c6a2c-plugins-conf\") on node \"crc\" DevicePath \"\"" Oct 05 21:16:51 crc kubenswrapper[4754]: I1005 21:16:51.824331 4754 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1d96e280-8c26-41ea-ab29-64ce5b9c6a2c-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 21:16:51 crc kubenswrapper[4754]: I1005 21:16:51.824339 4754 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/1d96e280-8c26-41ea-ab29-64ce5b9c6a2c-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Oct 05 21:16:51 crc kubenswrapper[4754]: I1005 21:16:51.824365 4754 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Oct 05 21:16:51 crc kubenswrapper[4754]: I1005 21:16:51.824373 4754 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/1d96e280-8c26-41ea-ab29-64ce5b9c6a2c-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Oct 05 21:16:51 crc kubenswrapper[4754]: I1005 21:16:51.834376 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1d96e280-8c26-41ea-ab29-64ce5b9c6a2c-server-conf" (OuterVolumeSpecName: "server-conf") pod "1d96e280-8c26-41ea-ab29-64ce5b9c6a2c" (UID: "1d96e280-8c26-41ea-ab29-64ce5b9c6a2c"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 21:16:51 crc kubenswrapper[4754]: I1005 21:16:51.857831 4754 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Oct 05 21:16:51 crc kubenswrapper[4754]: I1005 21:16:51.887765 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d96e280-8c26-41ea-ab29-64ce5b9c6a2c-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "1d96e280-8c26-41ea-ab29-64ce5b9c6a2c" (UID: "1d96e280-8c26-41ea-ab29-64ce5b9c6a2c"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:16:51 crc kubenswrapper[4754]: I1005 21:16:51.891000 4754 scope.go:117] "RemoveContainer" containerID="493f2a671031a6ad481270c980a4f16b3d2567e2e2da7677960e8f9fba245eaf" Oct 05 21:16:51 crc kubenswrapper[4754]: I1005 21:16:51.926270 4754 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Oct 05 21:16:51 crc kubenswrapper[4754]: I1005 21:16:51.926302 4754 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/1d96e280-8c26-41ea-ab29-64ce5b9c6a2c-server-conf\") on node \"crc\" DevicePath \"\"" Oct 05 21:16:51 crc kubenswrapper[4754]: I1005 21:16:51.926311 4754 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/1d96e280-8c26-41ea-ab29-64ce5b9c6a2c-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Oct 05 21:16:51 crc kubenswrapper[4754]: I1005 21:16:51.942306 4754 scope.go:117] "RemoveContainer" containerID="1a0e1b3cbb7432cf1c8d8d5b93db680fd2af14a635594bd4eb551779f7e4ba30" Oct 05 21:16:52 crc kubenswrapper[4754]: I1005 21:16:52.018473 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"39aff8a0-2905-4ad3-ad2d-5bf6b5de3858","Type":"ContainerDied","Data":"2e640693fa23c07efabf6eb4ac474cd2b47ab78643ef4513ff944ad3ce0b5dc9"} Oct 05 21:16:52 crc kubenswrapper[4754]: I1005 21:16:52.018532 4754 scope.go:117] "RemoveContainer" containerID="e4012712472c94cd11327163abb6c6f5c55cff5c6625d3242d418e93290fa5bf" Oct 05 21:16:52 crc kubenswrapper[4754]: I1005 21:16:52.018648 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 05 21:16:52 crc kubenswrapper[4754]: I1005 21:16:52.032078 4754 generic.go:334] "Generic (PLEG): container finished" podID="1d96e280-8c26-41ea-ab29-64ce5b9c6a2c" containerID="ea5291016f4053d149506d9f4e1a17c9d6281d48b900045f7a5bc743cc386a39" exitCode=0 Oct 05 21:16:52 crc kubenswrapper[4754]: I1005 21:16:52.032123 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"1d96e280-8c26-41ea-ab29-64ce5b9c6a2c","Type":"ContainerDied","Data":"ea5291016f4053d149506d9f4e1a17c9d6281d48b900045f7a5bc743cc386a39"} Oct 05 21:16:52 crc kubenswrapper[4754]: I1005 21:16:52.032152 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"1d96e280-8c26-41ea-ab29-64ce5b9c6a2c","Type":"ContainerDied","Data":"5709134c4abd234ef35cfa828d09fdbab849e2251253f921d5bf35df040753b0"} Oct 05 21:16:52 crc kubenswrapper[4754]: I1005 21:16:52.032205 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 05 21:16:52 crc kubenswrapper[4754]: I1005 21:16:52.069040 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 05 21:16:52 crc kubenswrapper[4754]: I1005 21:16:52.085640 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 05 21:16:52 crc kubenswrapper[4754]: I1005 21:16:52.100632 4754 scope.go:117] "RemoveContainer" containerID="ea5291016f4053d149506d9f4e1a17c9d6281d48b900045f7a5bc743cc386a39" Oct 05 21:16:52 crc kubenswrapper[4754]: I1005 21:16:52.104067 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 05 21:16:52 crc kubenswrapper[4754]: I1005 21:16:52.111119 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 05 21:16:52 crc kubenswrapper[4754]: I1005 21:16:52.125817 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Oct 05 21:16:52 crc kubenswrapper[4754]: E1005 21:16:52.126295 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39aff8a0-2905-4ad3-ad2d-5bf6b5de3858" containerName="rabbitmq" Oct 05 21:16:52 crc kubenswrapper[4754]: I1005 21:16:52.126307 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="39aff8a0-2905-4ad3-ad2d-5bf6b5de3858" containerName="rabbitmq" Oct 05 21:16:52 crc kubenswrapper[4754]: E1005 21:16:52.126339 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27cd6be5-f5eb-4158-bdd1-37f4df3d530d" containerName="horizon" Oct 05 21:16:52 crc kubenswrapper[4754]: I1005 21:16:52.126346 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="27cd6be5-f5eb-4158-bdd1-37f4df3d530d" containerName="horizon" Oct 05 21:16:52 crc kubenswrapper[4754]: E1005 21:16:52.126364 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27cd6be5-f5eb-4158-bdd1-37f4df3d530d" containerName="horizon" Oct 05 21:16:52 crc kubenswrapper[4754]: I1005 21:16:52.126371 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="27cd6be5-f5eb-4158-bdd1-37f4df3d530d" containerName="horizon" Oct 05 21:16:52 crc kubenswrapper[4754]: E1005 21:16:52.126383 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27cd6be5-f5eb-4158-bdd1-37f4df3d530d" containerName="horizon" Oct 05 21:16:52 crc kubenswrapper[4754]: I1005 21:16:52.126390 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="27cd6be5-f5eb-4158-bdd1-37f4df3d530d" containerName="horizon" Oct 05 21:16:52 crc kubenswrapper[4754]: E1005 21:16:52.126406 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27cd6be5-f5eb-4158-bdd1-37f4df3d530d" containerName="horizon" Oct 05 21:16:52 crc kubenswrapper[4754]: I1005 21:16:52.126413 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="27cd6be5-f5eb-4158-bdd1-37f4df3d530d" containerName="horizon" Oct 05 21:16:52 crc kubenswrapper[4754]: E1005 21:16:52.126421 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39aff8a0-2905-4ad3-ad2d-5bf6b5de3858" containerName="setup-container" Oct 05 21:16:52 crc kubenswrapper[4754]: I1005 21:16:52.126427 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="39aff8a0-2905-4ad3-ad2d-5bf6b5de3858" containerName="setup-container" Oct 05 21:16:52 crc kubenswrapper[4754]: E1005 21:16:52.126435 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d96e280-8c26-41ea-ab29-64ce5b9c6a2c" containerName="rabbitmq" Oct 05 21:16:52 crc kubenswrapper[4754]: I1005 21:16:52.126440 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d96e280-8c26-41ea-ab29-64ce5b9c6a2c" containerName="rabbitmq" Oct 05 21:16:52 crc kubenswrapper[4754]: E1005 21:16:52.126454 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27cd6be5-f5eb-4158-bdd1-37f4df3d530d" containerName="horizon-log" Oct 05 21:16:52 crc kubenswrapper[4754]: I1005 21:16:52.126461 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="27cd6be5-f5eb-4158-bdd1-37f4df3d530d" containerName="horizon-log" Oct 05 21:16:52 crc kubenswrapper[4754]: E1005 21:16:52.126475 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d96e280-8c26-41ea-ab29-64ce5b9c6a2c" containerName="setup-container" Oct 05 21:16:52 crc kubenswrapper[4754]: I1005 21:16:52.126481 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d96e280-8c26-41ea-ab29-64ce5b9c6a2c" containerName="setup-container" Oct 05 21:16:52 crc kubenswrapper[4754]: I1005 21:16:52.126680 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="27cd6be5-f5eb-4158-bdd1-37f4df3d530d" containerName="horizon" Oct 05 21:16:52 crc kubenswrapper[4754]: I1005 21:16:52.126694 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="27cd6be5-f5eb-4158-bdd1-37f4df3d530d" containerName="horizon" Oct 05 21:16:52 crc kubenswrapper[4754]: I1005 21:16:52.126710 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="27cd6be5-f5eb-4158-bdd1-37f4df3d530d" containerName="horizon" Oct 05 21:16:52 crc kubenswrapper[4754]: I1005 21:16:52.126716 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="27cd6be5-f5eb-4158-bdd1-37f4df3d530d" containerName="horizon-log" Oct 05 21:16:52 crc kubenswrapper[4754]: I1005 21:16:52.126729 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="39aff8a0-2905-4ad3-ad2d-5bf6b5de3858" containerName="rabbitmq" Oct 05 21:16:52 crc kubenswrapper[4754]: I1005 21:16:52.126738 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d96e280-8c26-41ea-ab29-64ce5b9c6a2c" containerName="rabbitmq" Oct 05 21:16:52 crc kubenswrapper[4754]: I1005 21:16:52.127110 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="27cd6be5-f5eb-4158-bdd1-37f4df3d530d" containerName="horizon" Oct 05 21:16:52 crc kubenswrapper[4754]: I1005 21:16:52.128579 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 05 21:16:52 crc kubenswrapper[4754]: I1005 21:16:52.130228 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-dm69s" Oct 05 21:16:52 crc kubenswrapper[4754]: I1005 21:16:52.130353 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 05 21:16:52 crc kubenswrapper[4754]: I1005 21:16:52.131391 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Oct 05 21:16:52 crc kubenswrapper[4754]: I1005 21:16:52.131610 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Oct 05 21:16:52 crc kubenswrapper[4754]: I1005 21:16:52.132393 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 05 21:16:52 crc kubenswrapper[4754]: I1005 21:16:52.133395 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Oct 05 21:16:52 crc kubenswrapper[4754]: I1005 21:16:52.146211 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 05 21:16:52 crc kubenswrapper[4754]: I1005 21:16:52.147618 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Oct 05 21:16:52 crc kubenswrapper[4754]: I1005 21:16:52.147861 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Oct 05 21:16:52 crc kubenswrapper[4754]: I1005 21:16:52.148029 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Oct 05 21:16:52 crc kubenswrapper[4754]: I1005 21:16:52.148147 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Oct 05 21:16:52 crc kubenswrapper[4754]: I1005 21:16:52.148297 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Oct 05 21:16:52 crc kubenswrapper[4754]: I1005 21:16:52.148437 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Oct 05 21:16:52 crc kubenswrapper[4754]: I1005 21:16:52.148626 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Oct 05 21:16:52 crc kubenswrapper[4754]: I1005 21:16:52.149466 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Oct 05 21:16:52 crc kubenswrapper[4754]: I1005 21:16:52.149719 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-bkl2n" Oct 05 21:16:52 crc kubenswrapper[4754]: I1005 21:16:52.155463 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 05 21:16:52 crc kubenswrapper[4754]: I1005 21:16:52.172211 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Oct 05 21:16:52 crc kubenswrapper[4754]: I1005 21:16:52.210811 4754 scope.go:117] "RemoveContainer" containerID="ea5291016f4053d149506d9f4e1a17c9d6281d48b900045f7a5bc743cc386a39" Oct 05 21:16:52 crc kubenswrapper[4754]: E1005 21:16:52.212278 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ea5291016f4053d149506d9f4e1a17c9d6281d48b900045f7a5bc743cc386a39\": container with ID starting with ea5291016f4053d149506d9f4e1a17c9d6281d48b900045f7a5bc743cc386a39 not found: ID does not exist" containerID="ea5291016f4053d149506d9f4e1a17c9d6281d48b900045f7a5bc743cc386a39" Oct 05 21:16:52 crc kubenswrapper[4754]: I1005 21:16:52.212321 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ea5291016f4053d149506d9f4e1a17c9d6281d48b900045f7a5bc743cc386a39"} err="failed to get container status \"ea5291016f4053d149506d9f4e1a17c9d6281d48b900045f7a5bc743cc386a39\": rpc error: code = NotFound desc = could not find container \"ea5291016f4053d149506d9f4e1a17c9d6281d48b900045f7a5bc743cc386a39\": container with ID starting with ea5291016f4053d149506d9f4e1a17c9d6281d48b900045f7a5bc743cc386a39 not found: ID does not exist" Oct 05 21:16:52 crc kubenswrapper[4754]: I1005 21:16:52.234760 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-server-0\" (UID: \"3a6824aa-cdc6-4598-b296-1db23f99b156\") " pod="openstack/rabbitmq-server-0" Oct 05 21:16:52 crc kubenswrapper[4754]: I1005 21:16:52.234798 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3a6824aa-cdc6-4598-b296-1db23f99b156-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"3a6824aa-cdc6-4598-b296-1db23f99b156\") " pod="openstack/rabbitmq-server-0" Oct 05 21:16:52 crc kubenswrapper[4754]: I1005 21:16:52.234837 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/7eb0b89b-e87d-4d5e-9939-d3a0f97f94db-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"7eb0b89b-e87d-4d5e-9939-d3a0f97f94db\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 21:16:52 crc kubenswrapper[4754]: I1005 21:16:52.234860 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/7eb0b89b-e87d-4d5e-9939-d3a0f97f94db-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"7eb0b89b-e87d-4d5e-9939-d3a0f97f94db\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 21:16:52 crc kubenswrapper[4754]: I1005 21:16:52.234880 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3a6824aa-cdc6-4598-b296-1db23f99b156-pod-info\") pod \"rabbitmq-server-0\" (UID: \"3a6824aa-cdc6-4598-b296-1db23f99b156\") " pod="openstack/rabbitmq-server-0" Oct 05 21:16:52 crc kubenswrapper[4754]: I1005 21:16:52.234902 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/7eb0b89b-e87d-4d5e-9939-d3a0f97f94db-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"7eb0b89b-e87d-4d5e-9939-d3a0f97f94db\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 21:16:52 crc kubenswrapper[4754]: I1005 21:16:52.234938 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3a6824aa-cdc6-4598-b296-1db23f99b156-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"3a6824aa-cdc6-4598-b296-1db23f99b156\") " pod="openstack/rabbitmq-server-0" Oct 05 21:16:52 crc kubenswrapper[4754]: I1005 21:16:52.234953 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3a6824aa-cdc6-4598-b296-1db23f99b156-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"3a6824aa-cdc6-4598-b296-1db23f99b156\") " pod="openstack/rabbitmq-server-0" Oct 05 21:16:52 crc kubenswrapper[4754]: I1005 21:16:52.234972 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/7eb0b89b-e87d-4d5e-9939-d3a0f97f94db-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"7eb0b89b-e87d-4d5e-9939-d3a0f97f94db\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 21:16:52 crc kubenswrapper[4754]: I1005 21:16:52.234988 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/7eb0b89b-e87d-4d5e-9939-d3a0f97f94db-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"7eb0b89b-e87d-4d5e-9939-d3a0f97f94db\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 21:16:52 crc kubenswrapper[4754]: I1005 21:16:52.235015 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/7eb0b89b-e87d-4d5e-9939-d3a0f97f94db-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"7eb0b89b-e87d-4d5e-9939-d3a0f97f94db\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 21:16:52 crc kubenswrapper[4754]: I1005 21:16:52.235041 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3a6824aa-cdc6-4598-b296-1db23f99b156-config-data\") pod \"rabbitmq-server-0\" (UID: \"3a6824aa-cdc6-4598-b296-1db23f99b156\") " pod="openstack/rabbitmq-server-0" Oct 05 21:16:52 crc kubenswrapper[4754]: I1005 21:16:52.235057 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7eb0b89b-e87d-4d5e-9939-d3a0f97f94db-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"7eb0b89b-e87d-4d5e-9939-d3a0f97f94db\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 21:16:52 crc kubenswrapper[4754]: I1005 21:16:52.235183 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4cwf5\" (UniqueName: \"kubernetes.io/projected/7eb0b89b-e87d-4d5e-9939-d3a0f97f94db-kube-api-access-4cwf5\") pod \"rabbitmq-cell1-server-0\" (UID: \"7eb0b89b-e87d-4d5e-9939-d3a0f97f94db\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 21:16:52 crc kubenswrapper[4754]: I1005 21:16:52.235230 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3a6824aa-cdc6-4598-b296-1db23f99b156-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"3a6824aa-cdc6-4598-b296-1db23f99b156\") " pod="openstack/rabbitmq-server-0" Oct 05 21:16:52 crc kubenswrapper[4754]: I1005 21:16:52.235277 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"7eb0b89b-e87d-4d5e-9939-d3a0f97f94db\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 21:16:52 crc kubenswrapper[4754]: I1005 21:16:52.235584 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3a6824aa-cdc6-4598-b296-1db23f99b156-server-conf\") pod \"rabbitmq-server-0\" (UID: \"3a6824aa-cdc6-4598-b296-1db23f99b156\") " pod="openstack/rabbitmq-server-0" Oct 05 21:16:52 crc kubenswrapper[4754]: I1005 21:16:52.235620 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/3a6824aa-cdc6-4598-b296-1db23f99b156-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"3a6824aa-cdc6-4598-b296-1db23f99b156\") " pod="openstack/rabbitmq-server-0" Oct 05 21:16:52 crc kubenswrapper[4754]: I1005 21:16:52.235653 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/7eb0b89b-e87d-4d5e-9939-d3a0f97f94db-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"7eb0b89b-e87d-4d5e-9939-d3a0f97f94db\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 21:16:52 crc kubenswrapper[4754]: I1005 21:16:52.235723 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3a6824aa-cdc6-4598-b296-1db23f99b156-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"3a6824aa-cdc6-4598-b296-1db23f99b156\") " pod="openstack/rabbitmq-server-0" Oct 05 21:16:52 crc kubenswrapper[4754]: I1005 21:16:52.235760 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/7eb0b89b-e87d-4d5e-9939-d3a0f97f94db-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"7eb0b89b-e87d-4d5e-9939-d3a0f97f94db\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 21:16:52 crc kubenswrapper[4754]: I1005 21:16:52.235920 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-blv5c\" (UniqueName: \"kubernetes.io/projected/3a6824aa-cdc6-4598-b296-1db23f99b156-kube-api-access-blv5c\") pod \"rabbitmq-server-0\" (UID: \"3a6824aa-cdc6-4598-b296-1db23f99b156\") " pod="openstack/rabbitmq-server-0" Oct 05 21:16:52 crc kubenswrapper[4754]: I1005 21:16:52.339624 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-blv5c\" (UniqueName: \"kubernetes.io/projected/3a6824aa-cdc6-4598-b296-1db23f99b156-kube-api-access-blv5c\") pod \"rabbitmq-server-0\" (UID: \"3a6824aa-cdc6-4598-b296-1db23f99b156\") " pod="openstack/rabbitmq-server-0" Oct 05 21:16:52 crc kubenswrapper[4754]: I1005 21:16:52.339964 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-server-0\" (UID: \"3a6824aa-cdc6-4598-b296-1db23f99b156\") " pod="openstack/rabbitmq-server-0" Oct 05 21:16:52 crc kubenswrapper[4754]: I1005 21:16:52.340039 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3a6824aa-cdc6-4598-b296-1db23f99b156-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"3a6824aa-cdc6-4598-b296-1db23f99b156\") " pod="openstack/rabbitmq-server-0" Oct 05 21:16:52 crc kubenswrapper[4754]: I1005 21:16:52.340157 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/7eb0b89b-e87d-4d5e-9939-d3a0f97f94db-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"7eb0b89b-e87d-4d5e-9939-d3a0f97f94db\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 21:16:52 crc kubenswrapper[4754]: I1005 21:16:52.340236 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/7eb0b89b-e87d-4d5e-9939-d3a0f97f94db-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"7eb0b89b-e87d-4d5e-9939-d3a0f97f94db\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 21:16:52 crc kubenswrapper[4754]: I1005 21:16:52.340283 4754 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-server-0\" (UID: \"3a6824aa-cdc6-4598-b296-1db23f99b156\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/rabbitmq-server-0" Oct 05 21:16:52 crc kubenswrapper[4754]: I1005 21:16:52.343935 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/7eb0b89b-e87d-4d5e-9939-d3a0f97f94db-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"7eb0b89b-e87d-4d5e-9939-d3a0f97f94db\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 21:16:52 crc kubenswrapper[4754]: I1005 21:16:52.340292 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3a6824aa-cdc6-4598-b296-1db23f99b156-pod-info\") pod \"rabbitmq-server-0\" (UID: \"3a6824aa-cdc6-4598-b296-1db23f99b156\") " pod="openstack/rabbitmq-server-0" Oct 05 21:16:52 crc kubenswrapper[4754]: I1005 21:16:52.357877 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/7eb0b89b-e87d-4d5e-9939-d3a0f97f94db-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"7eb0b89b-e87d-4d5e-9939-d3a0f97f94db\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 21:16:52 crc kubenswrapper[4754]: I1005 21:16:52.358047 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3a6824aa-cdc6-4598-b296-1db23f99b156-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"3a6824aa-cdc6-4598-b296-1db23f99b156\") " pod="openstack/rabbitmq-server-0" Oct 05 21:16:52 crc kubenswrapper[4754]: I1005 21:16:52.358081 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3a6824aa-cdc6-4598-b296-1db23f99b156-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"3a6824aa-cdc6-4598-b296-1db23f99b156\") " pod="openstack/rabbitmq-server-0" Oct 05 21:16:52 crc kubenswrapper[4754]: I1005 21:16:52.358147 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/7eb0b89b-e87d-4d5e-9939-d3a0f97f94db-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"7eb0b89b-e87d-4d5e-9939-d3a0f97f94db\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 21:16:52 crc kubenswrapper[4754]: I1005 21:16:52.358186 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/7eb0b89b-e87d-4d5e-9939-d3a0f97f94db-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"7eb0b89b-e87d-4d5e-9939-d3a0f97f94db\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 21:16:52 crc kubenswrapper[4754]: I1005 21:16:52.358313 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/7eb0b89b-e87d-4d5e-9939-d3a0f97f94db-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"7eb0b89b-e87d-4d5e-9939-d3a0f97f94db\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 21:16:52 crc kubenswrapper[4754]: I1005 21:16:52.358413 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3a6824aa-cdc6-4598-b296-1db23f99b156-config-data\") pod \"rabbitmq-server-0\" (UID: \"3a6824aa-cdc6-4598-b296-1db23f99b156\") " pod="openstack/rabbitmq-server-0" Oct 05 21:16:52 crc kubenswrapper[4754]: I1005 21:16:52.358446 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7eb0b89b-e87d-4d5e-9939-d3a0f97f94db-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"7eb0b89b-e87d-4d5e-9939-d3a0f97f94db\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 21:16:52 crc kubenswrapper[4754]: I1005 21:16:52.358503 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4cwf5\" (UniqueName: \"kubernetes.io/projected/7eb0b89b-e87d-4d5e-9939-d3a0f97f94db-kube-api-access-4cwf5\") pod \"rabbitmq-cell1-server-0\" (UID: \"7eb0b89b-e87d-4d5e-9939-d3a0f97f94db\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 21:16:52 crc kubenswrapper[4754]: I1005 21:16:52.358531 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3a6824aa-cdc6-4598-b296-1db23f99b156-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"3a6824aa-cdc6-4598-b296-1db23f99b156\") " pod="openstack/rabbitmq-server-0" Oct 05 21:16:52 crc kubenswrapper[4754]: I1005 21:16:52.358572 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"7eb0b89b-e87d-4d5e-9939-d3a0f97f94db\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 21:16:52 crc kubenswrapper[4754]: I1005 21:16:52.358756 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3a6824aa-cdc6-4598-b296-1db23f99b156-server-conf\") pod \"rabbitmq-server-0\" (UID: \"3a6824aa-cdc6-4598-b296-1db23f99b156\") " pod="openstack/rabbitmq-server-0" Oct 05 21:16:52 crc kubenswrapper[4754]: I1005 21:16:52.358805 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/3a6824aa-cdc6-4598-b296-1db23f99b156-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"3a6824aa-cdc6-4598-b296-1db23f99b156\") " pod="openstack/rabbitmq-server-0" Oct 05 21:16:52 crc kubenswrapper[4754]: I1005 21:16:52.358849 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/7eb0b89b-e87d-4d5e-9939-d3a0f97f94db-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"7eb0b89b-e87d-4d5e-9939-d3a0f97f94db\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 21:16:52 crc kubenswrapper[4754]: I1005 21:16:52.358934 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3a6824aa-cdc6-4598-b296-1db23f99b156-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"3a6824aa-cdc6-4598-b296-1db23f99b156\") " pod="openstack/rabbitmq-server-0" Oct 05 21:16:52 crc kubenswrapper[4754]: I1005 21:16:52.358995 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/7eb0b89b-e87d-4d5e-9939-d3a0f97f94db-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"7eb0b89b-e87d-4d5e-9939-d3a0f97f94db\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 21:16:52 crc kubenswrapper[4754]: I1005 21:16:52.362847 4754 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"7eb0b89b-e87d-4d5e-9939-d3a0f97f94db\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/rabbitmq-cell1-server-0" Oct 05 21:16:52 crc kubenswrapper[4754]: I1005 21:16:52.344375 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/7eb0b89b-e87d-4d5e-9939-d3a0f97f94db-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"7eb0b89b-e87d-4d5e-9939-d3a0f97f94db\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 21:16:52 crc kubenswrapper[4754]: I1005 21:16:52.365991 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3a6824aa-cdc6-4598-b296-1db23f99b156-config-data\") pod \"rabbitmq-server-0\" (UID: \"3a6824aa-cdc6-4598-b296-1db23f99b156\") " pod="openstack/rabbitmq-server-0" Oct 05 21:16:52 crc kubenswrapper[4754]: I1005 21:16:52.344657 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3a6824aa-cdc6-4598-b296-1db23f99b156-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"3a6824aa-cdc6-4598-b296-1db23f99b156\") " pod="openstack/rabbitmq-server-0" Oct 05 21:16:52 crc kubenswrapper[4754]: I1005 21:16:52.355353 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3a6824aa-cdc6-4598-b296-1db23f99b156-pod-info\") pod \"rabbitmq-server-0\" (UID: \"3a6824aa-cdc6-4598-b296-1db23f99b156\") " pod="openstack/rabbitmq-server-0" Oct 05 21:16:52 crc kubenswrapper[4754]: I1005 21:16:52.369042 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3a6824aa-cdc6-4598-b296-1db23f99b156-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"3a6824aa-cdc6-4598-b296-1db23f99b156\") " pod="openstack/rabbitmq-server-0" Oct 05 21:16:52 crc kubenswrapper[4754]: I1005 21:16:52.372092 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/7eb0b89b-e87d-4d5e-9939-d3a0f97f94db-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"7eb0b89b-e87d-4d5e-9939-d3a0f97f94db\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 21:16:52 crc kubenswrapper[4754]: I1005 21:16:52.375149 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3a6824aa-cdc6-4598-b296-1db23f99b156-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"3a6824aa-cdc6-4598-b296-1db23f99b156\") " pod="openstack/rabbitmq-server-0" Oct 05 21:16:52 crc kubenswrapper[4754]: I1005 21:16:52.377388 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/7eb0b89b-e87d-4d5e-9939-d3a0f97f94db-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"7eb0b89b-e87d-4d5e-9939-d3a0f97f94db\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 21:16:52 crc kubenswrapper[4754]: I1005 21:16:52.378116 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7eb0b89b-e87d-4d5e-9939-d3a0f97f94db-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"7eb0b89b-e87d-4d5e-9939-d3a0f97f94db\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 21:16:52 crc kubenswrapper[4754]: I1005 21:16:52.378321 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/7eb0b89b-e87d-4d5e-9939-d3a0f97f94db-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"7eb0b89b-e87d-4d5e-9939-d3a0f97f94db\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 21:16:52 crc kubenswrapper[4754]: I1005 21:16:52.378927 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3a6824aa-cdc6-4598-b296-1db23f99b156-server-conf\") pod \"rabbitmq-server-0\" (UID: \"3a6824aa-cdc6-4598-b296-1db23f99b156\") " pod="openstack/rabbitmq-server-0" Oct 05 21:16:52 crc kubenswrapper[4754]: I1005 21:16:52.386393 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/7eb0b89b-e87d-4d5e-9939-d3a0f97f94db-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"7eb0b89b-e87d-4d5e-9939-d3a0f97f94db\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 21:16:52 crc kubenswrapper[4754]: I1005 21:16:52.388123 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/3a6824aa-cdc6-4598-b296-1db23f99b156-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"3a6824aa-cdc6-4598-b296-1db23f99b156\") " pod="openstack/rabbitmq-server-0" Oct 05 21:16:52 crc kubenswrapper[4754]: I1005 21:16:52.393890 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-blv5c\" (UniqueName: \"kubernetes.io/projected/3a6824aa-cdc6-4598-b296-1db23f99b156-kube-api-access-blv5c\") pod \"rabbitmq-server-0\" (UID: \"3a6824aa-cdc6-4598-b296-1db23f99b156\") " pod="openstack/rabbitmq-server-0" Oct 05 21:16:52 crc kubenswrapper[4754]: I1005 21:16:52.394802 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/7eb0b89b-e87d-4d5e-9939-d3a0f97f94db-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"7eb0b89b-e87d-4d5e-9939-d3a0f97f94db\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 21:16:52 crc kubenswrapper[4754]: I1005 21:16:52.398733 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/7eb0b89b-e87d-4d5e-9939-d3a0f97f94db-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"7eb0b89b-e87d-4d5e-9939-d3a0f97f94db\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 21:16:52 crc kubenswrapper[4754]: I1005 21:16:52.406406 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3a6824aa-cdc6-4598-b296-1db23f99b156-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"3a6824aa-cdc6-4598-b296-1db23f99b156\") " pod="openstack/rabbitmq-server-0" Oct 05 21:16:52 crc kubenswrapper[4754]: I1005 21:16:52.416110 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3a6824aa-cdc6-4598-b296-1db23f99b156-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"3a6824aa-cdc6-4598-b296-1db23f99b156\") " pod="openstack/rabbitmq-server-0" Oct 05 21:16:52 crc kubenswrapper[4754]: I1005 21:16:52.416663 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4cwf5\" (UniqueName: \"kubernetes.io/projected/7eb0b89b-e87d-4d5e-9939-d3a0f97f94db-kube-api-access-4cwf5\") pod \"rabbitmq-cell1-server-0\" (UID: \"7eb0b89b-e87d-4d5e-9939-d3a0f97f94db\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 21:16:52 crc kubenswrapper[4754]: I1005 21:16:52.441962 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-server-0\" (UID: \"3a6824aa-cdc6-4598-b296-1db23f99b156\") " pod="openstack/rabbitmq-server-0" Oct 05 21:16:52 crc kubenswrapper[4754]: I1005 21:16:52.450903 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"7eb0b89b-e87d-4d5e-9939-d3a0f97f94db\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 21:16:52 crc kubenswrapper[4754]: I1005 21:16:52.482291 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 05 21:16:52 crc kubenswrapper[4754]: I1005 21:16:52.495983 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 05 21:16:52 crc kubenswrapper[4754]: I1005 21:16:52.850183 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d96e280-8c26-41ea-ab29-64ce5b9c6a2c" path="/var/lib/kubelet/pods/1d96e280-8c26-41ea-ab29-64ce5b9c6a2c/volumes" Oct 05 21:16:52 crc kubenswrapper[4754]: I1005 21:16:52.853768 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="39aff8a0-2905-4ad3-ad2d-5bf6b5de3858" path="/var/lib/kubelet/pods/39aff8a0-2905-4ad3-ad2d-5bf6b5de3858/volumes" Oct 05 21:16:53 crc kubenswrapper[4754]: I1005 21:16:53.010319 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-67b789f86c-bbpr4"] Oct 05 21:16:53 crc kubenswrapper[4754]: I1005 21:16:53.012102 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67b789f86c-bbpr4" Oct 05 21:16:53 crc kubenswrapper[4754]: I1005 21:16:53.014825 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Oct 05 21:16:53 crc kubenswrapper[4754]: I1005 21:16:53.073263 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-67b789f86c-bbpr4"] Oct 05 21:16:53 crc kubenswrapper[4754]: I1005 21:16:53.075975 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3e67b2a1-6784-48f6-8f13-a60ff73a5b3f-dns-swift-storage-0\") pod \"dnsmasq-dns-67b789f86c-bbpr4\" (UID: \"3e67b2a1-6784-48f6-8f13-a60ff73a5b3f\") " pod="openstack/dnsmasq-dns-67b789f86c-bbpr4" Oct 05 21:16:53 crc kubenswrapper[4754]: I1005 21:16:53.076024 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3e67b2a1-6784-48f6-8f13-a60ff73a5b3f-dns-svc\") pod \"dnsmasq-dns-67b789f86c-bbpr4\" (UID: \"3e67b2a1-6784-48f6-8f13-a60ff73a5b3f\") " pod="openstack/dnsmasq-dns-67b789f86c-bbpr4" Oct 05 21:16:53 crc kubenswrapper[4754]: I1005 21:16:53.076051 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3e67b2a1-6784-48f6-8f13-a60ff73a5b3f-ovsdbserver-sb\") pod \"dnsmasq-dns-67b789f86c-bbpr4\" (UID: \"3e67b2a1-6784-48f6-8f13-a60ff73a5b3f\") " pod="openstack/dnsmasq-dns-67b789f86c-bbpr4" Oct 05 21:16:53 crc kubenswrapper[4754]: I1005 21:16:53.076076 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/3e67b2a1-6784-48f6-8f13-a60ff73a5b3f-openstack-edpm-ipam\") pod \"dnsmasq-dns-67b789f86c-bbpr4\" (UID: \"3e67b2a1-6784-48f6-8f13-a60ff73a5b3f\") " pod="openstack/dnsmasq-dns-67b789f86c-bbpr4" Oct 05 21:16:53 crc kubenswrapper[4754]: I1005 21:16:53.076114 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2vpqr\" (UniqueName: \"kubernetes.io/projected/3e67b2a1-6784-48f6-8f13-a60ff73a5b3f-kube-api-access-2vpqr\") pod \"dnsmasq-dns-67b789f86c-bbpr4\" (UID: \"3e67b2a1-6784-48f6-8f13-a60ff73a5b3f\") " pod="openstack/dnsmasq-dns-67b789f86c-bbpr4" Oct 05 21:16:53 crc kubenswrapper[4754]: I1005 21:16:53.076151 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3e67b2a1-6784-48f6-8f13-a60ff73a5b3f-ovsdbserver-nb\") pod \"dnsmasq-dns-67b789f86c-bbpr4\" (UID: \"3e67b2a1-6784-48f6-8f13-a60ff73a5b3f\") " pod="openstack/dnsmasq-dns-67b789f86c-bbpr4" Oct 05 21:16:53 crc kubenswrapper[4754]: I1005 21:16:53.076173 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3e67b2a1-6784-48f6-8f13-a60ff73a5b3f-config\") pod \"dnsmasq-dns-67b789f86c-bbpr4\" (UID: \"3e67b2a1-6784-48f6-8f13-a60ff73a5b3f\") " pod="openstack/dnsmasq-dns-67b789f86c-bbpr4" Oct 05 21:16:53 crc kubenswrapper[4754]: I1005 21:16:53.080635 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 05 21:16:53 crc kubenswrapper[4754]: I1005 21:16:53.157212 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 05 21:16:53 crc kubenswrapper[4754]: I1005 21:16:53.178011 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3e67b2a1-6784-48f6-8f13-a60ff73a5b3f-dns-swift-storage-0\") pod \"dnsmasq-dns-67b789f86c-bbpr4\" (UID: \"3e67b2a1-6784-48f6-8f13-a60ff73a5b3f\") " pod="openstack/dnsmasq-dns-67b789f86c-bbpr4" Oct 05 21:16:53 crc kubenswrapper[4754]: I1005 21:16:53.179403 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3e67b2a1-6784-48f6-8f13-a60ff73a5b3f-dns-svc\") pod \"dnsmasq-dns-67b789f86c-bbpr4\" (UID: \"3e67b2a1-6784-48f6-8f13-a60ff73a5b3f\") " pod="openstack/dnsmasq-dns-67b789f86c-bbpr4" Oct 05 21:16:53 crc kubenswrapper[4754]: I1005 21:16:53.179513 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3e67b2a1-6784-48f6-8f13-a60ff73a5b3f-ovsdbserver-sb\") pod \"dnsmasq-dns-67b789f86c-bbpr4\" (UID: \"3e67b2a1-6784-48f6-8f13-a60ff73a5b3f\") " pod="openstack/dnsmasq-dns-67b789f86c-bbpr4" Oct 05 21:16:53 crc kubenswrapper[4754]: I1005 21:16:53.179588 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/3e67b2a1-6784-48f6-8f13-a60ff73a5b3f-openstack-edpm-ipam\") pod \"dnsmasq-dns-67b789f86c-bbpr4\" (UID: \"3e67b2a1-6784-48f6-8f13-a60ff73a5b3f\") " pod="openstack/dnsmasq-dns-67b789f86c-bbpr4" Oct 05 21:16:53 crc kubenswrapper[4754]: I1005 21:16:53.179674 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2vpqr\" (UniqueName: \"kubernetes.io/projected/3e67b2a1-6784-48f6-8f13-a60ff73a5b3f-kube-api-access-2vpqr\") pod \"dnsmasq-dns-67b789f86c-bbpr4\" (UID: \"3e67b2a1-6784-48f6-8f13-a60ff73a5b3f\") " pod="openstack/dnsmasq-dns-67b789f86c-bbpr4" Oct 05 21:16:53 crc kubenswrapper[4754]: I1005 21:16:53.179774 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3e67b2a1-6784-48f6-8f13-a60ff73a5b3f-ovsdbserver-nb\") pod \"dnsmasq-dns-67b789f86c-bbpr4\" (UID: \"3e67b2a1-6784-48f6-8f13-a60ff73a5b3f\") " pod="openstack/dnsmasq-dns-67b789f86c-bbpr4" Oct 05 21:16:53 crc kubenswrapper[4754]: I1005 21:16:53.179845 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3e67b2a1-6784-48f6-8f13-a60ff73a5b3f-config\") pod \"dnsmasq-dns-67b789f86c-bbpr4\" (UID: \"3e67b2a1-6784-48f6-8f13-a60ff73a5b3f\") " pod="openstack/dnsmasq-dns-67b789f86c-bbpr4" Oct 05 21:16:53 crc kubenswrapper[4754]: I1005 21:16:53.180751 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3e67b2a1-6784-48f6-8f13-a60ff73a5b3f-config\") pod \"dnsmasq-dns-67b789f86c-bbpr4\" (UID: \"3e67b2a1-6784-48f6-8f13-a60ff73a5b3f\") " pod="openstack/dnsmasq-dns-67b789f86c-bbpr4" Oct 05 21:16:53 crc kubenswrapper[4754]: I1005 21:16:53.181325 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3e67b2a1-6784-48f6-8f13-a60ff73a5b3f-dns-swift-storage-0\") pod \"dnsmasq-dns-67b789f86c-bbpr4\" (UID: \"3e67b2a1-6784-48f6-8f13-a60ff73a5b3f\") " pod="openstack/dnsmasq-dns-67b789f86c-bbpr4" Oct 05 21:16:53 crc kubenswrapper[4754]: I1005 21:16:53.181919 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3e67b2a1-6784-48f6-8f13-a60ff73a5b3f-dns-svc\") pod \"dnsmasq-dns-67b789f86c-bbpr4\" (UID: \"3e67b2a1-6784-48f6-8f13-a60ff73a5b3f\") " pod="openstack/dnsmasq-dns-67b789f86c-bbpr4" Oct 05 21:16:53 crc kubenswrapper[4754]: I1005 21:16:53.182512 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3e67b2a1-6784-48f6-8f13-a60ff73a5b3f-ovsdbserver-sb\") pod \"dnsmasq-dns-67b789f86c-bbpr4\" (UID: \"3e67b2a1-6784-48f6-8f13-a60ff73a5b3f\") " pod="openstack/dnsmasq-dns-67b789f86c-bbpr4" Oct 05 21:16:53 crc kubenswrapper[4754]: I1005 21:16:53.183043 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/3e67b2a1-6784-48f6-8f13-a60ff73a5b3f-openstack-edpm-ipam\") pod \"dnsmasq-dns-67b789f86c-bbpr4\" (UID: \"3e67b2a1-6784-48f6-8f13-a60ff73a5b3f\") " pod="openstack/dnsmasq-dns-67b789f86c-bbpr4" Oct 05 21:16:53 crc kubenswrapper[4754]: I1005 21:16:53.184314 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3e67b2a1-6784-48f6-8f13-a60ff73a5b3f-ovsdbserver-nb\") pod \"dnsmasq-dns-67b789f86c-bbpr4\" (UID: \"3e67b2a1-6784-48f6-8f13-a60ff73a5b3f\") " pod="openstack/dnsmasq-dns-67b789f86c-bbpr4" Oct 05 21:16:53 crc kubenswrapper[4754]: I1005 21:16:53.206760 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2vpqr\" (UniqueName: \"kubernetes.io/projected/3e67b2a1-6784-48f6-8f13-a60ff73a5b3f-kube-api-access-2vpqr\") pod \"dnsmasq-dns-67b789f86c-bbpr4\" (UID: \"3e67b2a1-6784-48f6-8f13-a60ff73a5b3f\") " pod="openstack/dnsmasq-dns-67b789f86c-bbpr4" Oct 05 21:16:53 crc kubenswrapper[4754]: I1005 21:16:53.328797 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67b789f86c-bbpr4" Oct 05 21:16:53 crc kubenswrapper[4754]: W1005 21:16:53.622999 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3e67b2a1_6784_48f6_8f13_a60ff73a5b3f.slice/crio-e2bd1c256a38a6bfdb5f89bafface41ea1ac606775a28383dd19529f63324644 WatchSource:0}: Error finding container e2bd1c256a38a6bfdb5f89bafface41ea1ac606775a28383dd19529f63324644: Status 404 returned error can't find the container with id e2bd1c256a38a6bfdb5f89bafface41ea1ac606775a28383dd19529f63324644 Oct 05 21:16:53 crc kubenswrapper[4754]: I1005 21:16:53.623030 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-67b789f86c-bbpr4"] Oct 05 21:16:54 crc kubenswrapper[4754]: I1005 21:16:54.079218 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"3a6824aa-cdc6-4598-b296-1db23f99b156","Type":"ContainerStarted","Data":"4ad8f637848dd1728b42241aec6589b43fa4352d62b5dba2259c136c587941ea"} Oct 05 21:16:54 crc kubenswrapper[4754]: I1005 21:16:54.085336 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"7eb0b89b-e87d-4d5e-9939-d3a0f97f94db","Type":"ContainerStarted","Data":"ecd5fca83aaed92d1e37533328165ae0df1379998f7e6f877a27fbeb056dcf74"} Oct 05 21:16:54 crc kubenswrapper[4754]: I1005 21:16:54.086258 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67b789f86c-bbpr4" event={"ID":"3e67b2a1-6784-48f6-8f13-a60ff73a5b3f","Type":"ContainerStarted","Data":"e2bd1c256a38a6bfdb5f89bafface41ea1ac606775a28383dd19529f63324644"} Oct 05 21:16:55 crc kubenswrapper[4754]: I1005 21:16:55.101568 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"7eb0b89b-e87d-4d5e-9939-d3a0f97f94db","Type":"ContainerStarted","Data":"38617f8591cc6649098aaf91ad9211e78aea19ffaa457aff20d8f37df07fb8d8"} Oct 05 21:16:55 crc kubenswrapper[4754]: I1005 21:16:55.106030 4754 generic.go:334] "Generic (PLEG): container finished" podID="3e67b2a1-6784-48f6-8f13-a60ff73a5b3f" containerID="0d51a0adcbbed88886970f857a356b6e7accf7d5c7ddb5c8f98caf533b40ae7b" exitCode=0 Oct 05 21:16:55 crc kubenswrapper[4754]: I1005 21:16:55.106134 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67b789f86c-bbpr4" event={"ID":"3e67b2a1-6784-48f6-8f13-a60ff73a5b3f","Type":"ContainerDied","Data":"0d51a0adcbbed88886970f857a356b6e7accf7d5c7ddb5c8f98caf533b40ae7b"} Oct 05 21:16:55 crc kubenswrapper[4754]: I1005 21:16:55.108126 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"3a6824aa-cdc6-4598-b296-1db23f99b156","Type":"ContainerStarted","Data":"8dc41f4cc18f442d9b09b9936c4c7e170a521bc1ec1b8a9f21546c137289a2fe"} Oct 05 21:16:56 crc kubenswrapper[4754]: I1005 21:16:56.126662 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67b789f86c-bbpr4" event={"ID":"3e67b2a1-6784-48f6-8f13-a60ff73a5b3f","Type":"ContainerStarted","Data":"0b3f93cc0a4899bd008a42d368002672c09e6fd37390555b3d61d3879292666e"} Oct 05 21:16:56 crc kubenswrapper[4754]: I1005 21:16:56.166281 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-67b789f86c-bbpr4" podStartSLOduration=4.166254729 podStartE2EDuration="4.166254729s" podCreationTimestamp="2025-10-05 21:16:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 21:16:56.15867044 +0000 UTC m=+1340.062789220" watchObservedRunningTime="2025-10-05 21:16:56.166254729 +0000 UTC m=+1340.070373469" Oct 05 21:16:57 crc kubenswrapper[4754]: I1005 21:16:57.137709 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-67b789f86c-bbpr4" Oct 05 21:17:03 crc kubenswrapper[4754]: I1005 21:17:03.331694 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-67b789f86c-bbpr4" Oct 05 21:17:03 crc kubenswrapper[4754]: I1005 21:17:03.438743 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-jlwmm"] Oct 05 21:17:03 crc kubenswrapper[4754]: I1005 21:17:03.439108 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-59cf4bdb65-jlwmm" podUID="ceb7a5cb-e7f2-4162-bd2b-7518c3ad4d52" containerName="dnsmasq-dns" containerID="cri-o://047bf50f6e906707a0a4e6953b750c33bfcddcbcf227b31d1f3ccb0f26a6997c" gracePeriod=10 Oct 05 21:17:03 crc kubenswrapper[4754]: I1005 21:17:03.630297 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-79dc84bdb7-ngcb6"] Oct 05 21:17:03 crc kubenswrapper[4754]: I1005 21:17:03.632191 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79dc84bdb7-ngcb6" Oct 05 21:17:03 crc kubenswrapper[4754]: I1005 21:17:03.662678 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-79dc84bdb7-ngcb6"] Oct 05 21:17:03 crc kubenswrapper[4754]: I1005 21:17:03.722013 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f031b1f6-304e-4419-a4b2-97a8415a530a-dns-swift-storage-0\") pod \"dnsmasq-dns-79dc84bdb7-ngcb6\" (UID: \"f031b1f6-304e-4419-a4b2-97a8415a530a\") " pod="openstack/dnsmasq-dns-79dc84bdb7-ngcb6" Oct 05 21:17:03 crc kubenswrapper[4754]: I1005 21:17:03.722089 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f031b1f6-304e-4419-a4b2-97a8415a530a-config\") pod \"dnsmasq-dns-79dc84bdb7-ngcb6\" (UID: \"f031b1f6-304e-4419-a4b2-97a8415a530a\") " pod="openstack/dnsmasq-dns-79dc84bdb7-ngcb6" Oct 05 21:17:03 crc kubenswrapper[4754]: I1005 21:17:03.722121 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f031b1f6-304e-4419-a4b2-97a8415a530a-dns-svc\") pod \"dnsmasq-dns-79dc84bdb7-ngcb6\" (UID: \"f031b1f6-304e-4419-a4b2-97a8415a530a\") " pod="openstack/dnsmasq-dns-79dc84bdb7-ngcb6" Oct 05 21:17:03 crc kubenswrapper[4754]: I1005 21:17:03.722284 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9klff\" (UniqueName: \"kubernetes.io/projected/f031b1f6-304e-4419-a4b2-97a8415a530a-kube-api-access-9klff\") pod \"dnsmasq-dns-79dc84bdb7-ngcb6\" (UID: \"f031b1f6-304e-4419-a4b2-97a8415a530a\") " pod="openstack/dnsmasq-dns-79dc84bdb7-ngcb6" Oct 05 21:17:03 crc kubenswrapper[4754]: I1005 21:17:03.722347 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f031b1f6-304e-4419-a4b2-97a8415a530a-ovsdbserver-nb\") pod \"dnsmasq-dns-79dc84bdb7-ngcb6\" (UID: \"f031b1f6-304e-4419-a4b2-97a8415a530a\") " pod="openstack/dnsmasq-dns-79dc84bdb7-ngcb6" Oct 05 21:17:03 crc kubenswrapper[4754]: I1005 21:17:03.722395 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f031b1f6-304e-4419-a4b2-97a8415a530a-ovsdbserver-sb\") pod \"dnsmasq-dns-79dc84bdb7-ngcb6\" (UID: \"f031b1f6-304e-4419-a4b2-97a8415a530a\") " pod="openstack/dnsmasq-dns-79dc84bdb7-ngcb6" Oct 05 21:17:03 crc kubenswrapper[4754]: I1005 21:17:03.722489 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/f031b1f6-304e-4419-a4b2-97a8415a530a-openstack-edpm-ipam\") pod \"dnsmasq-dns-79dc84bdb7-ngcb6\" (UID: \"f031b1f6-304e-4419-a4b2-97a8415a530a\") " pod="openstack/dnsmasq-dns-79dc84bdb7-ngcb6" Oct 05 21:17:03 crc kubenswrapper[4754]: I1005 21:17:03.825350 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9klff\" (UniqueName: \"kubernetes.io/projected/f031b1f6-304e-4419-a4b2-97a8415a530a-kube-api-access-9klff\") pod \"dnsmasq-dns-79dc84bdb7-ngcb6\" (UID: \"f031b1f6-304e-4419-a4b2-97a8415a530a\") " pod="openstack/dnsmasq-dns-79dc84bdb7-ngcb6" Oct 05 21:17:03 crc kubenswrapper[4754]: I1005 21:17:03.825457 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f031b1f6-304e-4419-a4b2-97a8415a530a-ovsdbserver-nb\") pod \"dnsmasq-dns-79dc84bdb7-ngcb6\" (UID: \"f031b1f6-304e-4419-a4b2-97a8415a530a\") " pod="openstack/dnsmasq-dns-79dc84bdb7-ngcb6" Oct 05 21:17:03 crc kubenswrapper[4754]: I1005 21:17:03.825516 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f031b1f6-304e-4419-a4b2-97a8415a530a-ovsdbserver-sb\") pod \"dnsmasq-dns-79dc84bdb7-ngcb6\" (UID: \"f031b1f6-304e-4419-a4b2-97a8415a530a\") " pod="openstack/dnsmasq-dns-79dc84bdb7-ngcb6" Oct 05 21:17:03 crc kubenswrapper[4754]: I1005 21:17:03.825560 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/f031b1f6-304e-4419-a4b2-97a8415a530a-openstack-edpm-ipam\") pod \"dnsmasq-dns-79dc84bdb7-ngcb6\" (UID: \"f031b1f6-304e-4419-a4b2-97a8415a530a\") " pod="openstack/dnsmasq-dns-79dc84bdb7-ngcb6" Oct 05 21:17:03 crc kubenswrapper[4754]: I1005 21:17:03.825615 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f031b1f6-304e-4419-a4b2-97a8415a530a-dns-swift-storage-0\") pod \"dnsmasq-dns-79dc84bdb7-ngcb6\" (UID: \"f031b1f6-304e-4419-a4b2-97a8415a530a\") " pod="openstack/dnsmasq-dns-79dc84bdb7-ngcb6" Oct 05 21:17:03 crc kubenswrapper[4754]: I1005 21:17:03.825647 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f031b1f6-304e-4419-a4b2-97a8415a530a-config\") pod \"dnsmasq-dns-79dc84bdb7-ngcb6\" (UID: \"f031b1f6-304e-4419-a4b2-97a8415a530a\") " pod="openstack/dnsmasq-dns-79dc84bdb7-ngcb6" Oct 05 21:17:03 crc kubenswrapper[4754]: I1005 21:17:03.825668 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f031b1f6-304e-4419-a4b2-97a8415a530a-dns-svc\") pod \"dnsmasq-dns-79dc84bdb7-ngcb6\" (UID: \"f031b1f6-304e-4419-a4b2-97a8415a530a\") " pod="openstack/dnsmasq-dns-79dc84bdb7-ngcb6" Oct 05 21:17:03 crc kubenswrapper[4754]: I1005 21:17:03.827931 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f031b1f6-304e-4419-a4b2-97a8415a530a-ovsdbserver-sb\") pod \"dnsmasq-dns-79dc84bdb7-ngcb6\" (UID: \"f031b1f6-304e-4419-a4b2-97a8415a530a\") " pod="openstack/dnsmasq-dns-79dc84bdb7-ngcb6" Oct 05 21:17:03 crc kubenswrapper[4754]: I1005 21:17:03.828168 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f031b1f6-304e-4419-a4b2-97a8415a530a-dns-swift-storage-0\") pod \"dnsmasq-dns-79dc84bdb7-ngcb6\" (UID: \"f031b1f6-304e-4419-a4b2-97a8415a530a\") " pod="openstack/dnsmasq-dns-79dc84bdb7-ngcb6" Oct 05 21:17:03 crc kubenswrapper[4754]: I1005 21:17:03.828802 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/f031b1f6-304e-4419-a4b2-97a8415a530a-openstack-edpm-ipam\") pod \"dnsmasq-dns-79dc84bdb7-ngcb6\" (UID: \"f031b1f6-304e-4419-a4b2-97a8415a530a\") " pod="openstack/dnsmasq-dns-79dc84bdb7-ngcb6" Oct 05 21:17:03 crc kubenswrapper[4754]: I1005 21:17:03.829358 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f031b1f6-304e-4419-a4b2-97a8415a530a-config\") pod \"dnsmasq-dns-79dc84bdb7-ngcb6\" (UID: \"f031b1f6-304e-4419-a4b2-97a8415a530a\") " pod="openstack/dnsmasq-dns-79dc84bdb7-ngcb6" Oct 05 21:17:03 crc kubenswrapper[4754]: I1005 21:17:03.831863 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f031b1f6-304e-4419-a4b2-97a8415a530a-ovsdbserver-nb\") pod \"dnsmasq-dns-79dc84bdb7-ngcb6\" (UID: \"f031b1f6-304e-4419-a4b2-97a8415a530a\") " pod="openstack/dnsmasq-dns-79dc84bdb7-ngcb6" Oct 05 21:17:03 crc kubenswrapper[4754]: I1005 21:17:03.839400 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f031b1f6-304e-4419-a4b2-97a8415a530a-dns-svc\") pod \"dnsmasq-dns-79dc84bdb7-ngcb6\" (UID: \"f031b1f6-304e-4419-a4b2-97a8415a530a\") " pod="openstack/dnsmasq-dns-79dc84bdb7-ngcb6" Oct 05 21:17:03 crc kubenswrapper[4754]: I1005 21:17:03.876668 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9klff\" (UniqueName: \"kubernetes.io/projected/f031b1f6-304e-4419-a4b2-97a8415a530a-kube-api-access-9klff\") pod \"dnsmasq-dns-79dc84bdb7-ngcb6\" (UID: \"f031b1f6-304e-4419-a4b2-97a8415a530a\") " pod="openstack/dnsmasq-dns-79dc84bdb7-ngcb6" Oct 05 21:17:03 crc kubenswrapper[4754]: I1005 21:17:03.958198 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79dc84bdb7-ngcb6" Oct 05 21:17:04 crc kubenswrapper[4754]: I1005 21:17:04.130275 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59cf4bdb65-jlwmm" Oct 05 21:17:04 crc kubenswrapper[4754]: I1005 21:17:04.132361 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ceb7a5cb-e7f2-4162-bd2b-7518c3ad4d52-ovsdbserver-sb\") pod \"ceb7a5cb-e7f2-4162-bd2b-7518c3ad4d52\" (UID: \"ceb7a5cb-e7f2-4162-bd2b-7518c3ad4d52\") " Oct 05 21:17:04 crc kubenswrapper[4754]: I1005 21:17:04.132402 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ceb7a5cb-e7f2-4162-bd2b-7518c3ad4d52-config\") pod \"ceb7a5cb-e7f2-4162-bd2b-7518c3ad4d52\" (UID: \"ceb7a5cb-e7f2-4162-bd2b-7518c3ad4d52\") " Oct 05 21:17:04 crc kubenswrapper[4754]: I1005 21:17:04.132464 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ceb7a5cb-e7f2-4162-bd2b-7518c3ad4d52-dns-svc\") pod \"ceb7a5cb-e7f2-4162-bd2b-7518c3ad4d52\" (UID: \"ceb7a5cb-e7f2-4162-bd2b-7518c3ad4d52\") " Oct 05 21:17:04 crc kubenswrapper[4754]: I1005 21:17:04.132515 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ceb7a5cb-e7f2-4162-bd2b-7518c3ad4d52-ovsdbserver-nb\") pod \"ceb7a5cb-e7f2-4162-bd2b-7518c3ad4d52\" (UID: \"ceb7a5cb-e7f2-4162-bd2b-7518c3ad4d52\") " Oct 05 21:17:04 crc kubenswrapper[4754]: I1005 21:17:04.132566 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ceb7a5cb-e7f2-4162-bd2b-7518c3ad4d52-dns-swift-storage-0\") pod \"ceb7a5cb-e7f2-4162-bd2b-7518c3ad4d52\" (UID: \"ceb7a5cb-e7f2-4162-bd2b-7518c3ad4d52\") " Oct 05 21:17:04 crc kubenswrapper[4754]: I1005 21:17:04.132606 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zb949\" (UniqueName: \"kubernetes.io/projected/ceb7a5cb-e7f2-4162-bd2b-7518c3ad4d52-kube-api-access-zb949\") pod \"ceb7a5cb-e7f2-4162-bd2b-7518c3ad4d52\" (UID: \"ceb7a5cb-e7f2-4162-bd2b-7518c3ad4d52\") " Oct 05 21:17:04 crc kubenswrapper[4754]: I1005 21:17:04.148890 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ceb7a5cb-e7f2-4162-bd2b-7518c3ad4d52-kube-api-access-zb949" (OuterVolumeSpecName: "kube-api-access-zb949") pod "ceb7a5cb-e7f2-4162-bd2b-7518c3ad4d52" (UID: "ceb7a5cb-e7f2-4162-bd2b-7518c3ad4d52"). InnerVolumeSpecName "kube-api-access-zb949". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:17:04 crc kubenswrapper[4754]: I1005 21:17:04.212460 4754 generic.go:334] "Generic (PLEG): container finished" podID="ceb7a5cb-e7f2-4162-bd2b-7518c3ad4d52" containerID="047bf50f6e906707a0a4e6953b750c33bfcddcbcf227b31d1f3ccb0f26a6997c" exitCode=0 Oct 05 21:17:04 crc kubenswrapper[4754]: I1005 21:17:04.212513 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cf4bdb65-jlwmm" event={"ID":"ceb7a5cb-e7f2-4162-bd2b-7518c3ad4d52","Type":"ContainerDied","Data":"047bf50f6e906707a0a4e6953b750c33bfcddcbcf227b31d1f3ccb0f26a6997c"} Oct 05 21:17:04 crc kubenswrapper[4754]: I1005 21:17:04.212540 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cf4bdb65-jlwmm" event={"ID":"ceb7a5cb-e7f2-4162-bd2b-7518c3ad4d52","Type":"ContainerDied","Data":"04bce3bfe7f2909d600d2f876cfed3d07c0847721902ce559cdc63be292a0a1c"} Oct 05 21:17:04 crc kubenswrapper[4754]: I1005 21:17:04.212557 4754 scope.go:117] "RemoveContainer" containerID="047bf50f6e906707a0a4e6953b750c33bfcddcbcf227b31d1f3ccb0f26a6997c" Oct 05 21:17:04 crc kubenswrapper[4754]: I1005 21:17:04.212662 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59cf4bdb65-jlwmm" Oct 05 21:17:04 crc kubenswrapper[4754]: I1005 21:17:04.222214 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ceb7a5cb-e7f2-4162-bd2b-7518c3ad4d52-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "ceb7a5cb-e7f2-4162-bd2b-7518c3ad4d52" (UID: "ceb7a5cb-e7f2-4162-bd2b-7518c3ad4d52"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 21:17:04 crc kubenswrapper[4754]: I1005 21:17:04.231510 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ceb7a5cb-e7f2-4162-bd2b-7518c3ad4d52-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ceb7a5cb-e7f2-4162-bd2b-7518c3ad4d52" (UID: "ceb7a5cb-e7f2-4162-bd2b-7518c3ad4d52"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 21:17:04 crc kubenswrapper[4754]: I1005 21:17:04.239786 4754 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ceb7a5cb-e7f2-4162-bd2b-7518c3ad4d52-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 05 21:17:04 crc kubenswrapper[4754]: I1005 21:17:04.239833 4754 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ceb7a5cb-e7f2-4162-bd2b-7518c3ad4d52-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 05 21:17:04 crc kubenswrapper[4754]: I1005 21:17:04.239843 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zb949\" (UniqueName: \"kubernetes.io/projected/ceb7a5cb-e7f2-4162-bd2b-7518c3ad4d52-kube-api-access-zb949\") on node \"crc\" DevicePath \"\"" Oct 05 21:17:04 crc kubenswrapper[4754]: I1005 21:17:04.248427 4754 scope.go:117] "RemoveContainer" containerID="f871f68953da0289a85d10ec6082b36f55672e3b4a72fc715962828c8d78ade1" Oct 05 21:17:04 crc kubenswrapper[4754]: I1005 21:17:04.250840 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ceb7a5cb-e7f2-4162-bd2b-7518c3ad4d52-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "ceb7a5cb-e7f2-4162-bd2b-7518c3ad4d52" (UID: "ceb7a5cb-e7f2-4162-bd2b-7518c3ad4d52"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 21:17:04 crc kubenswrapper[4754]: I1005 21:17:04.258895 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ceb7a5cb-e7f2-4162-bd2b-7518c3ad4d52-config" (OuterVolumeSpecName: "config") pod "ceb7a5cb-e7f2-4162-bd2b-7518c3ad4d52" (UID: "ceb7a5cb-e7f2-4162-bd2b-7518c3ad4d52"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 21:17:04 crc kubenswrapper[4754]: I1005 21:17:04.277907 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ceb7a5cb-e7f2-4162-bd2b-7518c3ad4d52-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "ceb7a5cb-e7f2-4162-bd2b-7518c3ad4d52" (UID: "ceb7a5cb-e7f2-4162-bd2b-7518c3ad4d52"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 21:17:04 crc kubenswrapper[4754]: I1005 21:17:04.278946 4754 scope.go:117] "RemoveContainer" containerID="047bf50f6e906707a0a4e6953b750c33bfcddcbcf227b31d1f3ccb0f26a6997c" Oct 05 21:17:04 crc kubenswrapper[4754]: E1005 21:17:04.282801 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"047bf50f6e906707a0a4e6953b750c33bfcddcbcf227b31d1f3ccb0f26a6997c\": container with ID starting with 047bf50f6e906707a0a4e6953b750c33bfcddcbcf227b31d1f3ccb0f26a6997c not found: ID does not exist" containerID="047bf50f6e906707a0a4e6953b750c33bfcddcbcf227b31d1f3ccb0f26a6997c" Oct 05 21:17:04 crc kubenswrapper[4754]: I1005 21:17:04.282839 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"047bf50f6e906707a0a4e6953b750c33bfcddcbcf227b31d1f3ccb0f26a6997c"} err="failed to get container status \"047bf50f6e906707a0a4e6953b750c33bfcddcbcf227b31d1f3ccb0f26a6997c\": rpc error: code = NotFound desc = could not find container \"047bf50f6e906707a0a4e6953b750c33bfcddcbcf227b31d1f3ccb0f26a6997c\": container with ID starting with 047bf50f6e906707a0a4e6953b750c33bfcddcbcf227b31d1f3ccb0f26a6997c not found: ID does not exist" Oct 05 21:17:04 crc kubenswrapper[4754]: I1005 21:17:04.282866 4754 scope.go:117] "RemoveContainer" containerID="f871f68953da0289a85d10ec6082b36f55672e3b4a72fc715962828c8d78ade1" Oct 05 21:17:04 crc kubenswrapper[4754]: E1005 21:17:04.283248 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f871f68953da0289a85d10ec6082b36f55672e3b4a72fc715962828c8d78ade1\": container with ID starting with f871f68953da0289a85d10ec6082b36f55672e3b4a72fc715962828c8d78ade1 not found: ID does not exist" containerID="f871f68953da0289a85d10ec6082b36f55672e3b4a72fc715962828c8d78ade1" Oct 05 21:17:04 crc kubenswrapper[4754]: I1005 21:17:04.283299 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f871f68953da0289a85d10ec6082b36f55672e3b4a72fc715962828c8d78ade1"} err="failed to get container status \"f871f68953da0289a85d10ec6082b36f55672e3b4a72fc715962828c8d78ade1\": rpc error: code = NotFound desc = could not find container \"f871f68953da0289a85d10ec6082b36f55672e3b4a72fc715962828c8d78ade1\": container with ID starting with f871f68953da0289a85d10ec6082b36f55672e3b4a72fc715962828c8d78ade1 not found: ID does not exist" Oct 05 21:17:04 crc kubenswrapper[4754]: I1005 21:17:04.341051 4754 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ceb7a5cb-e7f2-4162-bd2b-7518c3ad4d52-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 05 21:17:04 crc kubenswrapper[4754]: I1005 21:17:04.341088 4754 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ceb7a5cb-e7f2-4162-bd2b-7518c3ad4d52-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 05 21:17:04 crc kubenswrapper[4754]: I1005 21:17:04.341099 4754 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ceb7a5cb-e7f2-4162-bd2b-7518c3ad4d52-config\") on node \"crc\" DevicePath \"\"" Oct 05 21:17:04 crc kubenswrapper[4754]: I1005 21:17:04.554173 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-jlwmm"] Oct 05 21:17:04 crc kubenswrapper[4754]: I1005 21:17:04.559017 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-jlwmm"] Oct 05 21:17:04 crc kubenswrapper[4754]: I1005 21:17:04.599504 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-79dc84bdb7-ngcb6"] Oct 05 21:17:04 crc kubenswrapper[4754]: I1005 21:17:04.848591 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ceb7a5cb-e7f2-4162-bd2b-7518c3ad4d52" path="/var/lib/kubelet/pods/ceb7a5cb-e7f2-4162-bd2b-7518c3ad4d52/volumes" Oct 05 21:17:05 crc kubenswrapper[4754]: I1005 21:17:05.227052 4754 generic.go:334] "Generic (PLEG): container finished" podID="f031b1f6-304e-4419-a4b2-97a8415a530a" containerID="ae816ec55caf3eaa6987d5cbe5a056cf602dec12070f3706b0d5a69907844daa" exitCode=0 Oct 05 21:17:05 crc kubenswrapper[4754]: I1005 21:17:05.227107 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79dc84bdb7-ngcb6" event={"ID":"f031b1f6-304e-4419-a4b2-97a8415a530a","Type":"ContainerDied","Data":"ae816ec55caf3eaa6987d5cbe5a056cf602dec12070f3706b0d5a69907844daa"} Oct 05 21:17:05 crc kubenswrapper[4754]: I1005 21:17:05.227173 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79dc84bdb7-ngcb6" event={"ID":"f031b1f6-304e-4419-a4b2-97a8415a530a","Type":"ContainerStarted","Data":"77fe2a494808bf1102c060959ea45e9969de9ea7c41100d27929d960b1748d47"} Oct 05 21:17:05 crc kubenswrapper[4754]: I1005 21:17:05.244941 4754 patch_prober.go:28] interesting pod/machine-config-daemon-b2h9k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 21:17:05 crc kubenswrapper[4754]: I1005 21:17:05.245007 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 21:17:06 crc kubenswrapper[4754]: I1005 21:17:06.238689 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79dc84bdb7-ngcb6" event={"ID":"f031b1f6-304e-4419-a4b2-97a8415a530a","Type":"ContainerStarted","Data":"f559e75cdab214c73ccd72f114ccff960afdc375b7f548f95a1993389af43ceb"} Oct 05 21:17:06 crc kubenswrapper[4754]: I1005 21:17:06.239099 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-79dc84bdb7-ngcb6" Oct 05 21:17:06 crc kubenswrapper[4754]: I1005 21:17:06.267121 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-79dc84bdb7-ngcb6" podStartSLOduration=3.267095195 podStartE2EDuration="3.267095195s" podCreationTimestamp="2025-10-05 21:17:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 21:17:06.26538387 +0000 UTC m=+1350.169502600" watchObservedRunningTime="2025-10-05 21:17:06.267095195 +0000 UTC m=+1350.171213925" Oct 05 21:17:13 crc kubenswrapper[4754]: I1005 21:17:13.960691 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-79dc84bdb7-ngcb6" Oct 05 21:17:14 crc kubenswrapper[4754]: I1005 21:17:14.057672 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-67b789f86c-bbpr4"] Oct 05 21:17:14 crc kubenswrapper[4754]: I1005 21:17:14.057983 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-67b789f86c-bbpr4" podUID="3e67b2a1-6784-48f6-8f13-a60ff73a5b3f" containerName="dnsmasq-dns" containerID="cri-o://0b3f93cc0a4899bd008a42d368002672c09e6fd37390555b3d61d3879292666e" gracePeriod=10 Oct 05 21:17:14 crc kubenswrapper[4754]: I1005 21:17:14.339481 4754 generic.go:334] "Generic (PLEG): container finished" podID="3e67b2a1-6784-48f6-8f13-a60ff73a5b3f" containerID="0b3f93cc0a4899bd008a42d368002672c09e6fd37390555b3d61d3879292666e" exitCode=0 Oct 05 21:17:14 crc kubenswrapper[4754]: I1005 21:17:14.339866 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67b789f86c-bbpr4" event={"ID":"3e67b2a1-6784-48f6-8f13-a60ff73a5b3f","Type":"ContainerDied","Data":"0b3f93cc0a4899bd008a42d368002672c09e6fd37390555b3d61d3879292666e"} Oct 05 21:17:14 crc kubenswrapper[4754]: I1005 21:17:14.553933 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67b789f86c-bbpr4" Oct 05 21:17:14 crc kubenswrapper[4754]: I1005 21:17:14.682531 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2vpqr\" (UniqueName: \"kubernetes.io/projected/3e67b2a1-6784-48f6-8f13-a60ff73a5b3f-kube-api-access-2vpqr\") pod \"3e67b2a1-6784-48f6-8f13-a60ff73a5b3f\" (UID: \"3e67b2a1-6784-48f6-8f13-a60ff73a5b3f\") " Oct 05 21:17:14 crc kubenswrapper[4754]: I1005 21:17:14.682600 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3e67b2a1-6784-48f6-8f13-a60ff73a5b3f-dns-svc\") pod \"3e67b2a1-6784-48f6-8f13-a60ff73a5b3f\" (UID: \"3e67b2a1-6784-48f6-8f13-a60ff73a5b3f\") " Oct 05 21:17:14 crc kubenswrapper[4754]: I1005 21:17:14.682634 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3e67b2a1-6784-48f6-8f13-a60ff73a5b3f-ovsdbserver-nb\") pod \"3e67b2a1-6784-48f6-8f13-a60ff73a5b3f\" (UID: \"3e67b2a1-6784-48f6-8f13-a60ff73a5b3f\") " Oct 05 21:17:14 crc kubenswrapper[4754]: I1005 21:17:14.682700 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3e67b2a1-6784-48f6-8f13-a60ff73a5b3f-ovsdbserver-sb\") pod \"3e67b2a1-6784-48f6-8f13-a60ff73a5b3f\" (UID: \"3e67b2a1-6784-48f6-8f13-a60ff73a5b3f\") " Oct 05 21:17:14 crc kubenswrapper[4754]: I1005 21:17:14.682795 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/3e67b2a1-6784-48f6-8f13-a60ff73a5b3f-openstack-edpm-ipam\") pod \"3e67b2a1-6784-48f6-8f13-a60ff73a5b3f\" (UID: \"3e67b2a1-6784-48f6-8f13-a60ff73a5b3f\") " Oct 05 21:17:14 crc kubenswrapper[4754]: I1005 21:17:14.683625 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3e67b2a1-6784-48f6-8f13-a60ff73a5b3f-config\") pod \"3e67b2a1-6784-48f6-8f13-a60ff73a5b3f\" (UID: \"3e67b2a1-6784-48f6-8f13-a60ff73a5b3f\") " Oct 05 21:17:14 crc kubenswrapper[4754]: I1005 21:17:14.683738 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3e67b2a1-6784-48f6-8f13-a60ff73a5b3f-dns-swift-storage-0\") pod \"3e67b2a1-6784-48f6-8f13-a60ff73a5b3f\" (UID: \"3e67b2a1-6784-48f6-8f13-a60ff73a5b3f\") " Oct 05 21:17:14 crc kubenswrapper[4754]: I1005 21:17:14.706053 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3e67b2a1-6784-48f6-8f13-a60ff73a5b3f-kube-api-access-2vpqr" (OuterVolumeSpecName: "kube-api-access-2vpqr") pod "3e67b2a1-6784-48f6-8f13-a60ff73a5b3f" (UID: "3e67b2a1-6784-48f6-8f13-a60ff73a5b3f"). InnerVolumeSpecName "kube-api-access-2vpqr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:17:14 crc kubenswrapper[4754]: I1005 21:17:14.759454 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3e67b2a1-6784-48f6-8f13-a60ff73a5b3f-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "3e67b2a1-6784-48f6-8f13-a60ff73a5b3f" (UID: "3e67b2a1-6784-48f6-8f13-a60ff73a5b3f"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 21:17:14 crc kubenswrapper[4754]: I1005 21:17:14.762719 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3e67b2a1-6784-48f6-8f13-a60ff73a5b3f-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "3e67b2a1-6784-48f6-8f13-a60ff73a5b3f" (UID: "3e67b2a1-6784-48f6-8f13-a60ff73a5b3f"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 21:17:14 crc kubenswrapper[4754]: I1005 21:17:14.767188 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3e67b2a1-6784-48f6-8f13-a60ff73a5b3f-config" (OuterVolumeSpecName: "config") pod "3e67b2a1-6784-48f6-8f13-a60ff73a5b3f" (UID: "3e67b2a1-6784-48f6-8f13-a60ff73a5b3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 21:17:14 crc kubenswrapper[4754]: I1005 21:17:14.774035 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3e67b2a1-6784-48f6-8f13-a60ff73a5b3f-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "3e67b2a1-6784-48f6-8f13-a60ff73a5b3f" (UID: "3e67b2a1-6784-48f6-8f13-a60ff73a5b3f"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 21:17:14 crc kubenswrapper[4754]: I1005 21:17:14.787449 4754 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3e67b2a1-6784-48f6-8f13-a60ff73a5b3f-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 05 21:17:14 crc kubenswrapper[4754]: I1005 21:17:14.787570 4754 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3e67b2a1-6784-48f6-8f13-a60ff73a5b3f-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 05 21:17:14 crc kubenswrapper[4754]: I1005 21:17:14.794924 4754 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3e67b2a1-6784-48f6-8f13-a60ff73a5b3f-config\") on node \"crc\" DevicePath \"\"" Oct 05 21:17:14 crc kubenswrapper[4754]: I1005 21:17:14.794971 4754 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3e67b2a1-6784-48f6-8f13-a60ff73a5b3f-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 05 21:17:14 crc kubenswrapper[4754]: I1005 21:17:14.795020 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2vpqr\" (UniqueName: \"kubernetes.io/projected/3e67b2a1-6784-48f6-8f13-a60ff73a5b3f-kube-api-access-2vpqr\") on node \"crc\" DevicePath \"\"" Oct 05 21:17:14 crc kubenswrapper[4754]: I1005 21:17:14.792844 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3e67b2a1-6784-48f6-8f13-a60ff73a5b3f-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "3e67b2a1-6784-48f6-8f13-a60ff73a5b3f" (UID: "3e67b2a1-6784-48f6-8f13-a60ff73a5b3f"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 21:17:14 crc kubenswrapper[4754]: I1005 21:17:14.799549 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3e67b2a1-6784-48f6-8f13-a60ff73a5b3f-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "3e67b2a1-6784-48f6-8f13-a60ff73a5b3f" (UID: "3e67b2a1-6784-48f6-8f13-a60ff73a5b3f"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 21:17:14 crc kubenswrapper[4754]: I1005 21:17:14.899310 4754 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3e67b2a1-6784-48f6-8f13-a60ff73a5b3f-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 05 21:17:14 crc kubenswrapper[4754]: I1005 21:17:14.899350 4754 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/3e67b2a1-6784-48f6-8f13-a60ff73a5b3f-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Oct 05 21:17:15 crc kubenswrapper[4754]: I1005 21:17:15.362596 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67b789f86c-bbpr4" event={"ID":"3e67b2a1-6784-48f6-8f13-a60ff73a5b3f","Type":"ContainerDied","Data":"e2bd1c256a38a6bfdb5f89bafface41ea1ac606775a28383dd19529f63324644"} Oct 05 21:17:15 crc kubenswrapper[4754]: I1005 21:17:15.363002 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67b789f86c-bbpr4" Oct 05 21:17:15 crc kubenswrapper[4754]: I1005 21:17:15.363024 4754 scope.go:117] "RemoveContainer" containerID="0b3f93cc0a4899bd008a42d368002672c09e6fd37390555b3d61d3879292666e" Oct 05 21:17:15 crc kubenswrapper[4754]: I1005 21:17:15.405532 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-67b789f86c-bbpr4"] Oct 05 21:17:15 crc kubenswrapper[4754]: I1005 21:17:15.417254 4754 scope.go:117] "RemoveContainer" containerID="0d51a0adcbbed88886970f857a356b6e7accf7d5c7ddb5c8f98caf533b40ae7b" Oct 05 21:17:15 crc kubenswrapper[4754]: I1005 21:17:15.432934 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-67b789f86c-bbpr4"] Oct 05 21:17:16 crc kubenswrapper[4754]: I1005 21:17:16.858754 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3e67b2a1-6784-48f6-8f13-a60ff73a5b3f" path="/var/lib/kubelet/pods/3e67b2a1-6784-48f6-8f13-a60ff73a5b3f/volumes" Oct 05 21:17:28 crc kubenswrapper[4754]: I1005 21:17:28.505117 4754 generic.go:334] "Generic (PLEG): container finished" podID="3a6824aa-cdc6-4598-b296-1db23f99b156" containerID="8dc41f4cc18f442d9b09b9936c4c7e170a521bc1ec1b8a9f21546c137289a2fe" exitCode=0 Oct 05 21:17:28 crc kubenswrapper[4754]: I1005 21:17:28.505831 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"3a6824aa-cdc6-4598-b296-1db23f99b156","Type":"ContainerDied","Data":"8dc41f4cc18f442d9b09b9936c4c7e170a521bc1ec1b8a9f21546c137289a2fe"} Oct 05 21:17:28 crc kubenswrapper[4754]: I1005 21:17:28.511100 4754 generic.go:334] "Generic (PLEG): container finished" podID="7eb0b89b-e87d-4d5e-9939-d3a0f97f94db" containerID="38617f8591cc6649098aaf91ad9211e78aea19ffaa457aff20d8f37df07fb8d8" exitCode=0 Oct 05 21:17:28 crc kubenswrapper[4754]: I1005 21:17:28.511186 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"7eb0b89b-e87d-4d5e-9939-d3a0f97f94db","Type":"ContainerDied","Data":"38617f8591cc6649098aaf91ad9211e78aea19ffaa457aff20d8f37df07fb8d8"} Oct 05 21:17:29 crc kubenswrapper[4754]: I1005 21:17:29.529567 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"3a6824aa-cdc6-4598-b296-1db23f99b156","Type":"ContainerStarted","Data":"a5e9c2bc21d1ee8bcedb8edb409256a9cf471e372f6ad883b4bda1fd7f47aff9"} Oct 05 21:17:29 crc kubenswrapper[4754]: I1005 21:17:29.530156 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Oct 05 21:17:29 crc kubenswrapper[4754]: I1005 21:17:29.532637 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"7eb0b89b-e87d-4d5e-9939-d3a0f97f94db","Type":"ContainerStarted","Data":"15ce6c628c0ea7a001517a64eb08eacb7ea69b3a750e9289d88b1b1d7abcc2f9"} Oct 05 21:17:29 crc kubenswrapper[4754]: I1005 21:17:29.532856 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Oct 05 21:17:29 crc kubenswrapper[4754]: I1005 21:17:29.563444 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=37.563425421 podStartE2EDuration="37.563425421s" podCreationTimestamp="2025-10-05 21:16:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 21:17:29.556732275 +0000 UTC m=+1373.460851015" watchObservedRunningTime="2025-10-05 21:17:29.563425421 +0000 UTC m=+1373.467544141" Oct 05 21:17:29 crc kubenswrapper[4754]: I1005 21:17:29.582905 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=37.58288418 podStartE2EDuration="37.58288418s" podCreationTimestamp="2025-10-05 21:16:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 21:17:29.575859136 +0000 UTC m=+1373.479977856" watchObservedRunningTime="2025-10-05 21:17:29.58288418 +0000 UTC m=+1373.487002890" Oct 05 21:17:33 crc kubenswrapper[4754]: I1005 21:17:33.004883 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-bwlth"] Oct 05 21:17:33 crc kubenswrapper[4754]: E1005 21:17:33.006562 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ceb7a5cb-e7f2-4162-bd2b-7518c3ad4d52" containerName="dnsmasq-dns" Oct 05 21:17:33 crc kubenswrapper[4754]: I1005 21:17:33.006634 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="ceb7a5cb-e7f2-4162-bd2b-7518c3ad4d52" containerName="dnsmasq-dns" Oct 05 21:17:33 crc kubenswrapper[4754]: E1005 21:17:33.006703 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ceb7a5cb-e7f2-4162-bd2b-7518c3ad4d52" containerName="init" Oct 05 21:17:33 crc kubenswrapper[4754]: I1005 21:17:33.006754 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="ceb7a5cb-e7f2-4162-bd2b-7518c3ad4d52" containerName="init" Oct 05 21:17:33 crc kubenswrapper[4754]: E1005 21:17:33.006820 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e67b2a1-6784-48f6-8f13-a60ff73a5b3f" containerName="dnsmasq-dns" Oct 05 21:17:33 crc kubenswrapper[4754]: I1005 21:17:33.006873 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e67b2a1-6784-48f6-8f13-a60ff73a5b3f" containerName="dnsmasq-dns" Oct 05 21:17:33 crc kubenswrapper[4754]: E1005 21:17:33.006933 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e67b2a1-6784-48f6-8f13-a60ff73a5b3f" containerName="init" Oct 05 21:17:33 crc kubenswrapper[4754]: I1005 21:17:33.006982 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e67b2a1-6784-48f6-8f13-a60ff73a5b3f" containerName="init" Oct 05 21:17:33 crc kubenswrapper[4754]: I1005 21:17:33.007191 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="ceb7a5cb-e7f2-4162-bd2b-7518c3ad4d52" containerName="dnsmasq-dns" Oct 05 21:17:33 crc kubenswrapper[4754]: I1005 21:17:33.007320 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e67b2a1-6784-48f6-8f13-a60ff73a5b3f" containerName="dnsmasq-dns" Oct 05 21:17:33 crc kubenswrapper[4754]: I1005 21:17:33.008175 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-bwlth" Oct 05 21:17:33 crc kubenswrapper[4754]: I1005 21:17:33.012106 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-v54jd" Oct 05 21:17:33 crc kubenswrapper[4754]: I1005 21:17:33.012352 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 05 21:17:33 crc kubenswrapper[4754]: I1005 21:17:33.012550 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 05 21:17:33 crc kubenswrapper[4754]: I1005 21:17:33.012656 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 05 21:17:33 crc kubenswrapper[4754]: I1005 21:17:33.044538 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-bwlth"] Oct 05 21:17:33 crc kubenswrapper[4754]: I1005 21:17:33.087383 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/be10d33f-8bb1-4157-9a5f-1f5bbee194f0-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-bwlth\" (UID: \"be10d33f-8bb1-4157-9a5f-1f5bbee194f0\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-bwlth" Oct 05 21:17:33 crc kubenswrapper[4754]: I1005 21:17:33.087445 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d6hqk\" (UniqueName: \"kubernetes.io/projected/be10d33f-8bb1-4157-9a5f-1f5bbee194f0-kube-api-access-d6hqk\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-bwlth\" (UID: \"be10d33f-8bb1-4157-9a5f-1f5bbee194f0\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-bwlth" Oct 05 21:17:33 crc kubenswrapper[4754]: I1005 21:17:33.087687 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/be10d33f-8bb1-4157-9a5f-1f5bbee194f0-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-bwlth\" (UID: \"be10d33f-8bb1-4157-9a5f-1f5bbee194f0\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-bwlth" Oct 05 21:17:33 crc kubenswrapper[4754]: I1005 21:17:33.088029 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be10d33f-8bb1-4157-9a5f-1f5bbee194f0-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-bwlth\" (UID: \"be10d33f-8bb1-4157-9a5f-1f5bbee194f0\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-bwlth" Oct 05 21:17:33 crc kubenswrapper[4754]: I1005 21:17:33.190222 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be10d33f-8bb1-4157-9a5f-1f5bbee194f0-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-bwlth\" (UID: \"be10d33f-8bb1-4157-9a5f-1f5bbee194f0\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-bwlth" Oct 05 21:17:33 crc kubenswrapper[4754]: I1005 21:17:33.190309 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/be10d33f-8bb1-4157-9a5f-1f5bbee194f0-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-bwlth\" (UID: \"be10d33f-8bb1-4157-9a5f-1f5bbee194f0\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-bwlth" Oct 05 21:17:33 crc kubenswrapper[4754]: I1005 21:17:33.190345 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d6hqk\" (UniqueName: \"kubernetes.io/projected/be10d33f-8bb1-4157-9a5f-1f5bbee194f0-kube-api-access-d6hqk\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-bwlth\" (UID: \"be10d33f-8bb1-4157-9a5f-1f5bbee194f0\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-bwlth" Oct 05 21:17:33 crc kubenswrapper[4754]: I1005 21:17:33.190385 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/be10d33f-8bb1-4157-9a5f-1f5bbee194f0-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-bwlth\" (UID: \"be10d33f-8bb1-4157-9a5f-1f5bbee194f0\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-bwlth" Oct 05 21:17:33 crc kubenswrapper[4754]: I1005 21:17:33.195967 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be10d33f-8bb1-4157-9a5f-1f5bbee194f0-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-bwlth\" (UID: \"be10d33f-8bb1-4157-9a5f-1f5bbee194f0\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-bwlth" Oct 05 21:17:33 crc kubenswrapper[4754]: I1005 21:17:33.196798 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/be10d33f-8bb1-4157-9a5f-1f5bbee194f0-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-bwlth\" (UID: \"be10d33f-8bb1-4157-9a5f-1f5bbee194f0\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-bwlth" Oct 05 21:17:33 crc kubenswrapper[4754]: I1005 21:17:33.197667 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/be10d33f-8bb1-4157-9a5f-1f5bbee194f0-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-bwlth\" (UID: \"be10d33f-8bb1-4157-9a5f-1f5bbee194f0\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-bwlth" Oct 05 21:17:33 crc kubenswrapper[4754]: I1005 21:17:33.223627 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d6hqk\" (UniqueName: \"kubernetes.io/projected/be10d33f-8bb1-4157-9a5f-1f5bbee194f0-kube-api-access-d6hqk\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-bwlth\" (UID: \"be10d33f-8bb1-4157-9a5f-1f5bbee194f0\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-bwlth" Oct 05 21:17:33 crc kubenswrapper[4754]: I1005 21:17:33.332082 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-bwlth" Oct 05 21:17:34 crc kubenswrapper[4754]: I1005 21:17:34.018716 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-bwlth"] Oct 05 21:17:34 crc kubenswrapper[4754]: I1005 21:17:34.577568 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-bwlth" event={"ID":"be10d33f-8bb1-4157-9a5f-1f5bbee194f0","Type":"ContainerStarted","Data":"2034622146a3be6894a1e534b23f1cbd1b11adb1d787ad927faa6a1d1a0c7011"} Oct 05 21:17:35 crc kubenswrapper[4754]: I1005 21:17:35.244830 4754 patch_prober.go:28] interesting pod/machine-config-daemon-b2h9k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 21:17:35 crc kubenswrapper[4754]: I1005 21:17:35.244880 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 21:17:35 crc kubenswrapper[4754]: I1005 21:17:35.244914 4754 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" Oct 05 21:17:35 crc kubenswrapper[4754]: I1005 21:17:35.245623 4754 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a2229b13a2697fd212215ae5d196911669bc92f36ce70525da6f97d702c807b9"} pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 05 21:17:35 crc kubenswrapper[4754]: I1005 21:17:35.245672 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" containerName="machine-config-daemon" containerID="cri-o://a2229b13a2697fd212215ae5d196911669bc92f36ce70525da6f97d702c807b9" gracePeriod=600 Oct 05 21:17:35 crc kubenswrapper[4754]: I1005 21:17:35.595366 4754 generic.go:334] "Generic (PLEG): container finished" podID="be95a413-6a59-45b4-84b7-b43cae694a26" containerID="a2229b13a2697fd212215ae5d196911669bc92f36ce70525da6f97d702c807b9" exitCode=0 Oct 05 21:17:35 crc kubenswrapper[4754]: I1005 21:17:35.595457 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" event={"ID":"be95a413-6a59-45b4-84b7-b43cae694a26","Type":"ContainerDied","Data":"a2229b13a2697fd212215ae5d196911669bc92f36ce70525da6f97d702c807b9"} Oct 05 21:17:35 crc kubenswrapper[4754]: I1005 21:17:35.595694 4754 scope.go:117] "RemoveContainer" containerID="03710aff934f922f0a7167007c25b57fd90897b7f134dd149dd0d1c75cb91da2" Oct 05 21:17:36 crc kubenswrapper[4754]: I1005 21:17:36.609027 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" event={"ID":"be95a413-6a59-45b4-84b7-b43cae694a26","Type":"ContainerStarted","Data":"e8573d2f52a359de8bfb0cb5d4d1a72d653442930ef95a60a1d6d2dd509ab1ac"} Oct 05 21:17:42 crc kubenswrapper[4754]: I1005 21:17:42.486745 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Oct 05 21:17:42 crc kubenswrapper[4754]: I1005 21:17:42.501881 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Oct 05 21:17:46 crc kubenswrapper[4754]: I1005 21:17:46.830708 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 05 21:17:47 crc kubenswrapper[4754]: I1005 21:17:47.796465 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-bwlth" event={"ID":"be10d33f-8bb1-4157-9a5f-1f5bbee194f0","Type":"ContainerStarted","Data":"3581a3a923768663b16c41e12625d2d0d6fd0a576d92e54d35f7942c2f4553a6"} Oct 05 21:17:47 crc kubenswrapper[4754]: I1005 21:17:47.826615 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-bwlth" podStartSLOduration=3.01881296 podStartE2EDuration="15.826596928s" podCreationTimestamp="2025-10-05 21:17:32 +0000 UTC" firstStartedPulling="2025-10-05 21:17:34.020420593 +0000 UTC m=+1377.924539303" lastFinishedPulling="2025-10-05 21:17:46.828204561 +0000 UTC m=+1390.732323271" observedRunningTime="2025-10-05 21:17:47.818197288 +0000 UTC m=+1391.722315998" watchObservedRunningTime="2025-10-05 21:17:47.826596928 +0000 UTC m=+1391.730715638" Oct 05 21:17:52 crc kubenswrapper[4754]: I1005 21:17:52.286063 4754 scope.go:117] "RemoveContainer" containerID="4eb4ebe30f250163db5a231ad434d1d6d333d163269168a970eb63490f9dac7e" Oct 05 21:17:52 crc kubenswrapper[4754]: I1005 21:17:52.331136 4754 scope.go:117] "RemoveContainer" containerID="e0274a90d2ea4d614f3f772870b226b425ff9e29d203263f43d6f358f0642c31" Oct 05 21:17:52 crc kubenswrapper[4754]: I1005 21:17:52.399296 4754 scope.go:117] "RemoveContainer" containerID="70f654feef08dc667d5d46a2f784bbef7d1df0ec788d6b6e2bcab20afb46ee6b" Oct 05 21:17:59 crc kubenswrapper[4754]: I1005 21:17:59.931254 4754 generic.go:334] "Generic (PLEG): container finished" podID="be10d33f-8bb1-4157-9a5f-1f5bbee194f0" containerID="3581a3a923768663b16c41e12625d2d0d6fd0a576d92e54d35f7942c2f4553a6" exitCode=0 Oct 05 21:17:59 crc kubenswrapper[4754]: I1005 21:17:59.931319 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-bwlth" event={"ID":"be10d33f-8bb1-4157-9a5f-1f5bbee194f0","Type":"ContainerDied","Data":"3581a3a923768663b16c41e12625d2d0d6fd0a576d92e54d35f7942c2f4553a6"} Oct 05 21:18:01 crc kubenswrapper[4754]: I1005 21:18:01.659177 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-bwlth" Oct 05 21:18:01 crc kubenswrapper[4754]: I1005 21:18:01.748103 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be10d33f-8bb1-4157-9a5f-1f5bbee194f0-repo-setup-combined-ca-bundle\") pod \"be10d33f-8bb1-4157-9a5f-1f5bbee194f0\" (UID: \"be10d33f-8bb1-4157-9a5f-1f5bbee194f0\") " Oct 05 21:18:01 crc kubenswrapper[4754]: I1005 21:18:01.748407 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/be10d33f-8bb1-4157-9a5f-1f5bbee194f0-inventory\") pod \"be10d33f-8bb1-4157-9a5f-1f5bbee194f0\" (UID: \"be10d33f-8bb1-4157-9a5f-1f5bbee194f0\") " Oct 05 21:18:01 crc kubenswrapper[4754]: I1005 21:18:01.748446 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/be10d33f-8bb1-4157-9a5f-1f5bbee194f0-ssh-key\") pod \"be10d33f-8bb1-4157-9a5f-1f5bbee194f0\" (UID: \"be10d33f-8bb1-4157-9a5f-1f5bbee194f0\") " Oct 05 21:18:01 crc kubenswrapper[4754]: I1005 21:18:01.748587 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6hqk\" (UniqueName: \"kubernetes.io/projected/be10d33f-8bb1-4157-9a5f-1f5bbee194f0-kube-api-access-d6hqk\") pod \"be10d33f-8bb1-4157-9a5f-1f5bbee194f0\" (UID: \"be10d33f-8bb1-4157-9a5f-1f5bbee194f0\") " Oct 05 21:18:01 crc kubenswrapper[4754]: I1005 21:18:01.753941 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be10d33f-8bb1-4157-9a5f-1f5bbee194f0-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "be10d33f-8bb1-4157-9a5f-1f5bbee194f0" (UID: "be10d33f-8bb1-4157-9a5f-1f5bbee194f0"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:18:01 crc kubenswrapper[4754]: I1005 21:18:01.769848 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/be10d33f-8bb1-4157-9a5f-1f5bbee194f0-kube-api-access-d6hqk" (OuterVolumeSpecName: "kube-api-access-d6hqk") pod "be10d33f-8bb1-4157-9a5f-1f5bbee194f0" (UID: "be10d33f-8bb1-4157-9a5f-1f5bbee194f0"). InnerVolumeSpecName "kube-api-access-d6hqk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:18:01 crc kubenswrapper[4754]: I1005 21:18:01.779406 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be10d33f-8bb1-4157-9a5f-1f5bbee194f0-inventory" (OuterVolumeSpecName: "inventory") pod "be10d33f-8bb1-4157-9a5f-1f5bbee194f0" (UID: "be10d33f-8bb1-4157-9a5f-1f5bbee194f0"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:18:01 crc kubenswrapper[4754]: I1005 21:18:01.797337 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be10d33f-8bb1-4157-9a5f-1f5bbee194f0-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "be10d33f-8bb1-4157-9a5f-1f5bbee194f0" (UID: "be10d33f-8bb1-4157-9a5f-1f5bbee194f0"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:18:01 crc kubenswrapper[4754]: I1005 21:18:01.851032 4754 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be10d33f-8bb1-4157-9a5f-1f5bbee194f0-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 21:18:01 crc kubenswrapper[4754]: I1005 21:18:01.851063 4754 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/be10d33f-8bb1-4157-9a5f-1f5bbee194f0-inventory\") on node \"crc\" DevicePath \"\"" Oct 05 21:18:01 crc kubenswrapper[4754]: I1005 21:18:01.851073 4754 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/be10d33f-8bb1-4157-9a5f-1f5bbee194f0-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 05 21:18:01 crc kubenswrapper[4754]: I1005 21:18:01.851082 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6hqk\" (UniqueName: \"kubernetes.io/projected/be10d33f-8bb1-4157-9a5f-1f5bbee194f0-kube-api-access-d6hqk\") on node \"crc\" DevicePath \"\"" Oct 05 21:18:01 crc kubenswrapper[4754]: I1005 21:18:01.959010 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-bwlth" event={"ID":"be10d33f-8bb1-4157-9a5f-1f5bbee194f0","Type":"ContainerDied","Data":"2034622146a3be6894a1e534b23f1cbd1b11adb1d787ad927faa6a1d1a0c7011"} Oct 05 21:18:01 crc kubenswrapper[4754]: I1005 21:18:01.959059 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2034622146a3be6894a1e534b23f1cbd1b11adb1d787ad927faa6a1d1a0c7011" Oct 05 21:18:01 crc kubenswrapper[4754]: I1005 21:18:01.959058 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-bwlth" Oct 05 21:18:02 crc kubenswrapper[4754]: I1005 21:18:02.057314 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-xbw7s"] Oct 05 21:18:02 crc kubenswrapper[4754]: E1005 21:18:02.057828 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be10d33f-8bb1-4157-9a5f-1f5bbee194f0" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Oct 05 21:18:02 crc kubenswrapper[4754]: I1005 21:18:02.057843 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="be10d33f-8bb1-4157-9a5f-1f5bbee194f0" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Oct 05 21:18:02 crc kubenswrapper[4754]: I1005 21:18:02.058070 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="be10d33f-8bb1-4157-9a5f-1f5bbee194f0" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Oct 05 21:18:02 crc kubenswrapper[4754]: I1005 21:18:02.058905 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-xbw7s" Oct 05 21:18:02 crc kubenswrapper[4754]: I1005 21:18:02.062057 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 05 21:18:02 crc kubenswrapper[4754]: I1005 21:18:02.062818 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-v54jd" Oct 05 21:18:02 crc kubenswrapper[4754]: I1005 21:18:02.063114 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 05 21:18:02 crc kubenswrapper[4754]: I1005 21:18:02.069234 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 05 21:18:02 crc kubenswrapper[4754]: I1005 21:18:02.076371 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-xbw7s"] Oct 05 21:18:02 crc kubenswrapper[4754]: I1005 21:18:02.156609 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fb37244c-6380-4683-a2ed-a96dd32525c6-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-xbw7s\" (UID: \"fb37244c-6380-4683-a2ed-a96dd32525c6\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-xbw7s" Oct 05 21:18:02 crc kubenswrapper[4754]: I1005 21:18:02.156770 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ldk4j\" (UniqueName: \"kubernetes.io/projected/fb37244c-6380-4683-a2ed-a96dd32525c6-kube-api-access-ldk4j\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-xbw7s\" (UID: \"fb37244c-6380-4683-a2ed-a96dd32525c6\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-xbw7s" Oct 05 21:18:02 crc kubenswrapper[4754]: I1005 21:18:02.156812 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fb37244c-6380-4683-a2ed-a96dd32525c6-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-xbw7s\" (UID: \"fb37244c-6380-4683-a2ed-a96dd32525c6\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-xbw7s" Oct 05 21:18:02 crc kubenswrapper[4754]: I1005 21:18:02.258434 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fb37244c-6380-4683-a2ed-a96dd32525c6-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-xbw7s\" (UID: \"fb37244c-6380-4683-a2ed-a96dd32525c6\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-xbw7s" Oct 05 21:18:02 crc kubenswrapper[4754]: I1005 21:18:02.258562 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ldk4j\" (UniqueName: \"kubernetes.io/projected/fb37244c-6380-4683-a2ed-a96dd32525c6-kube-api-access-ldk4j\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-xbw7s\" (UID: \"fb37244c-6380-4683-a2ed-a96dd32525c6\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-xbw7s" Oct 05 21:18:02 crc kubenswrapper[4754]: I1005 21:18:02.258605 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fb37244c-6380-4683-a2ed-a96dd32525c6-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-xbw7s\" (UID: \"fb37244c-6380-4683-a2ed-a96dd32525c6\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-xbw7s" Oct 05 21:18:02 crc kubenswrapper[4754]: I1005 21:18:02.262098 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fb37244c-6380-4683-a2ed-a96dd32525c6-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-xbw7s\" (UID: \"fb37244c-6380-4683-a2ed-a96dd32525c6\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-xbw7s" Oct 05 21:18:02 crc kubenswrapper[4754]: I1005 21:18:02.263225 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fb37244c-6380-4683-a2ed-a96dd32525c6-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-xbw7s\" (UID: \"fb37244c-6380-4683-a2ed-a96dd32525c6\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-xbw7s" Oct 05 21:18:02 crc kubenswrapper[4754]: I1005 21:18:02.280467 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ldk4j\" (UniqueName: \"kubernetes.io/projected/fb37244c-6380-4683-a2ed-a96dd32525c6-kube-api-access-ldk4j\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-xbw7s\" (UID: \"fb37244c-6380-4683-a2ed-a96dd32525c6\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-xbw7s" Oct 05 21:18:02 crc kubenswrapper[4754]: I1005 21:18:02.396163 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-xbw7s" Oct 05 21:18:03 crc kubenswrapper[4754]: I1005 21:18:03.012429 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-xbw7s"] Oct 05 21:18:03 crc kubenswrapper[4754]: I1005 21:18:03.981855 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-xbw7s" event={"ID":"fb37244c-6380-4683-a2ed-a96dd32525c6","Type":"ContainerStarted","Data":"f14d059a0ef7cc39a3f9dba88c5d299aa0c7364df38760c9dc47c5f0d65cbcdc"} Oct 05 21:18:03 crc kubenswrapper[4754]: I1005 21:18:03.982303 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-xbw7s" event={"ID":"fb37244c-6380-4683-a2ed-a96dd32525c6","Type":"ContainerStarted","Data":"a55bd54acebd58ded78f52c7a23bb5a365db1b2cf070b7307624e06914ab5014"} Oct 05 21:18:04 crc kubenswrapper[4754]: I1005 21:18:04.005771 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-xbw7s" podStartSLOduration=1.515290624 podStartE2EDuration="2.005709497s" podCreationTimestamp="2025-10-05 21:18:02 +0000 UTC" firstStartedPulling="2025-10-05 21:18:03.045403008 +0000 UTC m=+1406.949521718" lastFinishedPulling="2025-10-05 21:18:03.535821861 +0000 UTC m=+1407.439940591" observedRunningTime="2025-10-05 21:18:03.998201751 +0000 UTC m=+1407.902320471" watchObservedRunningTime="2025-10-05 21:18:04.005709497 +0000 UTC m=+1407.909828217" Oct 05 21:18:06 crc kubenswrapper[4754]: E1005 21:18:06.647409 4754 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfb37244c_6380_4683_a2ed_a96dd32525c6.slice/crio-f14d059a0ef7cc39a3f9dba88c5d299aa0c7364df38760c9dc47c5f0d65cbcdc.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfb37244c_6380_4683_a2ed_a96dd32525c6.slice/crio-conmon-f14d059a0ef7cc39a3f9dba88c5d299aa0c7364df38760c9dc47c5f0d65cbcdc.scope\": RecentStats: unable to find data in memory cache]" Oct 05 21:18:07 crc kubenswrapper[4754]: I1005 21:18:07.023675 4754 generic.go:334] "Generic (PLEG): container finished" podID="fb37244c-6380-4683-a2ed-a96dd32525c6" containerID="f14d059a0ef7cc39a3f9dba88c5d299aa0c7364df38760c9dc47c5f0d65cbcdc" exitCode=0 Oct 05 21:18:07 crc kubenswrapper[4754]: I1005 21:18:07.023720 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-xbw7s" event={"ID":"fb37244c-6380-4683-a2ed-a96dd32525c6","Type":"ContainerDied","Data":"f14d059a0ef7cc39a3f9dba88c5d299aa0c7364df38760c9dc47c5f0d65cbcdc"} Oct 05 21:18:08 crc kubenswrapper[4754]: I1005 21:18:08.505677 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-xbw7s" Oct 05 21:18:08 crc kubenswrapper[4754]: I1005 21:18:08.604159 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ldk4j\" (UniqueName: \"kubernetes.io/projected/fb37244c-6380-4683-a2ed-a96dd32525c6-kube-api-access-ldk4j\") pod \"fb37244c-6380-4683-a2ed-a96dd32525c6\" (UID: \"fb37244c-6380-4683-a2ed-a96dd32525c6\") " Oct 05 21:18:08 crc kubenswrapper[4754]: I1005 21:18:08.604210 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fb37244c-6380-4683-a2ed-a96dd32525c6-ssh-key\") pod \"fb37244c-6380-4683-a2ed-a96dd32525c6\" (UID: \"fb37244c-6380-4683-a2ed-a96dd32525c6\") " Oct 05 21:18:08 crc kubenswrapper[4754]: I1005 21:18:08.604273 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fb37244c-6380-4683-a2ed-a96dd32525c6-inventory\") pod \"fb37244c-6380-4683-a2ed-a96dd32525c6\" (UID: \"fb37244c-6380-4683-a2ed-a96dd32525c6\") " Oct 05 21:18:08 crc kubenswrapper[4754]: I1005 21:18:08.618055 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fb37244c-6380-4683-a2ed-a96dd32525c6-kube-api-access-ldk4j" (OuterVolumeSpecName: "kube-api-access-ldk4j") pod "fb37244c-6380-4683-a2ed-a96dd32525c6" (UID: "fb37244c-6380-4683-a2ed-a96dd32525c6"). InnerVolumeSpecName "kube-api-access-ldk4j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:18:08 crc kubenswrapper[4754]: I1005 21:18:08.635248 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb37244c-6380-4683-a2ed-a96dd32525c6-inventory" (OuterVolumeSpecName: "inventory") pod "fb37244c-6380-4683-a2ed-a96dd32525c6" (UID: "fb37244c-6380-4683-a2ed-a96dd32525c6"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:18:08 crc kubenswrapper[4754]: I1005 21:18:08.662811 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb37244c-6380-4683-a2ed-a96dd32525c6-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "fb37244c-6380-4683-a2ed-a96dd32525c6" (UID: "fb37244c-6380-4683-a2ed-a96dd32525c6"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:18:08 crc kubenswrapper[4754]: I1005 21:18:08.706546 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ldk4j\" (UniqueName: \"kubernetes.io/projected/fb37244c-6380-4683-a2ed-a96dd32525c6-kube-api-access-ldk4j\") on node \"crc\" DevicePath \"\"" Oct 05 21:18:08 crc kubenswrapper[4754]: I1005 21:18:08.706586 4754 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fb37244c-6380-4683-a2ed-a96dd32525c6-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 05 21:18:08 crc kubenswrapper[4754]: I1005 21:18:08.706602 4754 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fb37244c-6380-4683-a2ed-a96dd32525c6-inventory\") on node \"crc\" DevicePath \"\"" Oct 05 21:18:09 crc kubenswrapper[4754]: I1005 21:18:09.053934 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-xbw7s" event={"ID":"fb37244c-6380-4683-a2ed-a96dd32525c6","Type":"ContainerDied","Data":"a55bd54acebd58ded78f52c7a23bb5a365db1b2cf070b7307624e06914ab5014"} Oct 05 21:18:09 crc kubenswrapper[4754]: I1005 21:18:09.054310 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a55bd54acebd58ded78f52c7a23bb5a365db1b2cf070b7307624e06914ab5014" Oct 05 21:18:09 crc kubenswrapper[4754]: I1005 21:18:09.054170 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-xbw7s" Oct 05 21:18:09 crc kubenswrapper[4754]: I1005 21:18:09.137295 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-6cdph"] Oct 05 21:18:09 crc kubenswrapper[4754]: E1005 21:18:09.138082 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb37244c-6380-4683-a2ed-a96dd32525c6" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Oct 05 21:18:09 crc kubenswrapper[4754]: I1005 21:18:09.138167 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb37244c-6380-4683-a2ed-a96dd32525c6" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Oct 05 21:18:09 crc kubenswrapper[4754]: I1005 21:18:09.138522 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb37244c-6380-4683-a2ed-a96dd32525c6" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Oct 05 21:18:09 crc kubenswrapper[4754]: I1005 21:18:09.139280 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-6cdph" Oct 05 21:18:09 crc kubenswrapper[4754]: I1005 21:18:09.144105 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 05 21:18:09 crc kubenswrapper[4754]: I1005 21:18:09.144407 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 05 21:18:09 crc kubenswrapper[4754]: I1005 21:18:09.147697 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-v54jd" Oct 05 21:18:09 crc kubenswrapper[4754]: I1005 21:18:09.149416 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 05 21:18:09 crc kubenswrapper[4754]: I1005 21:18:09.160681 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-6cdph"] Oct 05 21:18:09 crc kubenswrapper[4754]: I1005 21:18:09.318876 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dn8kz\" (UniqueName: \"kubernetes.io/projected/31cbd0a7-7e3d-4bd6-b84c-816431fdc633-kube-api-access-dn8kz\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-6cdph\" (UID: \"31cbd0a7-7e3d-4bd6-b84c-816431fdc633\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-6cdph" Oct 05 21:18:09 crc kubenswrapper[4754]: I1005 21:18:09.318965 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/31cbd0a7-7e3d-4bd6-b84c-816431fdc633-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-6cdph\" (UID: \"31cbd0a7-7e3d-4bd6-b84c-816431fdc633\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-6cdph" Oct 05 21:18:09 crc kubenswrapper[4754]: I1005 21:18:09.319125 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/31cbd0a7-7e3d-4bd6-b84c-816431fdc633-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-6cdph\" (UID: \"31cbd0a7-7e3d-4bd6-b84c-816431fdc633\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-6cdph" Oct 05 21:18:09 crc kubenswrapper[4754]: I1005 21:18:09.319205 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31cbd0a7-7e3d-4bd6-b84c-816431fdc633-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-6cdph\" (UID: \"31cbd0a7-7e3d-4bd6-b84c-816431fdc633\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-6cdph" Oct 05 21:18:09 crc kubenswrapper[4754]: I1005 21:18:09.421056 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dn8kz\" (UniqueName: \"kubernetes.io/projected/31cbd0a7-7e3d-4bd6-b84c-816431fdc633-kube-api-access-dn8kz\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-6cdph\" (UID: \"31cbd0a7-7e3d-4bd6-b84c-816431fdc633\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-6cdph" Oct 05 21:18:09 crc kubenswrapper[4754]: I1005 21:18:09.421144 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/31cbd0a7-7e3d-4bd6-b84c-816431fdc633-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-6cdph\" (UID: \"31cbd0a7-7e3d-4bd6-b84c-816431fdc633\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-6cdph" Oct 05 21:18:09 crc kubenswrapper[4754]: I1005 21:18:09.421382 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/31cbd0a7-7e3d-4bd6-b84c-816431fdc633-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-6cdph\" (UID: \"31cbd0a7-7e3d-4bd6-b84c-816431fdc633\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-6cdph" Oct 05 21:18:09 crc kubenswrapper[4754]: I1005 21:18:09.421519 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31cbd0a7-7e3d-4bd6-b84c-816431fdc633-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-6cdph\" (UID: \"31cbd0a7-7e3d-4bd6-b84c-816431fdc633\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-6cdph" Oct 05 21:18:09 crc kubenswrapper[4754]: I1005 21:18:09.428902 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31cbd0a7-7e3d-4bd6-b84c-816431fdc633-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-6cdph\" (UID: \"31cbd0a7-7e3d-4bd6-b84c-816431fdc633\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-6cdph" Oct 05 21:18:09 crc kubenswrapper[4754]: I1005 21:18:09.429700 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/31cbd0a7-7e3d-4bd6-b84c-816431fdc633-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-6cdph\" (UID: \"31cbd0a7-7e3d-4bd6-b84c-816431fdc633\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-6cdph" Oct 05 21:18:09 crc kubenswrapper[4754]: I1005 21:18:09.430874 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/31cbd0a7-7e3d-4bd6-b84c-816431fdc633-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-6cdph\" (UID: \"31cbd0a7-7e3d-4bd6-b84c-816431fdc633\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-6cdph" Oct 05 21:18:09 crc kubenswrapper[4754]: I1005 21:18:09.440511 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dn8kz\" (UniqueName: \"kubernetes.io/projected/31cbd0a7-7e3d-4bd6-b84c-816431fdc633-kube-api-access-dn8kz\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-6cdph\" (UID: \"31cbd0a7-7e3d-4bd6-b84c-816431fdc633\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-6cdph" Oct 05 21:18:09 crc kubenswrapper[4754]: I1005 21:18:09.460694 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-6cdph" Oct 05 21:18:10 crc kubenswrapper[4754]: I1005 21:18:10.017761 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-6cdph"] Oct 05 21:18:10 crc kubenswrapper[4754]: I1005 21:18:10.069712 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-6cdph" event={"ID":"31cbd0a7-7e3d-4bd6-b84c-816431fdc633","Type":"ContainerStarted","Data":"9a0ee2c55ece169b6695e7942283404a8970160f609be2d9a38e6862366d945d"} Oct 05 21:18:11 crc kubenswrapper[4754]: I1005 21:18:11.081113 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-6cdph" event={"ID":"31cbd0a7-7e3d-4bd6-b84c-816431fdc633","Type":"ContainerStarted","Data":"c0e9288c9e1c394fde84639e37dc3c511c39814f13c421cee5f5cb20af0e301a"} Oct 05 21:18:11 crc kubenswrapper[4754]: I1005 21:18:11.102783 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-6cdph" podStartSLOduration=1.426104788 podStartE2EDuration="2.102713638s" podCreationTimestamp="2025-10-05 21:18:09 +0000 UTC" firstStartedPulling="2025-10-05 21:18:10.028848534 +0000 UTC m=+1413.932967274" lastFinishedPulling="2025-10-05 21:18:10.705457404 +0000 UTC m=+1414.609576124" observedRunningTime="2025-10-05 21:18:11.100422238 +0000 UTC m=+1415.004540968" watchObservedRunningTime="2025-10-05 21:18:11.102713638 +0000 UTC m=+1415.006832358" Oct 05 21:18:49 crc kubenswrapper[4754]: I1005 21:18:49.421118 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-tflxk"] Oct 05 21:18:49 crc kubenswrapper[4754]: I1005 21:18:49.424172 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tflxk" Oct 05 21:18:49 crc kubenswrapper[4754]: I1005 21:18:49.433959 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-tflxk"] Oct 05 21:18:49 crc kubenswrapper[4754]: I1005 21:18:49.563512 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e43cd7ca-670e-47be-9bea-eee9f0680a5c-catalog-content\") pod \"community-operators-tflxk\" (UID: \"e43cd7ca-670e-47be-9bea-eee9f0680a5c\") " pod="openshift-marketplace/community-operators-tflxk" Oct 05 21:18:49 crc kubenswrapper[4754]: I1005 21:18:49.563626 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e43cd7ca-670e-47be-9bea-eee9f0680a5c-utilities\") pod \"community-operators-tflxk\" (UID: \"e43cd7ca-670e-47be-9bea-eee9f0680a5c\") " pod="openshift-marketplace/community-operators-tflxk" Oct 05 21:18:49 crc kubenswrapper[4754]: I1005 21:18:49.563768 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jpwm2\" (UniqueName: \"kubernetes.io/projected/e43cd7ca-670e-47be-9bea-eee9f0680a5c-kube-api-access-jpwm2\") pod \"community-operators-tflxk\" (UID: \"e43cd7ca-670e-47be-9bea-eee9f0680a5c\") " pod="openshift-marketplace/community-operators-tflxk" Oct 05 21:18:49 crc kubenswrapper[4754]: I1005 21:18:49.665627 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jpwm2\" (UniqueName: \"kubernetes.io/projected/e43cd7ca-670e-47be-9bea-eee9f0680a5c-kube-api-access-jpwm2\") pod \"community-operators-tflxk\" (UID: \"e43cd7ca-670e-47be-9bea-eee9f0680a5c\") " pod="openshift-marketplace/community-operators-tflxk" Oct 05 21:18:49 crc kubenswrapper[4754]: I1005 21:18:49.665706 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e43cd7ca-670e-47be-9bea-eee9f0680a5c-catalog-content\") pod \"community-operators-tflxk\" (UID: \"e43cd7ca-670e-47be-9bea-eee9f0680a5c\") " pod="openshift-marketplace/community-operators-tflxk" Oct 05 21:18:49 crc kubenswrapper[4754]: I1005 21:18:49.665767 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e43cd7ca-670e-47be-9bea-eee9f0680a5c-utilities\") pod \"community-operators-tflxk\" (UID: \"e43cd7ca-670e-47be-9bea-eee9f0680a5c\") " pod="openshift-marketplace/community-operators-tflxk" Oct 05 21:18:49 crc kubenswrapper[4754]: I1005 21:18:49.666320 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e43cd7ca-670e-47be-9bea-eee9f0680a5c-utilities\") pod \"community-operators-tflxk\" (UID: \"e43cd7ca-670e-47be-9bea-eee9f0680a5c\") " pod="openshift-marketplace/community-operators-tflxk" Oct 05 21:18:49 crc kubenswrapper[4754]: I1005 21:18:49.666644 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e43cd7ca-670e-47be-9bea-eee9f0680a5c-catalog-content\") pod \"community-operators-tflxk\" (UID: \"e43cd7ca-670e-47be-9bea-eee9f0680a5c\") " pod="openshift-marketplace/community-operators-tflxk" Oct 05 21:18:49 crc kubenswrapper[4754]: I1005 21:18:49.696585 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jpwm2\" (UniqueName: \"kubernetes.io/projected/e43cd7ca-670e-47be-9bea-eee9f0680a5c-kube-api-access-jpwm2\") pod \"community-operators-tflxk\" (UID: \"e43cd7ca-670e-47be-9bea-eee9f0680a5c\") " pod="openshift-marketplace/community-operators-tflxk" Oct 05 21:18:49 crc kubenswrapper[4754]: I1005 21:18:49.767400 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tflxk" Oct 05 21:18:50 crc kubenswrapper[4754]: I1005 21:18:50.312198 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-tflxk"] Oct 05 21:18:50 crc kubenswrapper[4754]: I1005 21:18:50.523691 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tflxk" event={"ID":"e43cd7ca-670e-47be-9bea-eee9f0680a5c","Type":"ContainerStarted","Data":"ea034a4ab0393fd2021b2f2259a0fa28d557799a9c6028f9081b885ef8bd42b4"} Oct 05 21:18:50 crc kubenswrapper[4754]: I1005 21:18:50.523737 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tflxk" event={"ID":"e43cd7ca-670e-47be-9bea-eee9f0680a5c","Type":"ContainerStarted","Data":"6b2034185b5a86bde57fcc6d4d7f7ec5791cd401c6a77245d2f053b92dc7fef7"} Oct 05 21:18:51 crc kubenswrapper[4754]: I1005 21:18:51.536842 4754 generic.go:334] "Generic (PLEG): container finished" podID="e43cd7ca-670e-47be-9bea-eee9f0680a5c" containerID="ea034a4ab0393fd2021b2f2259a0fa28d557799a9c6028f9081b885ef8bd42b4" exitCode=0 Oct 05 21:18:51 crc kubenswrapper[4754]: I1005 21:18:51.536948 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tflxk" event={"ID":"e43cd7ca-670e-47be-9bea-eee9f0680a5c","Type":"ContainerDied","Data":"ea034a4ab0393fd2021b2f2259a0fa28d557799a9c6028f9081b885ef8bd42b4"} Oct 05 21:18:51 crc kubenswrapper[4754]: I1005 21:18:51.537131 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tflxk" event={"ID":"e43cd7ca-670e-47be-9bea-eee9f0680a5c","Type":"ContainerStarted","Data":"f49120ff5c877b5fb3b6d1332525941ccdfd71ee47962057db14bb364f811ed6"} Oct 05 21:18:52 crc kubenswrapper[4754]: I1005 21:18:52.707113 4754 scope.go:117] "RemoveContainer" containerID="7a3be6d1860224c38bae686b11dc667c26380ccacdec9958d497ad2fde6b53c3" Oct 05 21:18:53 crc kubenswrapper[4754]: I1005 21:18:53.080116 4754 scope.go:117] "RemoveContainer" containerID="8e78a86fa9ef8e80c4b54e2989ab5675be879fe0716c3b91dc3600b2f0845d0c" Oct 05 21:18:53 crc kubenswrapper[4754]: I1005 21:18:53.105106 4754 scope.go:117] "RemoveContainer" containerID="3145d30e064762755f608d429f86b8dbe13676109d0cc8b22b9dc0af78ebbe01" Oct 05 21:18:53 crc kubenswrapper[4754]: I1005 21:18:53.132075 4754 scope.go:117] "RemoveContainer" containerID="90a1ac4ad2b42fce36308ae6b7c3ddb93e5d7a13682a2554e3b20086da7c736f" Oct 05 21:18:53 crc kubenswrapper[4754]: I1005 21:18:53.565102 4754 generic.go:334] "Generic (PLEG): container finished" podID="e43cd7ca-670e-47be-9bea-eee9f0680a5c" containerID="f49120ff5c877b5fb3b6d1332525941ccdfd71ee47962057db14bb364f811ed6" exitCode=0 Oct 05 21:18:53 crc kubenswrapper[4754]: I1005 21:18:53.565191 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tflxk" event={"ID":"e43cd7ca-670e-47be-9bea-eee9f0680a5c","Type":"ContainerDied","Data":"f49120ff5c877b5fb3b6d1332525941ccdfd71ee47962057db14bb364f811ed6"} Oct 05 21:18:54 crc kubenswrapper[4754]: I1005 21:18:54.579486 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tflxk" event={"ID":"e43cd7ca-670e-47be-9bea-eee9f0680a5c","Type":"ContainerStarted","Data":"5a78f629279752ea702966090f61cd23f31a05fe81729d10aa87ee74f54b702c"} Oct 05 21:18:57 crc kubenswrapper[4754]: I1005 21:18:57.288441 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-tflxk" podStartSLOduration=4.762810619 podStartE2EDuration="8.288410935s" podCreationTimestamp="2025-10-05 21:18:49 +0000 UTC" firstStartedPulling="2025-10-05 21:18:50.525568767 +0000 UTC m=+1454.429687477" lastFinishedPulling="2025-10-05 21:18:54.051169043 +0000 UTC m=+1457.955287793" observedRunningTime="2025-10-05 21:18:54.614745264 +0000 UTC m=+1458.518863974" watchObservedRunningTime="2025-10-05 21:18:57.288410935 +0000 UTC m=+1461.192529645" Oct 05 21:18:57 crc kubenswrapper[4754]: I1005 21:18:57.298828 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-v9tcs"] Oct 05 21:18:57 crc kubenswrapper[4754]: I1005 21:18:57.301694 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-v9tcs" Oct 05 21:18:57 crc kubenswrapper[4754]: I1005 21:18:57.318963 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-v9tcs"] Oct 05 21:18:57 crc kubenswrapper[4754]: I1005 21:18:57.443703 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8pblr\" (UniqueName: \"kubernetes.io/projected/a4e04343-a581-420e-80e6-b89287382b1a-kube-api-access-8pblr\") pod \"redhat-marketplace-v9tcs\" (UID: \"a4e04343-a581-420e-80e6-b89287382b1a\") " pod="openshift-marketplace/redhat-marketplace-v9tcs" Oct 05 21:18:57 crc kubenswrapper[4754]: I1005 21:18:57.444138 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a4e04343-a581-420e-80e6-b89287382b1a-catalog-content\") pod \"redhat-marketplace-v9tcs\" (UID: \"a4e04343-a581-420e-80e6-b89287382b1a\") " pod="openshift-marketplace/redhat-marketplace-v9tcs" Oct 05 21:18:57 crc kubenswrapper[4754]: I1005 21:18:57.444218 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a4e04343-a581-420e-80e6-b89287382b1a-utilities\") pod \"redhat-marketplace-v9tcs\" (UID: \"a4e04343-a581-420e-80e6-b89287382b1a\") " pod="openshift-marketplace/redhat-marketplace-v9tcs" Oct 05 21:18:57 crc kubenswrapper[4754]: I1005 21:18:57.546362 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a4e04343-a581-420e-80e6-b89287382b1a-catalog-content\") pod \"redhat-marketplace-v9tcs\" (UID: \"a4e04343-a581-420e-80e6-b89287382b1a\") " pod="openshift-marketplace/redhat-marketplace-v9tcs" Oct 05 21:18:57 crc kubenswrapper[4754]: I1005 21:18:57.546420 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a4e04343-a581-420e-80e6-b89287382b1a-utilities\") pod \"redhat-marketplace-v9tcs\" (UID: \"a4e04343-a581-420e-80e6-b89287382b1a\") " pod="openshift-marketplace/redhat-marketplace-v9tcs" Oct 05 21:18:57 crc kubenswrapper[4754]: I1005 21:18:57.546477 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8pblr\" (UniqueName: \"kubernetes.io/projected/a4e04343-a581-420e-80e6-b89287382b1a-kube-api-access-8pblr\") pod \"redhat-marketplace-v9tcs\" (UID: \"a4e04343-a581-420e-80e6-b89287382b1a\") " pod="openshift-marketplace/redhat-marketplace-v9tcs" Oct 05 21:18:57 crc kubenswrapper[4754]: I1005 21:18:57.546973 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a4e04343-a581-420e-80e6-b89287382b1a-catalog-content\") pod \"redhat-marketplace-v9tcs\" (UID: \"a4e04343-a581-420e-80e6-b89287382b1a\") " pod="openshift-marketplace/redhat-marketplace-v9tcs" Oct 05 21:18:57 crc kubenswrapper[4754]: I1005 21:18:57.547172 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a4e04343-a581-420e-80e6-b89287382b1a-utilities\") pod \"redhat-marketplace-v9tcs\" (UID: \"a4e04343-a581-420e-80e6-b89287382b1a\") " pod="openshift-marketplace/redhat-marketplace-v9tcs" Oct 05 21:18:57 crc kubenswrapper[4754]: I1005 21:18:57.570777 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8pblr\" (UniqueName: \"kubernetes.io/projected/a4e04343-a581-420e-80e6-b89287382b1a-kube-api-access-8pblr\") pod \"redhat-marketplace-v9tcs\" (UID: \"a4e04343-a581-420e-80e6-b89287382b1a\") " pod="openshift-marketplace/redhat-marketplace-v9tcs" Oct 05 21:18:57 crc kubenswrapper[4754]: I1005 21:18:57.623795 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-v9tcs" Oct 05 21:18:58 crc kubenswrapper[4754]: I1005 21:18:58.124747 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-v9tcs"] Oct 05 21:18:58 crc kubenswrapper[4754]: I1005 21:18:58.622077 4754 generic.go:334] "Generic (PLEG): container finished" podID="a4e04343-a581-420e-80e6-b89287382b1a" containerID="6868b6dafa8bb2356835b800f930f8421ab93b373af0ed103e14d6a47e12269c" exitCode=0 Oct 05 21:18:58 crc kubenswrapper[4754]: I1005 21:18:58.622164 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v9tcs" event={"ID":"a4e04343-a581-420e-80e6-b89287382b1a","Type":"ContainerDied","Data":"6868b6dafa8bb2356835b800f930f8421ab93b373af0ed103e14d6a47e12269c"} Oct 05 21:18:58 crc kubenswrapper[4754]: I1005 21:18:58.622609 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v9tcs" event={"ID":"a4e04343-a581-420e-80e6-b89287382b1a","Type":"ContainerStarted","Data":"2422e25ee892c27cb224ce72d3aa987387e7e89bb7f2fb0f50399539fd30b924"} Oct 05 21:18:59 crc kubenswrapper[4754]: I1005 21:18:59.633765 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v9tcs" event={"ID":"a4e04343-a581-420e-80e6-b89287382b1a","Type":"ContainerStarted","Data":"5487f14e98039f896024a9776dc9ab7717976acfec8ad06e0ec2a6638f9f5d18"} Oct 05 21:18:59 crc kubenswrapper[4754]: I1005 21:18:59.768477 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-tflxk" Oct 05 21:18:59 crc kubenswrapper[4754]: I1005 21:18:59.768555 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-tflxk" Oct 05 21:19:00 crc kubenswrapper[4754]: I1005 21:19:00.651575 4754 generic.go:334] "Generic (PLEG): container finished" podID="a4e04343-a581-420e-80e6-b89287382b1a" containerID="5487f14e98039f896024a9776dc9ab7717976acfec8ad06e0ec2a6638f9f5d18" exitCode=0 Oct 05 21:19:00 crc kubenswrapper[4754]: I1005 21:19:00.651649 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v9tcs" event={"ID":"a4e04343-a581-420e-80e6-b89287382b1a","Type":"ContainerDied","Data":"5487f14e98039f896024a9776dc9ab7717976acfec8ad06e0ec2a6638f9f5d18"} Oct 05 21:19:00 crc kubenswrapper[4754]: I1005 21:19:00.818557 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-tflxk" podUID="e43cd7ca-670e-47be-9bea-eee9f0680a5c" containerName="registry-server" probeResult="failure" output=< Oct 05 21:19:00 crc kubenswrapper[4754]: timeout: failed to connect service ":50051" within 1s Oct 05 21:19:00 crc kubenswrapper[4754]: > Oct 05 21:19:01 crc kubenswrapper[4754]: I1005 21:19:01.679207 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v9tcs" event={"ID":"a4e04343-a581-420e-80e6-b89287382b1a","Type":"ContainerStarted","Data":"1d6ad73c62eb0aee4172d1ee822ddd66168bb6ee0ee9c388a9452344ac4aa054"} Oct 05 21:19:01 crc kubenswrapper[4754]: I1005 21:19:01.714461 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-v9tcs" podStartSLOduration=2.250450765 podStartE2EDuration="4.714413514s" podCreationTimestamp="2025-10-05 21:18:57 +0000 UTC" firstStartedPulling="2025-10-05 21:18:58.625106405 +0000 UTC m=+1462.529225125" lastFinishedPulling="2025-10-05 21:19:01.089069164 +0000 UTC m=+1464.993187874" observedRunningTime="2025-10-05 21:19:01.702651223 +0000 UTC m=+1465.606769963" watchObservedRunningTime="2025-10-05 21:19:01.714413514 +0000 UTC m=+1465.618532224" Oct 05 21:19:07 crc kubenswrapper[4754]: I1005 21:19:07.624172 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-v9tcs" Oct 05 21:19:07 crc kubenswrapper[4754]: I1005 21:19:07.626436 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-v9tcs" Oct 05 21:19:07 crc kubenswrapper[4754]: I1005 21:19:07.693353 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-v9tcs" Oct 05 21:19:07 crc kubenswrapper[4754]: I1005 21:19:07.833065 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-v9tcs" Oct 05 21:19:07 crc kubenswrapper[4754]: I1005 21:19:07.959968 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-v9tcs"] Oct 05 21:19:09 crc kubenswrapper[4754]: I1005 21:19:09.766746 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-v9tcs" podUID="a4e04343-a581-420e-80e6-b89287382b1a" containerName="registry-server" containerID="cri-o://1d6ad73c62eb0aee4172d1ee822ddd66168bb6ee0ee9c388a9452344ac4aa054" gracePeriod=2 Oct 05 21:19:09 crc kubenswrapper[4754]: I1005 21:19:09.851776 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-tflxk" Oct 05 21:19:09 crc kubenswrapper[4754]: I1005 21:19:09.920684 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-tflxk" Oct 05 21:19:10 crc kubenswrapper[4754]: I1005 21:19:10.253768 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-v9tcs" Oct 05 21:19:10 crc kubenswrapper[4754]: I1005 21:19:10.321170 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8pblr\" (UniqueName: \"kubernetes.io/projected/a4e04343-a581-420e-80e6-b89287382b1a-kube-api-access-8pblr\") pod \"a4e04343-a581-420e-80e6-b89287382b1a\" (UID: \"a4e04343-a581-420e-80e6-b89287382b1a\") " Oct 05 21:19:10 crc kubenswrapper[4754]: I1005 21:19:10.321388 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a4e04343-a581-420e-80e6-b89287382b1a-catalog-content\") pod \"a4e04343-a581-420e-80e6-b89287382b1a\" (UID: \"a4e04343-a581-420e-80e6-b89287382b1a\") " Oct 05 21:19:10 crc kubenswrapper[4754]: I1005 21:19:10.321441 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a4e04343-a581-420e-80e6-b89287382b1a-utilities\") pod \"a4e04343-a581-420e-80e6-b89287382b1a\" (UID: \"a4e04343-a581-420e-80e6-b89287382b1a\") " Oct 05 21:19:10 crc kubenswrapper[4754]: I1005 21:19:10.322030 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a4e04343-a581-420e-80e6-b89287382b1a-utilities" (OuterVolumeSpecName: "utilities") pod "a4e04343-a581-420e-80e6-b89287382b1a" (UID: "a4e04343-a581-420e-80e6-b89287382b1a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 21:19:10 crc kubenswrapper[4754]: I1005 21:19:10.327784 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a4e04343-a581-420e-80e6-b89287382b1a-kube-api-access-8pblr" (OuterVolumeSpecName: "kube-api-access-8pblr") pod "a4e04343-a581-420e-80e6-b89287382b1a" (UID: "a4e04343-a581-420e-80e6-b89287382b1a"). InnerVolumeSpecName "kube-api-access-8pblr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:19:10 crc kubenswrapper[4754]: I1005 21:19:10.346870 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a4e04343-a581-420e-80e6-b89287382b1a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a4e04343-a581-420e-80e6-b89287382b1a" (UID: "a4e04343-a581-420e-80e6-b89287382b1a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 21:19:10 crc kubenswrapper[4754]: I1005 21:19:10.349560 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-tflxk"] Oct 05 21:19:10 crc kubenswrapper[4754]: I1005 21:19:10.424757 4754 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a4e04343-a581-420e-80e6-b89287382b1a-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 21:19:10 crc kubenswrapper[4754]: I1005 21:19:10.424818 4754 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a4e04343-a581-420e-80e6-b89287382b1a-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 21:19:10 crc kubenswrapper[4754]: I1005 21:19:10.424831 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8pblr\" (UniqueName: \"kubernetes.io/projected/a4e04343-a581-420e-80e6-b89287382b1a-kube-api-access-8pblr\") on node \"crc\" DevicePath \"\"" Oct 05 21:19:10 crc kubenswrapper[4754]: I1005 21:19:10.780460 4754 generic.go:334] "Generic (PLEG): container finished" podID="a4e04343-a581-420e-80e6-b89287382b1a" containerID="1d6ad73c62eb0aee4172d1ee822ddd66168bb6ee0ee9c388a9452344ac4aa054" exitCode=0 Oct 05 21:19:10 crc kubenswrapper[4754]: I1005 21:19:10.781756 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v9tcs" event={"ID":"a4e04343-a581-420e-80e6-b89287382b1a","Type":"ContainerDied","Data":"1d6ad73c62eb0aee4172d1ee822ddd66168bb6ee0ee9c388a9452344ac4aa054"} Oct 05 21:19:10 crc kubenswrapper[4754]: I1005 21:19:10.781826 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v9tcs" event={"ID":"a4e04343-a581-420e-80e6-b89287382b1a","Type":"ContainerDied","Data":"2422e25ee892c27cb224ce72d3aa987387e7e89bb7f2fb0f50399539fd30b924"} Oct 05 21:19:10 crc kubenswrapper[4754]: I1005 21:19:10.781857 4754 scope.go:117] "RemoveContainer" containerID="1d6ad73c62eb0aee4172d1ee822ddd66168bb6ee0ee9c388a9452344ac4aa054" Oct 05 21:19:10 crc kubenswrapper[4754]: I1005 21:19:10.781881 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-v9tcs" Oct 05 21:19:10 crc kubenswrapper[4754]: I1005 21:19:10.807830 4754 scope.go:117] "RemoveContainer" containerID="5487f14e98039f896024a9776dc9ab7717976acfec8ad06e0ec2a6638f9f5d18" Oct 05 21:19:10 crc kubenswrapper[4754]: I1005 21:19:10.880263 4754 scope.go:117] "RemoveContainer" containerID="6868b6dafa8bb2356835b800f930f8421ab93b373af0ed103e14d6a47e12269c" Oct 05 21:19:10 crc kubenswrapper[4754]: I1005 21:19:10.882871 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-v9tcs"] Oct 05 21:19:10 crc kubenswrapper[4754]: I1005 21:19:10.889834 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-v9tcs"] Oct 05 21:19:10 crc kubenswrapper[4754]: I1005 21:19:10.917588 4754 scope.go:117] "RemoveContainer" containerID="1d6ad73c62eb0aee4172d1ee822ddd66168bb6ee0ee9c388a9452344ac4aa054" Oct 05 21:19:10 crc kubenswrapper[4754]: E1005 21:19:10.921789 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1d6ad73c62eb0aee4172d1ee822ddd66168bb6ee0ee9c388a9452344ac4aa054\": container with ID starting with 1d6ad73c62eb0aee4172d1ee822ddd66168bb6ee0ee9c388a9452344ac4aa054 not found: ID does not exist" containerID="1d6ad73c62eb0aee4172d1ee822ddd66168bb6ee0ee9c388a9452344ac4aa054" Oct 05 21:19:10 crc kubenswrapper[4754]: I1005 21:19:10.921839 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d6ad73c62eb0aee4172d1ee822ddd66168bb6ee0ee9c388a9452344ac4aa054"} err="failed to get container status \"1d6ad73c62eb0aee4172d1ee822ddd66168bb6ee0ee9c388a9452344ac4aa054\": rpc error: code = NotFound desc = could not find container \"1d6ad73c62eb0aee4172d1ee822ddd66168bb6ee0ee9c388a9452344ac4aa054\": container with ID starting with 1d6ad73c62eb0aee4172d1ee822ddd66168bb6ee0ee9c388a9452344ac4aa054 not found: ID does not exist" Oct 05 21:19:10 crc kubenswrapper[4754]: I1005 21:19:10.921884 4754 scope.go:117] "RemoveContainer" containerID="5487f14e98039f896024a9776dc9ab7717976acfec8ad06e0ec2a6638f9f5d18" Oct 05 21:19:10 crc kubenswrapper[4754]: E1005 21:19:10.922429 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5487f14e98039f896024a9776dc9ab7717976acfec8ad06e0ec2a6638f9f5d18\": container with ID starting with 5487f14e98039f896024a9776dc9ab7717976acfec8ad06e0ec2a6638f9f5d18 not found: ID does not exist" containerID="5487f14e98039f896024a9776dc9ab7717976acfec8ad06e0ec2a6638f9f5d18" Oct 05 21:19:10 crc kubenswrapper[4754]: I1005 21:19:10.922518 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5487f14e98039f896024a9776dc9ab7717976acfec8ad06e0ec2a6638f9f5d18"} err="failed to get container status \"5487f14e98039f896024a9776dc9ab7717976acfec8ad06e0ec2a6638f9f5d18\": rpc error: code = NotFound desc = could not find container \"5487f14e98039f896024a9776dc9ab7717976acfec8ad06e0ec2a6638f9f5d18\": container with ID starting with 5487f14e98039f896024a9776dc9ab7717976acfec8ad06e0ec2a6638f9f5d18 not found: ID does not exist" Oct 05 21:19:10 crc kubenswrapper[4754]: I1005 21:19:10.922542 4754 scope.go:117] "RemoveContainer" containerID="6868b6dafa8bb2356835b800f930f8421ab93b373af0ed103e14d6a47e12269c" Oct 05 21:19:10 crc kubenswrapper[4754]: E1005 21:19:10.922825 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6868b6dafa8bb2356835b800f930f8421ab93b373af0ed103e14d6a47e12269c\": container with ID starting with 6868b6dafa8bb2356835b800f930f8421ab93b373af0ed103e14d6a47e12269c not found: ID does not exist" containerID="6868b6dafa8bb2356835b800f930f8421ab93b373af0ed103e14d6a47e12269c" Oct 05 21:19:10 crc kubenswrapper[4754]: I1005 21:19:10.922858 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6868b6dafa8bb2356835b800f930f8421ab93b373af0ed103e14d6a47e12269c"} err="failed to get container status \"6868b6dafa8bb2356835b800f930f8421ab93b373af0ed103e14d6a47e12269c\": rpc error: code = NotFound desc = could not find container \"6868b6dafa8bb2356835b800f930f8421ab93b373af0ed103e14d6a47e12269c\": container with ID starting with 6868b6dafa8bb2356835b800f930f8421ab93b373af0ed103e14d6a47e12269c not found: ID does not exist" Oct 05 21:19:11 crc kubenswrapper[4754]: I1005 21:19:11.793781 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-tflxk" podUID="e43cd7ca-670e-47be-9bea-eee9f0680a5c" containerName="registry-server" containerID="cri-o://5a78f629279752ea702966090f61cd23f31a05fe81729d10aa87ee74f54b702c" gracePeriod=2 Oct 05 21:19:12 crc kubenswrapper[4754]: I1005 21:19:12.254108 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tflxk" Oct 05 21:19:12 crc kubenswrapper[4754]: I1005 21:19:12.379192 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e43cd7ca-670e-47be-9bea-eee9f0680a5c-catalog-content\") pod \"e43cd7ca-670e-47be-9bea-eee9f0680a5c\" (UID: \"e43cd7ca-670e-47be-9bea-eee9f0680a5c\") " Oct 05 21:19:12 crc kubenswrapper[4754]: I1005 21:19:12.379246 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e43cd7ca-670e-47be-9bea-eee9f0680a5c-utilities\") pod \"e43cd7ca-670e-47be-9bea-eee9f0680a5c\" (UID: \"e43cd7ca-670e-47be-9bea-eee9f0680a5c\") " Oct 05 21:19:12 crc kubenswrapper[4754]: I1005 21:19:12.379532 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jpwm2\" (UniqueName: \"kubernetes.io/projected/e43cd7ca-670e-47be-9bea-eee9f0680a5c-kube-api-access-jpwm2\") pod \"e43cd7ca-670e-47be-9bea-eee9f0680a5c\" (UID: \"e43cd7ca-670e-47be-9bea-eee9f0680a5c\") " Oct 05 21:19:12 crc kubenswrapper[4754]: I1005 21:19:12.381767 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e43cd7ca-670e-47be-9bea-eee9f0680a5c-utilities" (OuterVolumeSpecName: "utilities") pod "e43cd7ca-670e-47be-9bea-eee9f0680a5c" (UID: "e43cd7ca-670e-47be-9bea-eee9f0680a5c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 21:19:12 crc kubenswrapper[4754]: I1005 21:19:12.395039 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e43cd7ca-670e-47be-9bea-eee9f0680a5c-kube-api-access-jpwm2" (OuterVolumeSpecName: "kube-api-access-jpwm2") pod "e43cd7ca-670e-47be-9bea-eee9f0680a5c" (UID: "e43cd7ca-670e-47be-9bea-eee9f0680a5c"). InnerVolumeSpecName "kube-api-access-jpwm2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:19:12 crc kubenswrapper[4754]: I1005 21:19:12.460971 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e43cd7ca-670e-47be-9bea-eee9f0680a5c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e43cd7ca-670e-47be-9bea-eee9f0680a5c" (UID: "e43cd7ca-670e-47be-9bea-eee9f0680a5c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 21:19:12 crc kubenswrapper[4754]: I1005 21:19:12.482347 4754 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e43cd7ca-670e-47be-9bea-eee9f0680a5c-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 21:19:12 crc kubenswrapper[4754]: I1005 21:19:12.482397 4754 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e43cd7ca-670e-47be-9bea-eee9f0680a5c-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 21:19:12 crc kubenswrapper[4754]: I1005 21:19:12.482413 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jpwm2\" (UniqueName: \"kubernetes.io/projected/e43cd7ca-670e-47be-9bea-eee9f0680a5c-kube-api-access-jpwm2\") on node \"crc\" DevicePath \"\"" Oct 05 21:19:12 crc kubenswrapper[4754]: I1005 21:19:12.812560 4754 generic.go:334] "Generic (PLEG): container finished" podID="e43cd7ca-670e-47be-9bea-eee9f0680a5c" containerID="5a78f629279752ea702966090f61cd23f31a05fe81729d10aa87ee74f54b702c" exitCode=0 Oct 05 21:19:12 crc kubenswrapper[4754]: I1005 21:19:12.812642 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tflxk" event={"ID":"e43cd7ca-670e-47be-9bea-eee9f0680a5c","Type":"ContainerDied","Data":"5a78f629279752ea702966090f61cd23f31a05fe81729d10aa87ee74f54b702c"} Oct 05 21:19:12 crc kubenswrapper[4754]: I1005 21:19:12.812687 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tflxk" event={"ID":"e43cd7ca-670e-47be-9bea-eee9f0680a5c","Type":"ContainerDied","Data":"6b2034185b5a86bde57fcc6d4d7f7ec5791cd401c6a77245d2f053b92dc7fef7"} Oct 05 21:19:12 crc kubenswrapper[4754]: I1005 21:19:12.812722 4754 scope.go:117] "RemoveContainer" containerID="5a78f629279752ea702966090f61cd23f31a05fe81729d10aa87ee74f54b702c" Oct 05 21:19:12 crc kubenswrapper[4754]: I1005 21:19:12.812952 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tflxk" Oct 05 21:19:12 crc kubenswrapper[4754]: I1005 21:19:12.869955 4754 scope.go:117] "RemoveContainer" containerID="f49120ff5c877b5fb3b6d1332525941ccdfd71ee47962057db14bb364f811ed6" Oct 05 21:19:12 crc kubenswrapper[4754]: I1005 21:19:12.872659 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a4e04343-a581-420e-80e6-b89287382b1a" path="/var/lib/kubelet/pods/a4e04343-a581-420e-80e6-b89287382b1a/volumes" Oct 05 21:19:12 crc kubenswrapper[4754]: I1005 21:19:12.881994 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-tflxk"] Oct 05 21:19:12 crc kubenswrapper[4754]: I1005 21:19:12.894861 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-tflxk"] Oct 05 21:19:12 crc kubenswrapper[4754]: I1005 21:19:12.931439 4754 scope.go:117] "RemoveContainer" containerID="ea034a4ab0393fd2021b2f2259a0fa28d557799a9c6028f9081b885ef8bd42b4" Oct 05 21:19:12 crc kubenswrapper[4754]: I1005 21:19:12.994818 4754 scope.go:117] "RemoveContainer" containerID="5a78f629279752ea702966090f61cd23f31a05fe81729d10aa87ee74f54b702c" Oct 05 21:19:12 crc kubenswrapper[4754]: E1005 21:19:12.996451 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5a78f629279752ea702966090f61cd23f31a05fe81729d10aa87ee74f54b702c\": container with ID starting with 5a78f629279752ea702966090f61cd23f31a05fe81729d10aa87ee74f54b702c not found: ID does not exist" containerID="5a78f629279752ea702966090f61cd23f31a05fe81729d10aa87ee74f54b702c" Oct 05 21:19:12 crc kubenswrapper[4754]: I1005 21:19:12.996538 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5a78f629279752ea702966090f61cd23f31a05fe81729d10aa87ee74f54b702c"} err="failed to get container status \"5a78f629279752ea702966090f61cd23f31a05fe81729d10aa87ee74f54b702c\": rpc error: code = NotFound desc = could not find container \"5a78f629279752ea702966090f61cd23f31a05fe81729d10aa87ee74f54b702c\": container with ID starting with 5a78f629279752ea702966090f61cd23f31a05fe81729d10aa87ee74f54b702c not found: ID does not exist" Oct 05 21:19:12 crc kubenswrapper[4754]: I1005 21:19:12.996583 4754 scope.go:117] "RemoveContainer" containerID="f49120ff5c877b5fb3b6d1332525941ccdfd71ee47962057db14bb364f811ed6" Oct 05 21:19:12 crc kubenswrapper[4754]: E1005 21:19:12.996961 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f49120ff5c877b5fb3b6d1332525941ccdfd71ee47962057db14bb364f811ed6\": container with ID starting with f49120ff5c877b5fb3b6d1332525941ccdfd71ee47962057db14bb364f811ed6 not found: ID does not exist" containerID="f49120ff5c877b5fb3b6d1332525941ccdfd71ee47962057db14bb364f811ed6" Oct 05 21:19:12 crc kubenswrapper[4754]: I1005 21:19:12.996987 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f49120ff5c877b5fb3b6d1332525941ccdfd71ee47962057db14bb364f811ed6"} err="failed to get container status \"f49120ff5c877b5fb3b6d1332525941ccdfd71ee47962057db14bb364f811ed6\": rpc error: code = NotFound desc = could not find container \"f49120ff5c877b5fb3b6d1332525941ccdfd71ee47962057db14bb364f811ed6\": container with ID starting with f49120ff5c877b5fb3b6d1332525941ccdfd71ee47962057db14bb364f811ed6 not found: ID does not exist" Oct 05 21:19:12 crc kubenswrapper[4754]: I1005 21:19:12.997010 4754 scope.go:117] "RemoveContainer" containerID="ea034a4ab0393fd2021b2f2259a0fa28d557799a9c6028f9081b885ef8bd42b4" Oct 05 21:19:12 crc kubenswrapper[4754]: E1005 21:19:12.997237 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ea034a4ab0393fd2021b2f2259a0fa28d557799a9c6028f9081b885ef8bd42b4\": container with ID starting with ea034a4ab0393fd2021b2f2259a0fa28d557799a9c6028f9081b885ef8bd42b4 not found: ID does not exist" containerID="ea034a4ab0393fd2021b2f2259a0fa28d557799a9c6028f9081b885ef8bd42b4" Oct 05 21:19:12 crc kubenswrapper[4754]: I1005 21:19:12.997263 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ea034a4ab0393fd2021b2f2259a0fa28d557799a9c6028f9081b885ef8bd42b4"} err="failed to get container status \"ea034a4ab0393fd2021b2f2259a0fa28d557799a9c6028f9081b885ef8bd42b4\": rpc error: code = NotFound desc = could not find container \"ea034a4ab0393fd2021b2f2259a0fa28d557799a9c6028f9081b885ef8bd42b4\": container with ID starting with ea034a4ab0393fd2021b2f2259a0fa28d557799a9c6028f9081b885ef8bd42b4 not found: ID does not exist" Oct 05 21:19:14 crc kubenswrapper[4754]: I1005 21:19:14.857736 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e43cd7ca-670e-47be-9bea-eee9f0680a5c" path="/var/lib/kubelet/pods/e43cd7ca-670e-47be-9bea-eee9f0680a5c/volumes" Oct 05 21:19:35 crc kubenswrapper[4754]: I1005 21:19:35.245475 4754 patch_prober.go:28] interesting pod/machine-config-daemon-b2h9k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 21:19:35 crc kubenswrapper[4754]: I1005 21:19:35.246555 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 21:20:05 crc kubenswrapper[4754]: I1005 21:20:05.245996 4754 patch_prober.go:28] interesting pod/machine-config-daemon-b2h9k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 21:20:05 crc kubenswrapper[4754]: I1005 21:20:05.247020 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 21:20:16 crc kubenswrapper[4754]: I1005 21:20:16.326548 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-k4ggq"] Oct 05 21:20:16 crc kubenswrapper[4754]: E1005 21:20:16.330934 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4e04343-a581-420e-80e6-b89287382b1a" containerName="extract-utilities" Oct 05 21:20:16 crc kubenswrapper[4754]: I1005 21:20:16.331101 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4e04343-a581-420e-80e6-b89287382b1a" containerName="extract-utilities" Oct 05 21:20:16 crc kubenswrapper[4754]: E1005 21:20:16.331150 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e43cd7ca-670e-47be-9bea-eee9f0680a5c" containerName="extract-utilities" Oct 05 21:20:16 crc kubenswrapper[4754]: I1005 21:20:16.331215 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="e43cd7ca-670e-47be-9bea-eee9f0680a5c" containerName="extract-utilities" Oct 05 21:20:16 crc kubenswrapper[4754]: E1005 21:20:16.331339 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e43cd7ca-670e-47be-9bea-eee9f0680a5c" containerName="registry-server" Oct 05 21:20:16 crc kubenswrapper[4754]: I1005 21:20:16.331406 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="e43cd7ca-670e-47be-9bea-eee9f0680a5c" containerName="registry-server" Oct 05 21:20:16 crc kubenswrapper[4754]: E1005 21:20:16.331481 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4e04343-a581-420e-80e6-b89287382b1a" containerName="extract-content" Oct 05 21:20:16 crc kubenswrapper[4754]: I1005 21:20:16.331590 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4e04343-a581-420e-80e6-b89287382b1a" containerName="extract-content" Oct 05 21:20:16 crc kubenswrapper[4754]: E1005 21:20:16.331706 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e43cd7ca-670e-47be-9bea-eee9f0680a5c" containerName="extract-content" Oct 05 21:20:16 crc kubenswrapper[4754]: I1005 21:20:16.331743 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="e43cd7ca-670e-47be-9bea-eee9f0680a5c" containerName="extract-content" Oct 05 21:20:16 crc kubenswrapper[4754]: E1005 21:20:16.331840 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4e04343-a581-420e-80e6-b89287382b1a" containerName="registry-server" Oct 05 21:20:16 crc kubenswrapper[4754]: I1005 21:20:16.331913 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4e04343-a581-420e-80e6-b89287382b1a" containerName="registry-server" Oct 05 21:20:16 crc kubenswrapper[4754]: I1005 21:20:16.333281 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="e43cd7ca-670e-47be-9bea-eee9f0680a5c" containerName="registry-server" Oct 05 21:20:16 crc kubenswrapper[4754]: I1005 21:20:16.333374 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="a4e04343-a581-420e-80e6-b89287382b1a" containerName="registry-server" Oct 05 21:20:16 crc kubenswrapper[4754]: I1005 21:20:16.337639 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-k4ggq" Oct 05 21:20:16 crc kubenswrapper[4754]: I1005 21:20:16.357470 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-k4ggq"] Oct 05 21:20:16 crc kubenswrapper[4754]: I1005 21:20:16.516267 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3faaf83b-b67f-454a-845f-d10861c6db57-catalog-content\") pod \"redhat-operators-k4ggq\" (UID: \"3faaf83b-b67f-454a-845f-d10861c6db57\") " pod="openshift-marketplace/redhat-operators-k4ggq" Oct 05 21:20:16 crc kubenswrapper[4754]: I1005 21:20:16.516559 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3faaf83b-b67f-454a-845f-d10861c6db57-utilities\") pod \"redhat-operators-k4ggq\" (UID: \"3faaf83b-b67f-454a-845f-d10861c6db57\") " pod="openshift-marketplace/redhat-operators-k4ggq" Oct 05 21:20:16 crc kubenswrapper[4754]: I1005 21:20:16.516659 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f48g8\" (UniqueName: \"kubernetes.io/projected/3faaf83b-b67f-454a-845f-d10861c6db57-kube-api-access-f48g8\") pod \"redhat-operators-k4ggq\" (UID: \"3faaf83b-b67f-454a-845f-d10861c6db57\") " pod="openshift-marketplace/redhat-operators-k4ggq" Oct 05 21:20:16 crc kubenswrapper[4754]: I1005 21:20:16.620173 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3faaf83b-b67f-454a-845f-d10861c6db57-utilities\") pod \"redhat-operators-k4ggq\" (UID: \"3faaf83b-b67f-454a-845f-d10861c6db57\") " pod="openshift-marketplace/redhat-operators-k4ggq" Oct 05 21:20:16 crc kubenswrapper[4754]: I1005 21:20:16.620856 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f48g8\" (UniqueName: \"kubernetes.io/projected/3faaf83b-b67f-454a-845f-d10861c6db57-kube-api-access-f48g8\") pod \"redhat-operators-k4ggq\" (UID: \"3faaf83b-b67f-454a-845f-d10861c6db57\") " pod="openshift-marketplace/redhat-operators-k4ggq" Oct 05 21:20:16 crc kubenswrapper[4754]: I1005 21:20:16.620946 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3faaf83b-b67f-454a-845f-d10861c6db57-utilities\") pod \"redhat-operators-k4ggq\" (UID: \"3faaf83b-b67f-454a-845f-d10861c6db57\") " pod="openshift-marketplace/redhat-operators-k4ggq" Oct 05 21:20:16 crc kubenswrapper[4754]: I1005 21:20:16.621204 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3faaf83b-b67f-454a-845f-d10861c6db57-catalog-content\") pod \"redhat-operators-k4ggq\" (UID: \"3faaf83b-b67f-454a-845f-d10861c6db57\") " pod="openshift-marketplace/redhat-operators-k4ggq" Oct 05 21:20:16 crc kubenswrapper[4754]: I1005 21:20:16.621560 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3faaf83b-b67f-454a-845f-d10861c6db57-catalog-content\") pod \"redhat-operators-k4ggq\" (UID: \"3faaf83b-b67f-454a-845f-d10861c6db57\") " pod="openshift-marketplace/redhat-operators-k4ggq" Oct 05 21:20:16 crc kubenswrapper[4754]: I1005 21:20:16.648122 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f48g8\" (UniqueName: \"kubernetes.io/projected/3faaf83b-b67f-454a-845f-d10861c6db57-kube-api-access-f48g8\") pod \"redhat-operators-k4ggq\" (UID: \"3faaf83b-b67f-454a-845f-d10861c6db57\") " pod="openshift-marketplace/redhat-operators-k4ggq" Oct 05 21:20:16 crc kubenswrapper[4754]: I1005 21:20:16.666222 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-k4ggq" Oct 05 21:20:17 crc kubenswrapper[4754]: I1005 21:20:17.225136 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-k4ggq"] Oct 05 21:20:17 crc kubenswrapper[4754]: I1005 21:20:17.752321 4754 generic.go:334] "Generic (PLEG): container finished" podID="3faaf83b-b67f-454a-845f-d10861c6db57" containerID="1214835d1a46d945d0ff1080e3cb3f8c028cb628bf9b4516a0eacdd4aa2039de" exitCode=0 Oct 05 21:20:17 crc kubenswrapper[4754]: I1005 21:20:17.752418 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k4ggq" event={"ID":"3faaf83b-b67f-454a-845f-d10861c6db57","Type":"ContainerDied","Data":"1214835d1a46d945d0ff1080e3cb3f8c028cb628bf9b4516a0eacdd4aa2039de"} Oct 05 21:20:17 crc kubenswrapper[4754]: I1005 21:20:17.752466 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k4ggq" event={"ID":"3faaf83b-b67f-454a-845f-d10861c6db57","Type":"ContainerStarted","Data":"61070943dfd24e6b210ae77f3fa9598d1160f42c5cf35bbed5a787faf57c2d86"} Oct 05 21:20:17 crc kubenswrapper[4754]: I1005 21:20:17.761027 4754 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 05 21:20:18 crc kubenswrapper[4754]: I1005 21:20:18.771562 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k4ggq" event={"ID":"3faaf83b-b67f-454a-845f-d10861c6db57","Type":"ContainerStarted","Data":"ea34437609298e591eafa0194e56303b000464466b9f13c2807aa7a8039e7bea"} Oct 05 21:20:22 crc kubenswrapper[4754]: I1005 21:20:22.829566 4754 generic.go:334] "Generic (PLEG): container finished" podID="3faaf83b-b67f-454a-845f-d10861c6db57" containerID="ea34437609298e591eafa0194e56303b000464466b9f13c2807aa7a8039e7bea" exitCode=0 Oct 05 21:20:22 crc kubenswrapper[4754]: I1005 21:20:22.829675 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k4ggq" event={"ID":"3faaf83b-b67f-454a-845f-d10861c6db57","Type":"ContainerDied","Data":"ea34437609298e591eafa0194e56303b000464466b9f13c2807aa7a8039e7bea"} Oct 05 21:20:23 crc kubenswrapper[4754]: I1005 21:20:23.853374 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k4ggq" event={"ID":"3faaf83b-b67f-454a-845f-d10861c6db57","Type":"ContainerStarted","Data":"cca9ded24faf0b7696fd4d12a04ee1f0822463109efb404012386e02667c1e01"} Oct 05 21:20:23 crc kubenswrapper[4754]: I1005 21:20:23.856164 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-qxm7h"] Oct 05 21:20:23 crc kubenswrapper[4754]: I1005 21:20:23.864253 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qxm7h" Oct 05 21:20:23 crc kubenswrapper[4754]: I1005 21:20:23.882733 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-qxm7h"] Oct 05 21:20:23 crc kubenswrapper[4754]: I1005 21:20:23.888211 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-k4ggq" podStartSLOduration=2.322133893 podStartE2EDuration="7.888197728s" podCreationTimestamp="2025-10-05 21:20:16 +0000 UTC" firstStartedPulling="2025-10-05 21:20:17.760662216 +0000 UTC m=+1541.664780926" lastFinishedPulling="2025-10-05 21:20:23.326726051 +0000 UTC m=+1547.230844761" observedRunningTime="2025-10-05 21:20:23.886060393 +0000 UTC m=+1547.790179103" watchObservedRunningTime="2025-10-05 21:20:23.888197728 +0000 UTC m=+1547.792316438" Oct 05 21:20:24 crc kubenswrapper[4754]: I1005 21:20:24.065645 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/679cb269-6827-4c74-8f4c-371103669036-utilities\") pod \"certified-operators-qxm7h\" (UID: \"679cb269-6827-4c74-8f4c-371103669036\") " pod="openshift-marketplace/certified-operators-qxm7h" Oct 05 21:20:24 crc kubenswrapper[4754]: I1005 21:20:24.065695 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/679cb269-6827-4c74-8f4c-371103669036-catalog-content\") pod \"certified-operators-qxm7h\" (UID: \"679cb269-6827-4c74-8f4c-371103669036\") " pod="openshift-marketplace/certified-operators-qxm7h" Oct 05 21:20:24 crc kubenswrapper[4754]: I1005 21:20:24.065818 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6vrwl\" (UniqueName: \"kubernetes.io/projected/679cb269-6827-4c74-8f4c-371103669036-kube-api-access-6vrwl\") pod \"certified-operators-qxm7h\" (UID: \"679cb269-6827-4c74-8f4c-371103669036\") " pod="openshift-marketplace/certified-operators-qxm7h" Oct 05 21:20:24 crc kubenswrapper[4754]: I1005 21:20:24.168314 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6vrwl\" (UniqueName: \"kubernetes.io/projected/679cb269-6827-4c74-8f4c-371103669036-kube-api-access-6vrwl\") pod \"certified-operators-qxm7h\" (UID: \"679cb269-6827-4c74-8f4c-371103669036\") " pod="openshift-marketplace/certified-operators-qxm7h" Oct 05 21:20:24 crc kubenswrapper[4754]: I1005 21:20:24.168453 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/679cb269-6827-4c74-8f4c-371103669036-utilities\") pod \"certified-operators-qxm7h\" (UID: \"679cb269-6827-4c74-8f4c-371103669036\") " pod="openshift-marketplace/certified-operators-qxm7h" Oct 05 21:20:24 crc kubenswrapper[4754]: I1005 21:20:24.168484 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/679cb269-6827-4c74-8f4c-371103669036-catalog-content\") pod \"certified-operators-qxm7h\" (UID: \"679cb269-6827-4c74-8f4c-371103669036\") " pod="openshift-marketplace/certified-operators-qxm7h" Oct 05 21:20:24 crc kubenswrapper[4754]: I1005 21:20:24.168961 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/679cb269-6827-4c74-8f4c-371103669036-catalog-content\") pod \"certified-operators-qxm7h\" (UID: \"679cb269-6827-4c74-8f4c-371103669036\") " pod="openshift-marketplace/certified-operators-qxm7h" Oct 05 21:20:24 crc kubenswrapper[4754]: I1005 21:20:24.168874 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/679cb269-6827-4c74-8f4c-371103669036-utilities\") pod \"certified-operators-qxm7h\" (UID: \"679cb269-6827-4c74-8f4c-371103669036\") " pod="openshift-marketplace/certified-operators-qxm7h" Oct 05 21:20:24 crc kubenswrapper[4754]: I1005 21:20:24.191338 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6vrwl\" (UniqueName: \"kubernetes.io/projected/679cb269-6827-4c74-8f4c-371103669036-kube-api-access-6vrwl\") pod \"certified-operators-qxm7h\" (UID: \"679cb269-6827-4c74-8f4c-371103669036\") " pod="openshift-marketplace/certified-operators-qxm7h" Oct 05 21:20:24 crc kubenswrapper[4754]: I1005 21:20:24.210969 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qxm7h" Oct 05 21:20:24 crc kubenswrapper[4754]: I1005 21:20:24.529767 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-qxm7h"] Oct 05 21:20:24 crc kubenswrapper[4754]: I1005 21:20:24.868754 4754 generic.go:334] "Generic (PLEG): container finished" podID="679cb269-6827-4c74-8f4c-371103669036" containerID="cb68190c97f028e9a410cd7f0199c3bf1062517938ca759818c37e1a6be29d10" exitCode=0 Oct 05 21:20:24 crc kubenswrapper[4754]: I1005 21:20:24.868865 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qxm7h" event={"ID":"679cb269-6827-4c74-8f4c-371103669036","Type":"ContainerDied","Data":"cb68190c97f028e9a410cd7f0199c3bf1062517938ca759818c37e1a6be29d10"} Oct 05 21:20:24 crc kubenswrapper[4754]: I1005 21:20:24.869118 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qxm7h" event={"ID":"679cb269-6827-4c74-8f4c-371103669036","Type":"ContainerStarted","Data":"8521d27760ed60156b151690454fd177ab039130f10c41e6b18b62523205d1d3"} Oct 05 21:20:25 crc kubenswrapper[4754]: I1005 21:20:25.915085 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qxm7h" event={"ID":"679cb269-6827-4c74-8f4c-371103669036","Type":"ContainerStarted","Data":"a7478730dadc023057598fde1c0c238fd9ed2e147f5918a39885f79182911762"} Oct 05 21:20:26 crc kubenswrapper[4754]: I1005 21:20:26.667413 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-k4ggq" Oct 05 21:20:26 crc kubenswrapper[4754]: I1005 21:20:26.667979 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-k4ggq" Oct 05 21:20:27 crc kubenswrapper[4754]: I1005 21:20:27.733521 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-k4ggq" podUID="3faaf83b-b67f-454a-845f-d10861c6db57" containerName="registry-server" probeResult="failure" output=< Oct 05 21:20:27 crc kubenswrapper[4754]: timeout: failed to connect service ":50051" within 1s Oct 05 21:20:27 crc kubenswrapper[4754]: > Oct 05 21:20:27 crc kubenswrapper[4754]: I1005 21:20:27.937802 4754 generic.go:334] "Generic (PLEG): container finished" podID="679cb269-6827-4c74-8f4c-371103669036" containerID="a7478730dadc023057598fde1c0c238fd9ed2e147f5918a39885f79182911762" exitCode=0 Oct 05 21:20:27 crc kubenswrapper[4754]: I1005 21:20:27.937868 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qxm7h" event={"ID":"679cb269-6827-4c74-8f4c-371103669036","Type":"ContainerDied","Data":"a7478730dadc023057598fde1c0c238fd9ed2e147f5918a39885f79182911762"} Oct 05 21:20:28 crc kubenswrapper[4754]: I1005 21:20:28.951678 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qxm7h" event={"ID":"679cb269-6827-4c74-8f4c-371103669036","Type":"ContainerStarted","Data":"9e0307e77e190aa70326f858ea3dd1eff68709084d286e520ebeb8f9b64202b9"} Oct 05 21:20:28 crc kubenswrapper[4754]: I1005 21:20:28.977105 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-qxm7h" podStartSLOduration=2.508352037 podStartE2EDuration="5.977079219s" podCreationTimestamp="2025-10-05 21:20:23 +0000 UTC" firstStartedPulling="2025-10-05 21:20:24.871334868 +0000 UTC m=+1548.775453578" lastFinishedPulling="2025-10-05 21:20:28.34006205 +0000 UTC m=+1552.244180760" observedRunningTime="2025-10-05 21:20:28.974450812 +0000 UTC m=+1552.878569532" watchObservedRunningTime="2025-10-05 21:20:28.977079219 +0000 UTC m=+1552.881197949" Oct 05 21:20:34 crc kubenswrapper[4754]: I1005 21:20:34.212068 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-qxm7h" Oct 05 21:20:34 crc kubenswrapper[4754]: I1005 21:20:34.212748 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-qxm7h" Oct 05 21:20:35 crc kubenswrapper[4754]: I1005 21:20:35.244739 4754 patch_prober.go:28] interesting pod/machine-config-daemon-b2h9k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 21:20:35 crc kubenswrapper[4754]: I1005 21:20:35.244821 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 21:20:35 crc kubenswrapper[4754]: I1005 21:20:35.244882 4754 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" Oct 05 21:20:35 crc kubenswrapper[4754]: I1005 21:20:35.245868 4754 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e8573d2f52a359de8bfb0cb5d4d1a72d653442930ef95a60a1d6d2dd509ab1ac"} pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 05 21:20:35 crc kubenswrapper[4754]: I1005 21:20:35.245969 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" containerName="machine-config-daemon" containerID="cri-o://e8573d2f52a359de8bfb0cb5d4d1a72d653442930ef95a60a1d6d2dd509ab1ac" gracePeriod=600 Oct 05 21:20:35 crc kubenswrapper[4754]: I1005 21:20:35.268854 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-qxm7h" podUID="679cb269-6827-4c74-8f4c-371103669036" containerName="registry-server" probeResult="failure" output=< Oct 05 21:20:35 crc kubenswrapper[4754]: timeout: failed to connect service ":50051" within 1s Oct 05 21:20:35 crc kubenswrapper[4754]: > Oct 05 21:20:35 crc kubenswrapper[4754]: E1005 21:20:35.378124 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 21:20:36 crc kubenswrapper[4754]: I1005 21:20:36.049667 4754 generic.go:334] "Generic (PLEG): container finished" podID="be95a413-6a59-45b4-84b7-b43cae694a26" containerID="e8573d2f52a359de8bfb0cb5d4d1a72d653442930ef95a60a1d6d2dd509ab1ac" exitCode=0 Oct 05 21:20:36 crc kubenswrapper[4754]: I1005 21:20:36.049772 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" event={"ID":"be95a413-6a59-45b4-84b7-b43cae694a26","Type":"ContainerDied","Data":"e8573d2f52a359de8bfb0cb5d4d1a72d653442930ef95a60a1d6d2dd509ab1ac"} Oct 05 21:20:36 crc kubenswrapper[4754]: I1005 21:20:36.050139 4754 scope.go:117] "RemoveContainer" containerID="a2229b13a2697fd212215ae5d196911669bc92f36ce70525da6f97d702c807b9" Oct 05 21:20:36 crc kubenswrapper[4754]: I1005 21:20:36.051444 4754 scope.go:117] "RemoveContainer" containerID="e8573d2f52a359de8bfb0cb5d4d1a72d653442930ef95a60a1d6d2dd509ab1ac" Oct 05 21:20:36 crc kubenswrapper[4754]: E1005 21:20:36.051907 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 21:20:37 crc kubenswrapper[4754]: I1005 21:20:37.735448 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-k4ggq" podUID="3faaf83b-b67f-454a-845f-d10861c6db57" containerName="registry-server" probeResult="failure" output=< Oct 05 21:20:37 crc kubenswrapper[4754]: timeout: failed to connect service ":50051" within 1s Oct 05 21:20:37 crc kubenswrapper[4754]: > Oct 05 21:20:44 crc kubenswrapper[4754]: I1005 21:20:44.259077 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-qxm7h" Oct 05 21:20:44 crc kubenswrapper[4754]: I1005 21:20:44.311171 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-qxm7h" Oct 05 21:20:44 crc kubenswrapper[4754]: I1005 21:20:44.510392 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-qxm7h"] Oct 05 21:20:46 crc kubenswrapper[4754]: I1005 21:20:46.158446 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-qxm7h" podUID="679cb269-6827-4c74-8f4c-371103669036" containerName="registry-server" containerID="cri-o://9e0307e77e190aa70326f858ea3dd1eff68709084d286e520ebeb8f9b64202b9" gracePeriod=2 Oct 05 21:20:46 crc kubenswrapper[4754]: I1005 21:20:46.651969 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qxm7h" Oct 05 21:20:46 crc kubenswrapper[4754]: I1005 21:20:46.776612 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/679cb269-6827-4c74-8f4c-371103669036-catalog-content\") pod \"679cb269-6827-4c74-8f4c-371103669036\" (UID: \"679cb269-6827-4c74-8f4c-371103669036\") " Oct 05 21:20:46 crc kubenswrapper[4754]: I1005 21:20:46.776966 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6vrwl\" (UniqueName: \"kubernetes.io/projected/679cb269-6827-4c74-8f4c-371103669036-kube-api-access-6vrwl\") pod \"679cb269-6827-4c74-8f4c-371103669036\" (UID: \"679cb269-6827-4c74-8f4c-371103669036\") " Oct 05 21:20:46 crc kubenswrapper[4754]: I1005 21:20:46.776998 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/679cb269-6827-4c74-8f4c-371103669036-utilities\") pod \"679cb269-6827-4c74-8f4c-371103669036\" (UID: \"679cb269-6827-4c74-8f4c-371103669036\") " Oct 05 21:20:46 crc kubenswrapper[4754]: I1005 21:20:46.778119 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/679cb269-6827-4c74-8f4c-371103669036-utilities" (OuterVolumeSpecName: "utilities") pod "679cb269-6827-4c74-8f4c-371103669036" (UID: "679cb269-6827-4c74-8f4c-371103669036"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 21:20:46 crc kubenswrapper[4754]: I1005 21:20:46.791669 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/679cb269-6827-4c74-8f4c-371103669036-kube-api-access-6vrwl" (OuterVolumeSpecName: "kube-api-access-6vrwl") pod "679cb269-6827-4c74-8f4c-371103669036" (UID: "679cb269-6827-4c74-8f4c-371103669036"). InnerVolumeSpecName "kube-api-access-6vrwl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:20:46 crc kubenswrapper[4754]: I1005 21:20:46.825838 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/679cb269-6827-4c74-8f4c-371103669036-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "679cb269-6827-4c74-8f4c-371103669036" (UID: "679cb269-6827-4c74-8f4c-371103669036"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 21:20:46 crc kubenswrapper[4754]: I1005 21:20:46.879371 4754 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/679cb269-6827-4c74-8f4c-371103669036-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 21:20:46 crc kubenswrapper[4754]: I1005 21:20:46.879423 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6vrwl\" (UniqueName: \"kubernetes.io/projected/679cb269-6827-4c74-8f4c-371103669036-kube-api-access-6vrwl\") on node \"crc\" DevicePath \"\"" Oct 05 21:20:46 crc kubenswrapper[4754]: I1005 21:20:46.879447 4754 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/679cb269-6827-4c74-8f4c-371103669036-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 21:20:47 crc kubenswrapper[4754]: I1005 21:20:47.175371 4754 generic.go:334] "Generic (PLEG): container finished" podID="679cb269-6827-4c74-8f4c-371103669036" containerID="9e0307e77e190aa70326f858ea3dd1eff68709084d286e520ebeb8f9b64202b9" exitCode=0 Oct 05 21:20:47 crc kubenswrapper[4754]: I1005 21:20:47.175440 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qxm7h" event={"ID":"679cb269-6827-4c74-8f4c-371103669036","Type":"ContainerDied","Data":"9e0307e77e190aa70326f858ea3dd1eff68709084d286e520ebeb8f9b64202b9"} Oct 05 21:20:47 crc kubenswrapper[4754]: I1005 21:20:47.175487 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qxm7h" event={"ID":"679cb269-6827-4c74-8f4c-371103669036","Type":"ContainerDied","Data":"8521d27760ed60156b151690454fd177ab039130f10c41e6b18b62523205d1d3"} Oct 05 21:20:47 crc kubenswrapper[4754]: I1005 21:20:47.175540 4754 scope.go:117] "RemoveContainer" containerID="9e0307e77e190aa70326f858ea3dd1eff68709084d286e520ebeb8f9b64202b9" Oct 05 21:20:47 crc kubenswrapper[4754]: I1005 21:20:47.175485 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qxm7h" Oct 05 21:20:47 crc kubenswrapper[4754]: I1005 21:20:47.212671 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-qxm7h"] Oct 05 21:20:47 crc kubenswrapper[4754]: I1005 21:20:47.225760 4754 scope.go:117] "RemoveContainer" containerID="a7478730dadc023057598fde1c0c238fd9ed2e147f5918a39885f79182911762" Oct 05 21:20:47 crc kubenswrapper[4754]: I1005 21:20:47.230313 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-qxm7h"] Oct 05 21:20:47 crc kubenswrapper[4754]: I1005 21:20:47.252758 4754 scope.go:117] "RemoveContainer" containerID="cb68190c97f028e9a410cd7f0199c3bf1062517938ca759818c37e1a6be29d10" Oct 05 21:20:47 crc kubenswrapper[4754]: I1005 21:20:47.300909 4754 scope.go:117] "RemoveContainer" containerID="9e0307e77e190aa70326f858ea3dd1eff68709084d286e520ebeb8f9b64202b9" Oct 05 21:20:47 crc kubenswrapper[4754]: E1005 21:20:47.301670 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9e0307e77e190aa70326f858ea3dd1eff68709084d286e520ebeb8f9b64202b9\": container with ID starting with 9e0307e77e190aa70326f858ea3dd1eff68709084d286e520ebeb8f9b64202b9 not found: ID does not exist" containerID="9e0307e77e190aa70326f858ea3dd1eff68709084d286e520ebeb8f9b64202b9" Oct 05 21:20:47 crc kubenswrapper[4754]: I1005 21:20:47.301787 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9e0307e77e190aa70326f858ea3dd1eff68709084d286e520ebeb8f9b64202b9"} err="failed to get container status \"9e0307e77e190aa70326f858ea3dd1eff68709084d286e520ebeb8f9b64202b9\": rpc error: code = NotFound desc = could not find container \"9e0307e77e190aa70326f858ea3dd1eff68709084d286e520ebeb8f9b64202b9\": container with ID starting with 9e0307e77e190aa70326f858ea3dd1eff68709084d286e520ebeb8f9b64202b9 not found: ID does not exist" Oct 05 21:20:47 crc kubenswrapper[4754]: I1005 21:20:47.301890 4754 scope.go:117] "RemoveContainer" containerID="a7478730dadc023057598fde1c0c238fd9ed2e147f5918a39885f79182911762" Oct 05 21:20:47 crc kubenswrapper[4754]: E1005 21:20:47.302795 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a7478730dadc023057598fde1c0c238fd9ed2e147f5918a39885f79182911762\": container with ID starting with a7478730dadc023057598fde1c0c238fd9ed2e147f5918a39885f79182911762 not found: ID does not exist" containerID="a7478730dadc023057598fde1c0c238fd9ed2e147f5918a39885f79182911762" Oct 05 21:20:47 crc kubenswrapper[4754]: I1005 21:20:47.302826 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a7478730dadc023057598fde1c0c238fd9ed2e147f5918a39885f79182911762"} err="failed to get container status \"a7478730dadc023057598fde1c0c238fd9ed2e147f5918a39885f79182911762\": rpc error: code = NotFound desc = could not find container \"a7478730dadc023057598fde1c0c238fd9ed2e147f5918a39885f79182911762\": container with ID starting with a7478730dadc023057598fde1c0c238fd9ed2e147f5918a39885f79182911762 not found: ID does not exist" Oct 05 21:20:47 crc kubenswrapper[4754]: I1005 21:20:47.302850 4754 scope.go:117] "RemoveContainer" containerID="cb68190c97f028e9a410cd7f0199c3bf1062517938ca759818c37e1a6be29d10" Oct 05 21:20:47 crc kubenswrapper[4754]: E1005 21:20:47.303378 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cb68190c97f028e9a410cd7f0199c3bf1062517938ca759818c37e1a6be29d10\": container with ID starting with cb68190c97f028e9a410cd7f0199c3bf1062517938ca759818c37e1a6be29d10 not found: ID does not exist" containerID="cb68190c97f028e9a410cd7f0199c3bf1062517938ca759818c37e1a6be29d10" Oct 05 21:20:47 crc kubenswrapper[4754]: I1005 21:20:47.303423 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cb68190c97f028e9a410cd7f0199c3bf1062517938ca759818c37e1a6be29d10"} err="failed to get container status \"cb68190c97f028e9a410cd7f0199c3bf1062517938ca759818c37e1a6be29d10\": rpc error: code = NotFound desc = could not find container \"cb68190c97f028e9a410cd7f0199c3bf1062517938ca759818c37e1a6be29d10\": container with ID starting with cb68190c97f028e9a410cd7f0199c3bf1062517938ca759818c37e1a6be29d10 not found: ID does not exist" Oct 05 21:20:47 crc kubenswrapper[4754]: I1005 21:20:47.722078 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-k4ggq" podUID="3faaf83b-b67f-454a-845f-d10861c6db57" containerName="registry-server" probeResult="failure" output=< Oct 05 21:20:47 crc kubenswrapper[4754]: timeout: failed to connect service ":50051" within 1s Oct 05 21:20:47 crc kubenswrapper[4754]: > Oct 05 21:20:47 crc kubenswrapper[4754]: I1005 21:20:47.837889 4754 scope.go:117] "RemoveContainer" containerID="e8573d2f52a359de8bfb0cb5d4d1a72d653442930ef95a60a1d6d2dd509ab1ac" Oct 05 21:20:47 crc kubenswrapper[4754]: E1005 21:20:47.838696 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 21:20:48 crc kubenswrapper[4754]: I1005 21:20:48.850651 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="679cb269-6827-4c74-8f4c-371103669036" path="/var/lib/kubelet/pods/679cb269-6827-4c74-8f4c-371103669036/volumes" Oct 05 21:20:53 crc kubenswrapper[4754]: I1005 21:20:53.539934 4754 scope.go:117] "RemoveContainer" containerID="67ec590285159d5dd7ffbf3414c3f3154b09a08128f4538272dd9316d970ffd3" Oct 05 21:20:53 crc kubenswrapper[4754]: I1005 21:20:53.581425 4754 scope.go:117] "RemoveContainer" containerID="41c9b85f7408628829308d18d3aeee6968858f7de70656a6fe724f64f802b04f" Oct 05 21:20:53 crc kubenswrapper[4754]: I1005 21:20:53.613343 4754 scope.go:117] "RemoveContainer" containerID="c81e9c627545c8ae7fec11f2768aba757f9b5f96a53410f2bbb3e6958d352382" Oct 05 21:20:53 crc kubenswrapper[4754]: I1005 21:20:53.645454 4754 scope.go:117] "RemoveContainer" containerID="e6ac7457d8a80fcb4fc42b1237603228691d5ff6f9672e3467bd4592dc1d0478" Oct 05 21:20:56 crc kubenswrapper[4754]: I1005 21:20:56.054193 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-g8hwz"] Oct 05 21:20:56 crc kubenswrapper[4754]: I1005 21:20:56.067282 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-g8hwz"] Oct 05 21:20:56 crc kubenswrapper[4754]: I1005 21:20:56.758223 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-k4ggq" Oct 05 21:20:56 crc kubenswrapper[4754]: I1005 21:20:56.858209 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b8d124eb-a543-4988-9523-d0e37ca113b8" path="/var/lib/kubelet/pods/b8d124eb-a543-4988-9523-d0e37ca113b8/volumes" Oct 05 21:20:56 crc kubenswrapper[4754]: I1005 21:20:56.859683 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-k4ggq" Oct 05 21:20:57 crc kubenswrapper[4754]: I1005 21:20:57.005511 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-k4ggq"] Oct 05 21:20:57 crc kubenswrapper[4754]: I1005 21:20:57.034467 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-5h8xj"] Oct 05 21:20:57 crc kubenswrapper[4754]: I1005 21:20:57.046086 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-c7gj8"] Oct 05 21:20:57 crc kubenswrapper[4754]: I1005 21:20:57.056139 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-c7gj8"] Oct 05 21:20:57 crc kubenswrapper[4754]: I1005 21:20:57.064889 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-5h8xj"] Oct 05 21:20:58 crc kubenswrapper[4754]: I1005 21:20:58.299371 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-k4ggq" podUID="3faaf83b-b67f-454a-845f-d10861c6db57" containerName="registry-server" containerID="cri-o://cca9ded24faf0b7696fd4d12a04ee1f0822463109efb404012386e02667c1e01" gracePeriod=2 Oct 05 21:20:58 crc kubenswrapper[4754]: I1005 21:20:58.796508 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-k4ggq" Oct 05 21:20:58 crc kubenswrapper[4754]: I1005 21:20:58.882131 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="28dab52f-2bfa-409f-8ec3-b47ae0641ce2" path="/var/lib/kubelet/pods/28dab52f-2bfa-409f-8ec3-b47ae0641ce2/volumes" Oct 05 21:20:58 crc kubenswrapper[4754]: I1005 21:20:58.883711 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cee454de-74cf-4913-8792-a231ddd4c691" path="/var/lib/kubelet/pods/cee454de-74cf-4913-8792-a231ddd4c691/volumes" Oct 05 21:20:58 crc kubenswrapper[4754]: I1005 21:20:58.961057 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3faaf83b-b67f-454a-845f-d10861c6db57-catalog-content\") pod \"3faaf83b-b67f-454a-845f-d10861c6db57\" (UID: \"3faaf83b-b67f-454a-845f-d10861c6db57\") " Oct 05 21:20:58 crc kubenswrapper[4754]: I1005 21:20:58.961100 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3faaf83b-b67f-454a-845f-d10861c6db57-utilities\") pod \"3faaf83b-b67f-454a-845f-d10861c6db57\" (UID: \"3faaf83b-b67f-454a-845f-d10861c6db57\") " Oct 05 21:20:58 crc kubenswrapper[4754]: I1005 21:20:58.961176 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f48g8\" (UniqueName: \"kubernetes.io/projected/3faaf83b-b67f-454a-845f-d10861c6db57-kube-api-access-f48g8\") pod \"3faaf83b-b67f-454a-845f-d10861c6db57\" (UID: \"3faaf83b-b67f-454a-845f-d10861c6db57\") " Oct 05 21:20:58 crc kubenswrapper[4754]: I1005 21:20:58.964903 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3faaf83b-b67f-454a-845f-d10861c6db57-utilities" (OuterVolumeSpecName: "utilities") pod "3faaf83b-b67f-454a-845f-d10861c6db57" (UID: "3faaf83b-b67f-454a-845f-d10861c6db57"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 21:20:58 crc kubenswrapper[4754]: I1005 21:20:58.966885 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3faaf83b-b67f-454a-845f-d10861c6db57-kube-api-access-f48g8" (OuterVolumeSpecName: "kube-api-access-f48g8") pod "3faaf83b-b67f-454a-845f-d10861c6db57" (UID: "3faaf83b-b67f-454a-845f-d10861c6db57"). InnerVolumeSpecName "kube-api-access-f48g8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:20:59 crc kubenswrapper[4754]: I1005 21:20:59.038230 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3faaf83b-b67f-454a-845f-d10861c6db57-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3faaf83b-b67f-454a-845f-d10861c6db57" (UID: "3faaf83b-b67f-454a-845f-d10861c6db57"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 21:20:59 crc kubenswrapper[4754]: I1005 21:20:59.064146 4754 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3faaf83b-b67f-454a-845f-d10861c6db57-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 21:20:59 crc kubenswrapper[4754]: I1005 21:20:59.064174 4754 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3faaf83b-b67f-454a-845f-d10861c6db57-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 21:20:59 crc kubenswrapper[4754]: I1005 21:20:59.064185 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f48g8\" (UniqueName: \"kubernetes.io/projected/3faaf83b-b67f-454a-845f-d10861c6db57-kube-api-access-f48g8\") on node \"crc\" DevicePath \"\"" Oct 05 21:20:59 crc kubenswrapper[4754]: I1005 21:20:59.317286 4754 generic.go:334] "Generic (PLEG): container finished" podID="3faaf83b-b67f-454a-845f-d10861c6db57" containerID="cca9ded24faf0b7696fd4d12a04ee1f0822463109efb404012386e02667c1e01" exitCode=0 Oct 05 21:20:59 crc kubenswrapper[4754]: I1005 21:20:59.317350 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k4ggq" event={"ID":"3faaf83b-b67f-454a-845f-d10861c6db57","Type":"ContainerDied","Data":"cca9ded24faf0b7696fd4d12a04ee1f0822463109efb404012386e02667c1e01"} Oct 05 21:20:59 crc kubenswrapper[4754]: I1005 21:20:59.317392 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k4ggq" event={"ID":"3faaf83b-b67f-454a-845f-d10861c6db57","Type":"ContainerDied","Data":"61070943dfd24e6b210ae77f3fa9598d1160f42c5cf35bbed5a787faf57c2d86"} Oct 05 21:20:59 crc kubenswrapper[4754]: I1005 21:20:59.317423 4754 scope.go:117] "RemoveContainer" containerID="cca9ded24faf0b7696fd4d12a04ee1f0822463109efb404012386e02667c1e01" Oct 05 21:20:59 crc kubenswrapper[4754]: I1005 21:20:59.317685 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-k4ggq" Oct 05 21:20:59 crc kubenswrapper[4754]: I1005 21:20:59.377394 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-k4ggq"] Oct 05 21:20:59 crc kubenswrapper[4754]: I1005 21:20:59.386528 4754 scope.go:117] "RemoveContainer" containerID="ea34437609298e591eafa0194e56303b000464466b9f13c2807aa7a8039e7bea" Oct 05 21:20:59 crc kubenswrapper[4754]: I1005 21:20:59.389554 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-k4ggq"] Oct 05 21:20:59 crc kubenswrapper[4754]: I1005 21:20:59.416327 4754 scope.go:117] "RemoveContainer" containerID="1214835d1a46d945d0ff1080e3cb3f8c028cb628bf9b4516a0eacdd4aa2039de" Oct 05 21:20:59 crc kubenswrapper[4754]: I1005 21:20:59.466031 4754 scope.go:117] "RemoveContainer" containerID="cca9ded24faf0b7696fd4d12a04ee1f0822463109efb404012386e02667c1e01" Oct 05 21:20:59 crc kubenswrapper[4754]: E1005 21:20:59.466680 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cca9ded24faf0b7696fd4d12a04ee1f0822463109efb404012386e02667c1e01\": container with ID starting with cca9ded24faf0b7696fd4d12a04ee1f0822463109efb404012386e02667c1e01 not found: ID does not exist" containerID="cca9ded24faf0b7696fd4d12a04ee1f0822463109efb404012386e02667c1e01" Oct 05 21:20:59 crc kubenswrapper[4754]: I1005 21:20:59.466836 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cca9ded24faf0b7696fd4d12a04ee1f0822463109efb404012386e02667c1e01"} err="failed to get container status \"cca9ded24faf0b7696fd4d12a04ee1f0822463109efb404012386e02667c1e01\": rpc error: code = NotFound desc = could not find container \"cca9ded24faf0b7696fd4d12a04ee1f0822463109efb404012386e02667c1e01\": container with ID starting with cca9ded24faf0b7696fd4d12a04ee1f0822463109efb404012386e02667c1e01 not found: ID does not exist" Oct 05 21:20:59 crc kubenswrapper[4754]: I1005 21:20:59.466946 4754 scope.go:117] "RemoveContainer" containerID="ea34437609298e591eafa0194e56303b000464466b9f13c2807aa7a8039e7bea" Oct 05 21:20:59 crc kubenswrapper[4754]: E1005 21:20:59.467449 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ea34437609298e591eafa0194e56303b000464466b9f13c2807aa7a8039e7bea\": container with ID starting with ea34437609298e591eafa0194e56303b000464466b9f13c2807aa7a8039e7bea not found: ID does not exist" containerID="ea34437609298e591eafa0194e56303b000464466b9f13c2807aa7a8039e7bea" Oct 05 21:20:59 crc kubenswrapper[4754]: I1005 21:20:59.467506 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ea34437609298e591eafa0194e56303b000464466b9f13c2807aa7a8039e7bea"} err="failed to get container status \"ea34437609298e591eafa0194e56303b000464466b9f13c2807aa7a8039e7bea\": rpc error: code = NotFound desc = could not find container \"ea34437609298e591eafa0194e56303b000464466b9f13c2807aa7a8039e7bea\": container with ID starting with ea34437609298e591eafa0194e56303b000464466b9f13c2807aa7a8039e7bea not found: ID does not exist" Oct 05 21:20:59 crc kubenswrapper[4754]: I1005 21:20:59.467532 4754 scope.go:117] "RemoveContainer" containerID="1214835d1a46d945d0ff1080e3cb3f8c028cb628bf9b4516a0eacdd4aa2039de" Oct 05 21:20:59 crc kubenswrapper[4754]: E1005 21:20:59.467859 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1214835d1a46d945d0ff1080e3cb3f8c028cb628bf9b4516a0eacdd4aa2039de\": container with ID starting with 1214835d1a46d945d0ff1080e3cb3f8c028cb628bf9b4516a0eacdd4aa2039de not found: ID does not exist" containerID="1214835d1a46d945d0ff1080e3cb3f8c028cb628bf9b4516a0eacdd4aa2039de" Oct 05 21:20:59 crc kubenswrapper[4754]: I1005 21:20:59.467982 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1214835d1a46d945d0ff1080e3cb3f8c028cb628bf9b4516a0eacdd4aa2039de"} err="failed to get container status \"1214835d1a46d945d0ff1080e3cb3f8c028cb628bf9b4516a0eacdd4aa2039de\": rpc error: code = NotFound desc = could not find container \"1214835d1a46d945d0ff1080e3cb3f8c028cb628bf9b4516a0eacdd4aa2039de\": container with ID starting with 1214835d1a46d945d0ff1080e3cb3f8c028cb628bf9b4516a0eacdd4aa2039de not found: ID does not exist" Oct 05 21:21:00 crc kubenswrapper[4754]: I1005 21:21:00.856662 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3faaf83b-b67f-454a-845f-d10861c6db57" path="/var/lib/kubelet/pods/3faaf83b-b67f-454a-845f-d10861c6db57/volumes" Oct 05 21:21:02 crc kubenswrapper[4754]: I1005 21:21:02.838227 4754 scope.go:117] "RemoveContainer" containerID="e8573d2f52a359de8bfb0cb5d4d1a72d653442930ef95a60a1d6d2dd509ab1ac" Oct 05 21:21:02 crc kubenswrapper[4754]: E1005 21:21:02.840583 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 21:21:07 crc kubenswrapper[4754]: I1005 21:21:07.035771 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-5428-account-create-gbvmw"] Oct 05 21:21:07 crc kubenswrapper[4754]: I1005 21:21:07.044329 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-5428-account-create-gbvmw"] Oct 05 21:21:08 crc kubenswrapper[4754]: I1005 21:21:08.048044 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-e07b-account-create-wkgh8"] Oct 05 21:21:08 crc kubenswrapper[4754]: I1005 21:21:08.060004 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-8228-account-create-pf59x"] Oct 05 21:21:08 crc kubenswrapper[4754]: I1005 21:21:08.077907 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-8228-account-create-pf59x"] Oct 05 21:21:08 crc kubenswrapper[4754]: I1005 21:21:08.089448 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-e07b-account-create-wkgh8"] Oct 05 21:21:08 crc kubenswrapper[4754]: I1005 21:21:08.853139 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="21391bdb-ce3c-45ed-bc20-1bccdbcc3c72" path="/var/lib/kubelet/pods/21391bdb-ce3c-45ed-bc20-1bccdbcc3c72/volumes" Oct 05 21:21:08 crc kubenswrapper[4754]: I1005 21:21:08.855777 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b380a91-da43-41a4-afcf-3d41e99b094f" path="/var/lib/kubelet/pods/5b380a91-da43-41a4-afcf-3d41e99b094f/volumes" Oct 05 21:21:08 crc kubenswrapper[4754]: I1005 21:21:08.859309 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e3ddc43f-cb95-4106-a5ff-09fc6320aad8" path="/var/lib/kubelet/pods/e3ddc43f-cb95-4106-a5ff-09fc6320aad8/volumes" Oct 05 21:21:14 crc kubenswrapper[4754]: I1005 21:21:14.843571 4754 scope.go:117] "RemoveContainer" containerID="e8573d2f52a359de8bfb0cb5d4d1a72d653442930ef95a60a1d6d2dd509ab1ac" Oct 05 21:21:14 crc kubenswrapper[4754]: E1005 21:21:14.844255 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 21:21:25 crc kubenswrapper[4754]: I1005 21:21:25.837681 4754 scope.go:117] "RemoveContainer" containerID="e8573d2f52a359de8bfb0cb5d4d1a72d653442930ef95a60a1d6d2dd509ab1ac" Oct 05 21:21:25 crc kubenswrapper[4754]: E1005 21:21:25.838616 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 21:21:30 crc kubenswrapper[4754]: I1005 21:21:30.041759 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-679wz"] Oct 05 21:21:30 crc kubenswrapper[4754]: I1005 21:21:30.055479 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-679wz"] Oct 05 21:21:30 crc kubenswrapper[4754]: I1005 21:21:30.873145 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9a72c674-5986-45a3-aae6-7f15de18efed" path="/var/lib/kubelet/pods/9a72c674-5986-45a3-aae6-7f15de18efed/volumes" Oct 05 21:21:36 crc kubenswrapper[4754]: I1005 21:21:36.047463 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-5mdmj"] Oct 05 21:21:36 crc kubenswrapper[4754]: I1005 21:21:36.057819 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-qpgn5"] Oct 05 21:21:36 crc kubenswrapper[4754]: I1005 21:21:36.072232 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-5mdmj"] Oct 05 21:21:36 crc kubenswrapper[4754]: I1005 21:21:36.081240 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-qpgn5"] Oct 05 21:21:36 crc kubenswrapper[4754]: I1005 21:21:36.845034 4754 scope.go:117] "RemoveContainer" containerID="e8573d2f52a359de8bfb0cb5d4d1a72d653442930ef95a60a1d6d2dd509ab1ac" Oct 05 21:21:36 crc kubenswrapper[4754]: E1005 21:21:36.845385 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 21:21:36 crc kubenswrapper[4754]: I1005 21:21:36.860648 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0f34b855-e478-41a0-b7b4-28dcf667b7c2" path="/var/lib/kubelet/pods/0f34b855-e478-41a0-b7b4-28dcf667b7c2/volumes" Oct 05 21:21:36 crc kubenswrapper[4754]: I1005 21:21:36.865103 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c0678fc1-b079-4379-8bbc-8848c3848f02" path="/var/lib/kubelet/pods/c0678fc1-b079-4379-8bbc-8848c3848f02/volumes" Oct 05 21:21:41 crc kubenswrapper[4754]: I1005 21:21:41.030640 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-9fzft"] Oct 05 21:21:41 crc kubenswrapper[4754]: I1005 21:21:41.040630 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-9fzft"] Oct 05 21:21:42 crc kubenswrapper[4754]: I1005 21:21:42.030109 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-dvzwp"] Oct 05 21:21:42 crc kubenswrapper[4754]: I1005 21:21:42.038789 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-dvzwp"] Oct 05 21:21:42 crc kubenswrapper[4754]: I1005 21:21:42.852993 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87ecff30-d447-417b-9fe5-97cd6fd64c93" path="/var/lib/kubelet/pods/87ecff30-d447-417b-9fe5-97cd6fd64c93/volumes" Oct 05 21:21:42 crc kubenswrapper[4754]: I1005 21:21:42.864407 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d20989a7-98b9-4bf2-9f85-d72890ef7e1a" path="/var/lib/kubelet/pods/d20989a7-98b9-4bf2-9f85-d72890ef7e1a/volumes" Oct 05 21:21:44 crc kubenswrapper[4754]: I1005 21:21:44.892874 4754 generic.go:334] "Generic (PLEG): container finished" podID="31cbd0a7-7e3d-4bd6-b84c-816431fdc633" containerID="c0e9288c9e1c394fde84639e37dc3c511c39814f13c421cee5f5cb20af0e301a" exitCode=0 Oct 05 21:21:44 crc kubenswrapper[4754]: I1005 21:21:44.893016 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-6cdph" event={"ID":"31cbd0a7-7e3d-4bd6-b84c-816431fdc633","Type":"ContainerDied","Data":"c0e9288c9e1c394fde84639e37dc3c511c39814f13c421cee5f5cb20af0e301a"} Oct 05 21:21:46 crc kubenswrapper[4754]: I1005 21:21:46.045722 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-3385-account-create-62sw4"] Oct 05 21:21:46 crc kubenswrapper[4754]: I1005 21:21:46.052244 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-3385-account-create-62sw4"] Oct 05 21:21:46 crc kubenswrapper[4754]: I1005 21:21:46.401029 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-6cdph" Oct 05 21:21:46 crc kubenswrapper[4754]: I1005 21:21:46.529012 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dn8kz\" (UniqueName: \"kubernetes.io/projected/31cbd0a7-7e3d-4bd6-b84c-816431fdc633-kube-api-access-dn8kz\") pod \"31cbd0a7-7e3d-4bd6-b84c-816431fdc633\" (UID: \"31cbd0a7-7e3d-4bd6-b84c-816431fdc633\") " Oct 05 21:21:46 crc kubenswrapper[4754]: I1005 21:21:46.529112 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/31cbd0a7-7e3d-4bd6-b84c-816431fdc633-inventory\") pod \"31cbd0a7-7e3d-4bd6-b84c-816431fdc633\" (UID: \"31cbd0a7-7e3d-4bd6-b84c-816431fdc633\") " Oct 05 21:21:46 crc kubenswrapper[4754]: I1005 21:21:46.529618 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31cbd0a7-7e3d-4bd6-b84c-816431fdc633-bootstrap-combined-ca-bundle\") pod \"31cbd0a7-7e3d-4bd6-b84c-816431fdc633\" (UID: \"31cbd0a7-7e3d-4bd6-b84c-816431fdc633\") " Oct 05 21:21:46 crc kubenswrapper[4754]: I1005 21:21:46.529731 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/31cbd0a7-7e3d-4bd6-b84c-816431fdc633-ssh-key\") pod \"31cbd0a7-7e3d-4bd6-b84c-816431fdc633\" (UID: \"31cbd0a7-7e3d-4bd6-b84c-816431fdc633\") " Oct 05 21:21:46 crc kubenswrapper[4754]: I1005 21:21:46.535776 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31cbd0a7-7e3d-4bd6-b84c-816431fdc633-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "31cbd0a7-7e3d-4bd6-b84c-816431fdc633" (UID: "31cbd0a7-7e3d-4bd6-b84c-816431fdc633"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:21:46 crc kubenswrapper[4754]: I1005 21:21:46.537551 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31cbd0a7-7e3d-4bd6-b84c-816431fdc633-kube-api-access-dn8kz" (OuterVolumeSpecName: "kube-api-access-dn8kz") pod "31cbd0a7-7e3d-4bd6-b84c-816431fdc633" (UID: "31cbd0a7-7e3d-4bd6-b84c-816431fdc633"). InnerVolumeSpecName "kube-api-access-dn8kz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:21:46 crc kubenswrapper[4754]: I1005 21:21:46.558809 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31cbd0a7-7e3d-4bd6-b84c-816431fdc633-inventory" (OuterVolumeSpecName: "inventory") pod "31cbd0a7-7e3d-4bd6-b84c-816431fdc633" (UID: "31cbd0a7-7e3d-4bd6-b84c-816431fdc633"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:21:46 crc kubenswrapper[4754]: I1005 21:21:46.568535 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31cbd0a7-7e3d-4bd6-b84c-816431fdc633-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "31cbd0a7-7e3d-4bd6-b84c-816431fdc633" (UID: "31cbd0a7-7e3d-4bd6-b84c-816431fdc633"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:21:46 crc kubenswrapper[4754]: I1005 21:21:46.633745 4754 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31cbd0a7-7e3d-4bd6-b84c-816431fdc633-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 21:21:46 crc kubenswrapper[4754]: I1005 21:21:46.633785 4754 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/31cbd0a7-7e3d-4bd6-b84c-816431fdc633-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 05 21:21:46 crc kubenswrapper[4754]: I1005 21:21:46.633796 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dn8kz\" (UniqueName: \"kubernetes.io/projected/31cbd0a7-7e3d-4bd6-b84c-816431fdc633-kube-api-access-dn8kz\") on node \"crc\" DevicePath \"\"" Oct 05 21:21:46 crc kubenswrapper[4754]: I1005 21:21:46.633805 4754 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/31cbd0a7-7e3d-4bd6-b84c-816431fdc633-inventory\") on node \"crc\" DevicePath \"\"" Oct 05 21:21:46 crc kubenswrapper[4754]: I1005 21:21:46.859635 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b7ded8c6-2bd5-46ca-b36f-68ccbd92366d" path="/var/lib/kubelet/pods/b7ded8c6-2bd5-46ca-b36f-68ccbd92366d/volumes" Oct 05 21:21:46 crc kubenswrapper[4754]: I1005 21:21:46.926353 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-6cdph" event={"ID":"31cbd0a7-7e3d-4bd6-b84c-816431fdc633","Type":"ContainerDied","Data":"9a0ee2c55ece169b6695e7942283404a8970160f609be2d9a38e6862366d945d"} Oct 05 21:21:46 crc kubenswrapper[4754]: I1005 21:21:46.926426 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9a0ee2c55ece169b6695e7942283404a8970160f609be2d9a38e6862366d945d" Oct 05 21:21:46 crc kubenswrapper[4754]: I1005 21:21:46.926578 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-6cdph" Oct 05 21:21:47 crc kubenswrapper[4754]: I1005 21:21:47.046617 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-dfec-account-create-qdgw4"] Oct 05 21:21:47 crc kubenswrapper[4754]: I1005 21:21:47.094014 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-dfec-account-create-qdgw4"] Oct 05 21:21:47 crc kubenswrapper[4754]: I1005 21:21:47.094102 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-5x9f9"] Oct 05 21:21:47 crc kubenswrapper[4754]: E1005 21:21:47.095717 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3faaf83b-b67f-454a-845f-d10861c6db57" containerName="registry-server" Oct 05 21:21:47 crc kubenswrapper[4754]: I1005 21:21:47.095752 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="3faaf83b-b67f-454a-845f-d10861c6db57" containerName="registry-server" Oct 05 21:21:47 crc kubenswrapper[4754]: E1005 21:21:47.095799 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31cbd0a7-7e3d-4bd6-b84c-816431fdc633" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Oct 05 21:21:47 crc kubenswrapper[4754]: I1005 21:21:47.095810 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="31cbd0a7-7e3d-4bd6-b84c-816431fdc633" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Oct 05 21:21:47 crc kubenswrapper[4754]: E1005 21:21:47.095820 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="679cb269-6827-4c74-8f4c-371103669036" containerName="extract-content" Oct 05 21:21:47 crc kubenswrapper[4754]: I1005 21:21:47.095829 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="679cb269-6827-4c74-8f4c-371103669036" containerName="extract-content" Oct 05 21:21:47 crc kubenswrapper[4754]: E1005 21:21:47.095852 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="679cb269-6827-4c74-8f4c-371103669036" containerName="registry-server" Oct 05 21:21:47 crc kubenswrapper[4754]: I1005 21:21:47.095860 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="679cb269-6827-4c74-8f4c-371103669036" containerName="registry-server" Oct 05 21:21:47 crc kubenswrapper[4754]: E1005 21:21:47.095886 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3faaf83b-b67f-454a-845f-d10861c6db57" containerName="extract-utilities" Oct 05 21:21:47 crc kubenswrapper[4754]: I1005 21:21:47.095897 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="3faaf83b-b67f-454a-845f-d10861c6db57" containerName="extract-utilities" Oct 05 21:21:47 crc kubenswrapper[4754]: E1005 21:21:47.095922 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="679cb269-6827-4c74-8f4c-371103669036" containerName="extract-utilities" Oct 05 21:21:47 crc kubenswrapper[4754]: I1005 21:21:47.095931 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="679cb269-6827-4c74-8f4c-371103669036" containerName="extract-utilities" Oct 05 21:21:47 crc kubenswrapper[4754]: E1005 21:21:47.095974 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3faaf83b-b67f-454a-845f-d10861c6db57" containerName="extract-content" Oct 05 21:21:47 crc kubenswrapper[4754]: I1005 21:21:47.095983 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="3faaf83b-b67f-454a-845f-d10861c6db57" containerName="extract-content" Oct 05 21:21:47 crc kubenswrapper[4754]: I1005 21:21:47.096665 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="31cbd0a7-7e3d-4bd6-b84c-816431fdc633" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Oct 05 21:21:47 crc kubenswrapper[4754]: I1005 21:21:47.096709 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="679cb269-6827-4c74-8f4c-371103669036" containerName="registry-server" Oct 05 21:21:47 crc kubenswrapper[4754]: I1005 21:21:47.096732 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="3faaf83b-b67f-454a-845f-d10861c6db57" containerName="registry-server" Oct 05 21:21:47 crc kubenswrapper[4754]: I1005 21:21:47.098147 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-5x9f9" Oct 05 21:21:47 crc kubenswrapper[4754]: I1005 21:21:47.104092 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 05 21:21:47 crc kubenswrapper[4754]: I1005 21:21:47.104378 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 05 21:21:47 crc kubenswrapper[4754]: I1005 21:21:47.125177 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 05 21:21:47 crc kubenswrapper[4754]: I1005 21:21:47.125379 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-v54jd" Oct 05 21:21:47 crc kubenswrapper[4754]: I1005 21:21:47.142318 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-5x9f9"] Oct 05 21:21:47 crc kubenswrapper[4754]: I1005 21:21:47.147066 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tmd76\" (UniqueName: \"kubernetes.io/projected/920f98be-147d-4b28-a82c-fa77c86fd8f1-kube-api-access-tmd76\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-5x9f9\" (UID: \"920f98be-147d-4b28-a82c-fa77c86fd8f1\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-5x9f9" Oct 05 21:21:47 crc kubenswrapper[4754]: I1005 21:21:47.147112 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/920f98be-147d-4b28-a82c-fa77c86fd8f1-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-5x9f9\" (UID: \"920f98be-147d-4b28-a82c-fa77c86fd8f1\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-5x9f9" Oct 05 21:21:47 crc kubenswrapper[4754]: I1005 21:21:47.147146 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/920f98be-147d-4b28-a82c-fa77c86fd8f1-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-5x9f9\" (UID: \"920f98be-147d-4b28-a82c-fa77c86fd8f1\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-5x9f9" Oct 05 21:21:47 crc kubenswrapper[4754]: I1005 21:21:47.250101 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tmd76\" (UniqueName: \"kubernetes.io/projected/920f98be-147d-4b28-a82c-fa77c86fd8f1-kube-api-access-tmd76\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-5x9f9\" (UID: \"920f98be-147d-4b28-a82c-fa77c86fd8f1\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-5x9f9" Oct 05 21:21:47 crc kubenswrapper[4754]: I1005 21:21:47.250213 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/920f98be-147d-4b28-a82c-fa77c86fd8f1-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-5x9f9\" (UID: \"920f98be-147d-4b28-a82c-fa77c86fd8f1\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-5x9f9" Oct 05 21:21:47 crc kubenswrapper[4754]: I1005 21:21:47.250255 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/920f98be-147d-4b28-a82c-fa77c86fd8f1-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-5x9f9\" (UID: \"920f98be-147d-4b28-a82c-fa77c86fd8f1\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-5x9f9" Oct 05 21:21:47 crc kubenswrapper[4754]: I1005 21:21:47.260190 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/920f98be-147d-4b28-a82c-fa77c86fd8f1-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-5x9f9\" (UID: \"920f98be-147d-4b28-a82c-fa77c86fd8f1\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-5x9f9" Oct 05 21:21:47 crc kubenswrapper[4754]: I1005 21:21:47.260786 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/920f98be-147d-4b28-a82c-fa77c86fd8f1-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-5x9f9\" (UID: \"920f98be-147d-4b28-a82c-fa77c86fd8f1\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-5x9f9" Oct 05 21:21:47 crc kubenswrapper[4754]: I1005 21:21:47.269006 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tmd76\" (UniqueName: \"kubernetes.io/projected/920f98be-147d-4b28-a82c-fa77c86fd8f1-kube-api-access-tmd76\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-5x9f9\" (UID: \"920f98be-147d-4b28-a82c-fa77c86fd8f1\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-5x9f9" Oct 05 21:21:47 crc kubenswrapper[4754]: I1005 21:21:47.428549 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-5x9f9" Oct 05 21:21:48 crc kubenswrapper[4754]: I1005 21:21:48.021059 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-5x9f9"] Oct 05 21:21:48 crc kubenswrapper[4754]: W1005 21:21:48.032559 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod920f98be_147d_4b28_a82c_fa77c86fd8f1.slice/crio-630d6dcfc056eaeadb60aa44af7cfa076ab2323b76451b0a041d9131f12240e6 WatchSource:0}: Error finding container 630d6dcfc056eaeadb60aa44af7cfa076ab2323b76451b0a041d9131f12240e6: Status 404 returned error can't find the container with id 630d6dcfc056eaeadb60aa44af7cfa076ab2323b76451b0a041d9131f12240e6 Oct 05 21:21:48 crc kubenswrapper[4754]: I1005 21:21:48.855360 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a454365f-a32c-427d-a2d1-cecff2501134" path="/var/lib/kubelet/pods/a454365f-a32c-427d-a2d1-cecff2501134/volumes" Oct 05 21:21:48 crc kubenswrapper[4754]: I1005 21:21:48.955093 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-5x9f9" event={"ID":"920f98be-147d-4b28-a82c-fa77c86fd8f1","Type":"ContainerStarted","Data":"b5020fd322def11f70227cb35b0c7af36eb88704e15108a38abedfbc0be0d2f4"} Oct 05 21:21:48 crc kubenswrapper[4754]: I1005 21:21:48.955415 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-5x9f9" event={"ID":"920f98be-147d-4b28-a82c-fa77c86fd8f1","Type":"ContainerStarted","Data":"630d6dcfc056eaeadb60aa44af7cfa076ab2323b76451b0a041d9131f12240e6"} Oct 05 21:21:48 crc kubenswrapper[4754]: I1005 21:21:48.977206 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-5x9f9" podStartSLOduration=1.4695276050000001 podStartE2EDuration="1.977184236s" podCreationTimestamp="2025-10-05 21:21:47 +0000 UTC" firstStartedPulling="2025-10-05 21:21:48.037004624 +0000 UTC m=+1631.941123344" lastFinishedPulling="2025-10-05 21:21:48.544661225 +0000 UTC m=+1632.448779975" observedRunningTime="2025-10-05 21:21:48.975336909 +0000 UTC m=+1632.879455639" watchObservedRunningTime="2025-10-05 21:21:48.977184236 +0000 UTC m=+1632.881302976" Oct 05 21:21:49 crc kubenswrapper[4754]: I1005 21:21:49.839718 4754 scope.go:117] "RemoveContainer" containerID="e8573d2f52a359de8bfb0cb5d4d1a72d653442930ef95a60a1d6d2dd509ab1ac" Oct 05 21:21:49 crc kubenswrapper[4754]: E1005 21:21:49.840478 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 21:21:53 crc kubenswrapper[4754]: I1005 21:21:53.742003 4754 scope.go:117] "RemoveContainer" containerID="fdd3e88cffd311578243f0016120a4d7c1b3556ae7bc395b06d22414814fa856" Oct 05 21:21:53 crc kubenswrapper[4754]: I1005 21:21:53.812913 4754 scope.go:117] "RemoveContainer" containerID="50d818cfdab3f39a595898fe80cb53f7246ed57225037a25fb76b9787eeb1f6a" Oct 05 21:21:53 crc kubenswrapper[4754]: I1005 21:21:53.833580 4754 scope.go:117] "RemoveContainer" containerID="05effd9954d08807f18e5f05169d35b21126fe1f01e7b273fe9f9107f13be6b4" Oct 05 21:21:53 crc kubenswrapper[4754]: I1005 21:21:53.882304 4754 scope.go:117] "RemoveContainer" containerID="e5cc49e57eb855cc1e05ded8fe904f5da764a35f1685c58e8914f9a18b0f54c0" Oct 05 21:21:53 crc kubenswrapper[4754]: I1005 21:21:53.920165 4754 scope.go:117] "RemoveContainer" containerID="99792af3e76c6ce02963c39a0a6f464ab5bf88393b412a908230ca3a3b9fac0a" Oct 05 21:21:53 crc kubenswrapper[4754]: I1005 21:21:53.968336 4754 scope.go:117] "RemoveContainer" containerID="62d7716283de84282a34e0fa4992d2b086d38ba6c83499f982fd51187d96878a" Oct 05 21:21:54 crc kubenswrapper[4754]: I1005 21:21:54.003961 4754 scope.go:117] "RemoveContainer" containerID="8667484195cb7e3c97d3a0017d2bb486586073eaa74b701c1abddcc59d96b189" Oct 05 21:21:54 crc kubenswrapper[4754]: I1005 21:21:54.045272 4754 scope.go:117] "RemoveContainer" containerID="89f838a779a6be03f73263b78506025ee248a5cc9827aa7981ac52d5d68aab78" Oct 05 21:21:54 crc kubenswrapper[4754]: I1005 21:21:54.069368 4754 scope.go:117] "RemoveContainer" containerID="41a7d5b3018f8f69bfd9504da8a203204ddacb0282e477631de02511aadb7446" Oct 05 21:21:54 crc kubenswrapper[4754]: I1005 21:21:54.091969 4754 scope.go:117] "RemoveContainer" containerID="0e0159d153289e7a8a9958ccd54336126d3fa18d6bb00f21b2ef5061991767b6" Oct 05 21:21:54 crc kubenswrapper[4754]: I1005 21:21:54.116159 4754 scope.go:117] "RemoveContainer" containerID="41b465148330f3be15a27262db77f02a260ba3a73ea227b2e2caa0a0795e677d" Oct 05 21:21:54 crc kubenswrapper[4754]: I1005 21:21:54.139218 4754 scope.go:117] "RemoveContainer" containerID="684677c5bedf56c3583beb7f226abfbf4d2708174fc5174c20b2441d2a43d98f" Oct 05 21:21:54 crc kubenswrapper[4754]: I1005 21:21:54.166486 4754 scope.go:117] "RemoveContainer" containerID="6ec2c24cd529e409522a7d5601c008441f4e50246937630801fbcd282d2978b4" Oct 05 21:22:02 crc kubenswrapper[4754]: I1005 21:22:02.837237 4754 scope.go:117] "RemoveContainer" containerID="e8573d2f52a359de8bfb0cb5d4d1a72d653442930ef95a60a1d6d2dd509ab1ac" Oct 05 21:22:02 crc kubenswrapper[4754]: E1005 21:22:02.838304 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 21:22:07 crc kubenswrapper[4754]: I1005 21:22:07.054805 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-3d83-account-create-hq42j"] Oct 05 21:22:07 crc kubenswrapper[4754]: I1005 21:22:07.074462 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-3d83-account-create-hq42j"] Oct 05 21:22:08 crc kubenswrapper[4754]: I1005 21:22:08.849138 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="be464a13-b373-46f0-bcfa-fcf9f491652e" path="/var/lib/kubelet/pods/be464a13-b373-46f0-bcfa-fcf9f491652e/volumes" Oct 05 21:22:13 crc kubenswrapper[4754]: I1005 21:22:13.035924 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-s4lpl"] Oct 05 21:22:13 crc kubenswrapper[4754]: I1005 21:22:13.052353 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-s4lpl"] Oct 05 21:22:14 crc kubenswrapper[4754]: I1005 21:22:14.864666 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8705e88f-0fb4-4e7c-be21-d7bed4a8bede" path="/var/lib/kubelet/pods/8705e88f-0fb4-4e7c-be21-d7bed4a8bede/volumes" Oct 05 21:22:17 crc kubenswrapper[4754]: I1005 21:22:17.838045 4754 scope.go:117] "RemoveContainer" containerID="e8573d2f52a359de8bfb0cb5d4d1a72d653442930ef95a60a1d6d2dd509ab1ac" Oct 05 21:22:17 crc kubenswrapper[4754]: E1005 21:22:17.839022 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 21:22:29 crc kubenswrapper[4754]: I1005 21:22:29.838451 4754 scope.go:117] "RemoveContainer" containerID="e8573d2f52a359de8bfb0cb5d4d1a72d653442930ef95a60a1d6d2dd509ab1ac" Oct 05 21:22:29 crc kubenswrapper[4754]: E1005 21:22:29.839679 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 21:22:44 crc kubenswrapper[4754]: I1005 21:22:44.838440 4754 scope.go:117] "RemoveContainer" containerID="e8573d2f52a359de8bfb0cb5d4d1a72d653442930ef95a60a1d6d2dd509ab1ac" Oct 05 21:22:44 crc kubenswrapper[4754]: E1005 21:22:44.839626 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 21:22:54 crc kubenswrapper[4754]: I1005 21:22:54.487174 4754 scope.go:117] "RemoveContainer" containerID="4f500128b3304f698cca44b4340c265d04313ef491ceec4ef0596e6aa8e57918" Oct 05 21:22:54 crc kubenswrapper[4754]: I1005 21:22:54.547183 4754 scope.go:117] "RemoveContainer" containerID="6c2f653b095b4d003d9c9987856ca69ed98dbc3f4baf9d8b18f7966747cb44a7" Oct 05 21:22:57 crc kubenswrapper[4754]: I1005 21:22:57.838183 4754 scope.go:117] "RemoveContainer" containerID="e8573d2f52a359de8bfb0cb5d4d1a72d653442930ef95a60a1d6d2dd509ab1ac" Oct 05 21:22:57 crc kubenswrapper[4754]: E1005 21:22:57.839349 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 21:23:11 crc kubenswrapper[4754]: I1005 21:23:11.838076 4754 scope.go:117] "RemoveContainer" containerID="e8573d2f52a359de8bfb0cb5d4d1a72d653442930ef95a60a1d6d2dd509ab1ac" Oct 05 21:23:11 crc kubenswrapper[4754]: E1005 21:23:11.839193 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 21:23:15 crc kubenswrapper[4754]: I1005 21:23:15.099274 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-6wlrw"] Oct 05 21:23:15 crc kubenswrapper[4754]: I1005 21:23:15.107082 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-6wlrw"] Oct 05 21:23:16 crc kubenswrapper[4754]: I1005 21:23:16.855689 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="99975fd1-2c5d-4a2d-9790-0ef13f9fc3f9" path="/var/lib/kubelet/pods/99975fd1-2c5d-4a2d-9790-0ef13f9fc3f9/volumes" Oct 05 21:23:19 crc kubenswrapper[4754]: I1005 21:23:19.031735 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-xncdg"] Oct 05 21:23:19 crc kubenswrapper[4754]: I1005 21:23:19.040184 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-xncdg"] Oct 05 21:23:20 crc kubenswrapper[4754]: I1005 21:23:20.870223 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="94d0ae23-e26d-4bca-b7f3-e4af07ac2722" path="/var/lib/kubelet/pods/94d0ae23-e26d-4bca-b7f3-e4af07ac2722/volumes" Oct 05 21:23:25 crc kubenswrapper[4754]: I1005 21:23:25.837947 4754 scope.go:117] "RemoveContainer" containerID="e8573d2f52a359de8bfb0cb5d4d1a72d653442930ef95a60a1d6d2dd509ab1ac" Oct 05 21:23:25 crc kubenswrapper[4754]: E1005 21:23:25.840645 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 21:23:28 crc kubenswrapper[4754]: I1005 21:23:28.054165 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-xlw68"] Oct 05 21:23:28 crc kubenswrapper[4754]: I1005 21:23:28.063330 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-xlw68"] Oct 05 21:23:28 crc kubenswrapper[4754]: I1005 21:23:28.847947 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="94c73d9b-e224-4788-995d-be11dcee7b3c" path="/var/lib/kubelet/pods/94c73d9b-e224-4788-995d-be11dcee7b3c/volumes" Oct 05 21:23:35 crc kubenswrapper[4754]: I1005 21:23:35.046674 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-drwtj"] Oct 05 21:23:35 crc kubenswrapper[4754]: I1005 21:23:35.067299 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-drwtj"] Oct 05 21:23:38 crc kubenswrapper[4754]: I1005 21:23:38.473137 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5594b283-6565-4ab3-89bd-9769992f6a20" path="/var/lib/kubelet/pods/5594b283-6565-4ab3-89bd-9769992f6a20/volumes" Oct 05 21:23:38 crc kubenswrapper[4754]: E1005 21:23:38.477194 4754 kubelet.go:2526] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="1.641s" Oct 05 21:23:38 crc kubenswrapper[4754]: I1005 21:23:38.478471 4754 scope.go:117] "RemoveContainer" containerID="e8573d2f52a359de8bfb0cb5d4d1a72d653442930ef95a60a1d6d2dd509ab1ac" Oct 05 21:23:38 crc kubenswrapper[4754]: E1005 21:23:38.478893 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 21:23:39 crc kubenswrapper[4754]: I1005 21:23:39.275564 4754 generic.go:334] "Generic (PLEG): container finished" podID="920f98be-147d-4b28-a82c-fa77c86fd8f1" containerID="b5020fd322def11f70227cb35b0c7af36eb88704e15108a38abedfbc0be0d2f4" exitCode=0 Oct 05 21:23:39 crc kubenswrapper[4754]: I1005 21:23:39.275611 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-5x9f9" event={"ID":"920f98be-147d-4b28-a82c-fa77c86fd8f1","Type":"ContainerDied","Data":"b5020fd322def11f70227cb35b0c7af36eb88704e15108a38abedfbc0be0d2f4"} Oct 05 21:23:40 crc kubenswrapper[4754]: I1005 21:23:40.746217 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-5x9f9" Oct 05 21:23:40 crc kubenswrapper[4754]: I1005 21:23:40.849812 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/920f98be-147d-4b28-a82c-fa77c86fd8f1-inventory\") pod \"920f98be-147d-4b28-a82c-fa77c86fd8f1\" (UID: \"920f98be-147d-4b28-a82c-fa77c86fd8f1\") " Oct 05 21:23:40 crc kubenswrapper[4754]: I1005 21:23:40.850097 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tmd76\" (UniqueName: \"kubernetes.io/projected/920f98be-147d-4b28-a82c-fa77c86fd8f1-kube-api-access-tmd76\") pod \"920f98be-147d-4b28-a82c-fa77c86fd8f1\" (UID: \"920f98be-147d-4b28-a82c-fa77c86fd8f1\") " Oct 05 21:23:40 crc kubenswrapper[4754]: I1005 21:23:40.850252 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/920f98be-147d-4b28-a82c-fa77c86fd8f1-ssh-key\") pod \"920f98be-147d-4b28-a82c-fa77c86fd8f1\" (UID: \"920f98be-147d-4b28-a82c-fa77c86fd8f1\") " Oct 05 21:23:40 crc kubenswrapper[4754]: I1005 21:23:40.864759 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/920f98be-147d-4b28-a82c-fa77c86fd8f1-kube-api-access-tmd76" (OuterVolumeSpecName: "kube-api-access-tmd76") pod "920f98be-147d-4b28-a82c-fa77c86fd8f1" (UID: "920f98be-147d-4b28-a82c-fa77c86fd8f1"). InnerVolumeSpecName "kube-api-access-tmd76". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:23:40 crc kubenswrapper[4754]: I1005 21:23:40.883195 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/920f98be-147d-4b28-a82c-fa77c86fd8f1-inventory" (OuterVolumeSpecName: "inventory") pod "920f98be-147d-4b28-a82c-fa77c86fd8f1" (UID: "920f98be-147d-4b28-a82c-fa77c86fd8f1"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:23:40 crc kubenswrapper[4754]: I1005 21:23:40.894977 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/920f98be-147d-4b28-a82c-fa77c86fd8f1-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "920f98be-147d-4b28-a82c-fa77c86fd8f1" (UID: "920f98be-147d-4b28-a82c-fa77c86fd8f1"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:23:40 crc kubenswrapper[4754]: I1005 21:23:40.954134 4754 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/920f98be-147d-4b28-a82c-fa77c86fd8f1-inventory\") on node \"crc\" DevicePath \"\"" Oct 05 21:23:40 crc kubenswrapper[4754]: I1005 21:23:40.954184 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tmd76\" (UniqueName: \"kubernetes.io/projected/920f98be-147d-4b28-a82c-fa77c86fd8f1-kube-api-access-tmd76\") on node \"crc\" DevicePath \"\"" Oct 05 21:23:40 crc kubenswrapper[4754]: I1005 21:23:40.954199 4754 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/920f98be-147d-4b28-a82c-fa77c86fd8f1-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 05 21:23:41 crc kubenswrapper[4754]: I1005 21:23:41.295980 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-5x9f9" event={"ID":"920f98be-147d-4b28-a82c-fa77c86fd8f1","Type":"ContainerDied","Data":"630d6dcfc056eaeadb60aa44af7cfa076ab2323b76451b0a041d9131f12240e6"} Oct 05 21:23:41 crc kubenswrapper[4754]: I1005 21:23:41.296020 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="630d6dcfc056eaeadb60aa44af7cfa076ab2323b76451b0a041d9131f12240e6" Oct 05 21:23:41 crc kubenswrapper[4754]: I1005 21:23:41.296030 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-5x9f9" Oct 05 21:23:41 crc kubenswrapper[4754]: I1005 21:23:41.396430 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-7xfdc"] Oct 05 21:23:41 crc kubenswrapper[4754]: E1005 21:23:41.401215 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="920f98be-147d-4b28-a82c-fa77c86fd8f1" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Oct 05 21:23:41 crc kubenswrapper[4754]: I1005 21:23:41.401336 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="920f98be-147d-4b28-a82c-fa77c86fd8f1" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Oct 05 21:23:41 crc kubenswrapper[4754]: I1005 21:23:41.401622 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="920f98be-147d-4b28-a82c-fa77c86fd8f1" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Oct 05 21:23:41 crc kubenswrapper[4754]: I1005 21:23:41.402354 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-7xfdc" Oct 05 21:23:41 crc kubenswrapper[4754]: I1005 21:23:41.406361 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 05 21:23:41 crc kubenswrapper[4754]: I1005 21:23:41.407093 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 05 21:23:41 crc kubenswrapper[4754]: I1005 21:23:41.407737 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 05 21:23:41 crc kubenswrapper[4754]: I1005 21:23:41.407961 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-v54jd" Oct 05 21:23:41 crc kubenswrapper[4754]: I1005 21:23:41.413535 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-7xfdc"] Oct 05 21:23:41 crc kubenswrapper[4754]: I1005 21:23:41.463550 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-khz87\" (UniqueName: \"kubernetes.io/projected/5421db8c-f81e-483e-ac15-c0b725a2c277-kube-api-access-khz87\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-7xfdc\" (UID: \"5421db8c-f81e-483e-ac15-c0b725a2c277\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-7xfdc" Oct 05 21:23:41 crc kubenswrapper[4754]: I1005 21:23:41.463612 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5421db8c-f81e-483e-ac15-c0b725a2c277-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-7xfdc\" (UID: \"5421db8c-f81e-483e-ac15-c0b725a2c277\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-7xfdc" Oct 05 21:23:41 crc kubenswrapper[4754]: I1005 21:23:41.463677 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5421db8c-f81e-483e-ac15-c0b725a2c277-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-7xfdc\" (UID: \"5421db8c-f81e-483e-ac15-c0b725a2c277\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-7xfdc" Oct 05 21:23:41 crc kubenswrapper[4754]: I1005 21:23:41.565734 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-khz87\" (UniqueName: \"kubernetes.io/projected/5421db8c-f81e-483e-ac15-c0b725a2c277-kube-api-access-khz87\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-7xfdc\" (UID: \"5421db8c-f81e-483e-ac15-c0b725a2c277\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-7xfdc" Oct 05 21:23:41 crc kubenswrapper[4754]: I1005 21:23:41.565796 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5421db8c-f81e-483e-ac15-c0b725a2c277-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-7xfdc\" (UID: \"5421db8c-f81e-483e-ac15-c0b725a2c277\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-7xfdc" Oct 05 21:23:41 crc kubenswrapper[4754]: I1005 21:23:41.565868 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5421db8c-f81e-483e-ac15-c0b725a2c277-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-7xfdc\" (UID: \"5421db8c-f81e-483e-ac15-c0b725a2c277\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-7xfdc" Oct 05 21:23:41 crc kubenswrapper[4754]: I1005 21:23:41.569573 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5421db8c-f81e-483e-ac15-c0b725a2c277-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-7xfdc\" (UID: \"5421db8c-f81e-483e-ac15-c0b725a2c277\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-7xfdc" Oct 05 21:23:41 crc kubenswrapper[4754]: I1005 21:23:41.569964 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5421db8c-f81e-483e-ac15-c0b725a2c277-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-7xfdc\" (UID: \"5421db8c-f81e-483e-ac15-c0b725a2c277\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-7xfdc" Oct 05 21:23:41 crc kubenswrapper[4754]: I1005 21:23:41.583076 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-khz87\" (UniqueName: \"kubernetes.io/projected/5421db8c-f81e-483e-ac15-c0b725a2c277-kube-api-access-khz87\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-7xfdc\" (UID: \"5421db8c-f81e-483e-ac15-c0b725a2c277\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-7xfdc" Oct 05 21:23:41 crc kubenswrapper[4754]: I1005 21:23:41.728756 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-7xfdc" Oct 05 21:23:42 crc kubenswrapper[4754]: I1005 21:23:42.242707 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-7xfdc"] Oct 05 21:23:42 crc kubenswrapper[4754]: I1005 21:23:42.305981 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-7xfdc" event={"ID":"5421db8c-f81e-483e-ac15-c0b725a2c277","Type":"ContainerStarted","Data":"f1e3f8fcd0cac274e04012db54d89f4d0669aaf142e413341b97cb9d94d8788a"} Oct 05 21:23:43 crc kubenswrapper[4754]: I1005 21:23:43.322203 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-7xfdc" event={"ID":"5421db8c-f81e-483e-ac15-c0b725a2c277","Type":"ContainerStarted","Data":"027af8341448492fa4329e6efd80705e5dfceb542de40d2b6147458f05e29f61"} Oct 05 21:23:43 crc kubenswrapper[4754]: I1005 21:23:43.344098 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-7xfdc" podStartSLOduration=1.746825353 podStartE2EDuration="2.344079328s" podCreationTimestamp="2025-10-05 21:23:41 +0000 UTC" firstStartedPulling="2025-10-05 21:23:42.24981462 +0000 UTC m=+1746.153933350" lastFinishedPulling="2025-10-05 21:23:42.847068615 +0000 UTC m=+1746.751187325" observedRunningTime="2025-10-05 21:23:43.343898893 +0000 UTC m=+1747.248017603" watchObservedRunningTime="2025-10-05 21:23:43.344079328 +0000 UTC m=+1747.248198048" Oct 05 21:23:49 crc kubenswrapper[4754]: I1005 21:23:49.837682 4754 scope.go:117] "RemoveContainer" containerID="e8573d2f52a359de8bfb0cb5d4d1a72d653442930ef95a60a1d6d2dd509ab1ac" Oct 05 21:23:49 crc kubenswrapper[4754]: E1005 21:23:49.839832 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 21:23:54 crc kubenswrapper[4754]: I1005 21:23:54.716391 4754 scope.go:117] "RemoveContainer" containerID="4302b6bd8251e7a7c27e4b9908256f84d6cc923b3112ce45c885fc471fb96198" Oct 05 21:23:54 crc kubenswrapper[4754]: I1005 21:23:54.766606 4754 scope.go:117] "RemoveContainer" containerID="35e7ec9f9fdd5504e8cead852c32e929a754fb8f8f40cf4c89ee0c93643b374b" Oct 05 21:23:54 crc kubenswrapper[4754]: I1005 21:23:54.801567 4754 scope.go:117] "RemoveContainer" containerID="907c4f77f15040be8d5f56e9c1085880e54be3a78590fa12671e33449f1ded81" Oct 05 21:23:54 crc kubenswrapper[4754]: I1005 21:23:54.907219 4754 scope.go:117] "RemoveContainer" containerID="beb03566dca75372fd1c3a5f05aae0ee0555748b120f181ee2ca7eb6e2d97a71" Oct 05 21:24:04 crc kubenswrapper[4754]: I1005 21:24:04.838980 4754 scope.go:117] "RemoveContainer" containerID="e8573d2f52a359de8bfb0cb5d4d1a72d653442930ef95a60a1d6d2dd509ab1ac" Oct 05 21:24:04 crc kubenswrapper[4754]: E1005 21:24:04.840670 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 21:24:09 crc kubenswrapper[4754]: I1005 21:24:09.050625 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-x2r84"] Oct 05 21:24:09 crc kubenswrapper[4754]: I1005 21:24:09.066491 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-x2r84"] Oct 05 21:24:10 crc kubenswrapper[4754]: I1005 21:24:10.069311 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-tfbkl"] Oct 05 21:24:10 crc kubenswrapper[4754]: I1005 21:24:10.076736 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-cwmnq"] Oct 05 21:24:10 crc kubenswrapper[4754]: I1005 21:24:10.086634 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-cwmnq"] Oct 05 21:24:10 crc kubenswrapper[4754]: I1005 21:24:10.093986 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-tfbkl"] Oct 05 21:24:10 crc kubenswrapper[4754]: I1005 21:24:10.855757 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="24624447-d035-4476-b248-b7776bb261cd" path="/var/lib/kubelet/pods/24624447-d035-4476-b248-b7776bb261cd/volumes" Oct 05 21:24:10 crc kubenswrapper[4754]: I1005 21:24:10.857998 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="339a3c5a-8feb-4381-ae02-dcff29fc61fc" path="/var/lib/kubelet/pods/339a3c5a-8feb-4381-ae02-dcff29fc61fc/volumes" Oct 05 21:24:10 crc kubenswrapper[4754]: I1005 21:24:10.859042 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5e45337f-b84d-473e-8530-942a17eb3c15" path="/var/lib/kubelet/pods/5e45337f-b84d-473e-8530-942a17eb3c15/volumes" Oct 05 21:24:17 crc kubenswrapper[4754]: I1005 21:24:17.839276 4754 scope.go:117] "RemoveContainer" containerID="e8573d2f52a359de8bfb0cb5d4d1a72d653442930ef95a60a1d6d2dd509ab1ac" Oct 05 21:24:17 crc kubenswrapper[4754]: E1005 21:24:17.840817 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 21:24:21 crc kubenswrapper[4754]: I1005 21:24:21.050243 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-89d4-account-create-vxv7h"] Oct 05 21:24:21 crc kubenswrapper[4754]: I1005 21:24:21.059287 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-89d4-account-create-vxv7h"] Oct 05 21:24:22 crc kubenswrapper[4754]: I1005 21:24:22.877162 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6f2f8cf7-43a7-4845-b9a5-d39680402e09" path="/var/lib/kubelet/pods/6f2f8cf7-43a7-4845-b9a5-d39680402e09/volumes" Oct 05 21:24:30 crc kubenswrapper[4754]: I1005 21:24:30.065213 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-e205-account-create-hh2rc"] Oct 05 21:24:30 crc kubenswrapper[4754]: I1005 21:24:30.082529 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-1d7a-account-create-dtspg"] Oct 05 21:24:30 crc kubenswrapper[4754]: I1005 21:24:30.091246 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-1d7a-account-create-dtspg"] Oct 05 21:24:30 crc kubenswrapper[4754]: I1005 21:24:30.101387 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-e205-account-create-hh2rc"] Oct 05 21:24:30 crc kubenswrapper[4754]: I1005 21:24:30.852456 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="48089a5e-fc86-40cd-96db-c522040598c4" path="/var/lib/kubelet/pods/48089a5e-fc86-40cd-96db-c522040598c4/volumes" Oct 05 21:24:30 crc kubenswrapper[4754]: I1005 21:24:30.854154 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fa9e9f3c-9ee0-426d-aafe-0c39c9980b53" path="/var/lib/kubelet/pods/fa9e9f3c-9ee0-426d-aafe-0c39c9980b53/volumes" Oct 05 21:24:31 crc kubenswrapper[4754]: I1005 21:24:31.837853 4754 scope.go:117] "RemoveContainer" containerID="e8573d2f52a359de8bfb0cb5d4d1a72d653442930ef95a60a1d6d2dd509ab1ac" Oct 05 21:24:31 crc kubenswrapper[4754]: E1005 21:24:31.838276 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 21:24:43 crc kubenswrapper[4754]: I1005 21:24:43.838245 4754 scope.go:117] "RemoveContainer" containerID="e8573d2f52a359de8bfb0cb5d4d1a72d653442930ef95a60a1d6d2dd509ab1ac" Oct 05 21:24:43 crc kubenswrapper[4754]: E1005 21:24:43.839250 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 21:24:54 crc kubenswrapper[4754]: I1005 21:24:54.030044 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-4p4pw"] Oct 05 21:24:54 crc kubenswrapper[4754]: I1005 21:24:54.050755 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-4p4pw"] Oct 05 21:24:54 crc kubenswrapper[4754]: I1005 21:24:54.849146 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="498207e0-f869-46d9-bdc9-d83f44d02385" path="/var/lib/kubelet/pods/498207e0-f869-46d9-bdc9-d83f44d02385/volumes" Oct 05 21:24:55 crc kubenswrapper[4754]: I1005 21:24:55.025416 4754 scope.go:117] "RemoveContainer" containerID="30976ef69a1385193880865c854658ecdb6d9fbdbbcf4197ad7bb575690f26d9" Oct 05 21:24:55 crc kubenswrapper[4754]: I1005 21:24:55.064912 4754 scope.go:117] "RemoveContainer" containerID="5c0c3769cc7e996e252f750662bfc170350b169e66fd2eb2db56ef545fd81f57" Oct 05 21:24:55 crc kubenswrapper[4754]: I1005 21:24:55.128139 4754 scope.go:117] "RemoveContainer" containerID="5d5ac350e26ebbf73c7b124d89e020d93d0c4cc3ee33bd5619260334c213e7d1" Oct 05 21:24:55 crc kubenswrapper[4754]: I1005 21:24:55.183035 4754 scope.go:117] "RemoveContainer" containerID="cc96f5fe37567878834101eddfd72e749514fc5b7abf47dabfa6c8684f47ac67" Oct 05 21:24:55 crc kubenswrapper[4754]: I1005 21:24:55.215665 4754 scope.go:117] "RemoveContainer" containerID="14be147d93ae83ceeb0bc2bb4a7a02b0e0b47263ec54292f2da834c0a8aeea03" Oct 05 21:24:55 crc kubenswrapper[4754]: I1005 21:24:55.254326 4754 scope.go:117] "RemoveContainer" containerID="d4d26a2314f7524c5a56cc71b40492123a49edd7c648d3266e69d2b6bffc2a7f" Oct 05 21:24:55 crc kubenswrapper[4754]: I1005 21:24:55.299589 4754 scope.go:117] "RemoveContainer" containerID="2e5929c9b37c6691cf5350c3d6078cf7fb5f65a0adafa8b9331f21166686756d" Oct 05 21:24:58 crc kubenswrapper[4754]: I1005 21:24:58.837438 4754 scope.go:117] "RemoveContainer" containerID="e8573d2f52a359de8bfb0cb5d4d1a72d653442930ef95a60a1d6d2dd509ab1ac" Oct 05 21:24:58 crc kubenswrapper[4754]: E1005 21:24:58.838249 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 21:25:05 crc kubenswrapper[4754]: I1005 21:25:05.215652 4754 generic.go:334] "Generic (PLEG): container finished" podID="5421db8c-f81e-483e-ac15-c0b725a2c277" containerID="027af8341448492fa4329e6efd80705e5dfceb542de40d2b6147458f05e29f61" exitCode=0 Oct 05 21:25:05 crc kubenswrapper[4754]: I1005 21:25:05.216233 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-7xfdc" event={"ID":"5421db8c-f81e-483e-ac15-c0b725a2c277","Type":"ContainerDied","Data":"027af8341448492fa4329e6efd80705e5dfceb542de40d2b6147458f05e29f61"} Oct 05 21:25:06 crc kubenswrapper[4754]: I1005 21:25:06.655399 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-7xfdc" Oct 05 21:25:06 crc kubenswrapper[4754]: I1005 21:25:06.841290 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5421db8c-f81e-483e-ac15-c0b725a2c277-ssh-key\") pod \"5421db8c-f81e-483e-ac15-c0b725a2c277\" (UID: \"5421db8c-f81e-483e-ac15-c0b725a2c277\") " Oct 05 21:25:06 crc kubenswrapper[4754]: I1005 21:25:06.841825 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-khz87\" (UniqueName: \"kubernetes.io/projected/5421db8c-f81e-483e-ac15-c0b725a2c277-kube-api-access-khz87\") pod \"5421db8c-f81e-483e-ac15-c0b725a2c277\" (UID: \"5421db8c-f81e-483e-ac15-c0b725a2c277\") " Oct 05 21:25:06 crc kubenswrapper[4754]: I1005 21:25:06.842339 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5421db8c-f81e-483e-ac15-c0b725a2c277-inventory\") pod \"5421db8c-f81e-483e-ac15-c0b725a2c277\" (UID: \"5421db8c-f81e-483e-ac15-c0b725a2c277\") " Oct 05 21:25:06 crc kubenswrapper[4754]: I1005 21:25:06.850776 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5421db8c-f81e-483e-ac15-c0b725a2c277-kube-api-access-khz87" (OuterVolumeSpecName: "kube-api-access-khz87") pod "5421db8c-f81e-483e-ac15-c0b725a2c277" (UID: "5421db8c-f81e-483e-ac15-c0b725a2c277"). InnerVolumeSpecName "kube-api-access-khz87". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:25:06 crc kubenswrapper[4754]: I1005 21:25:06.881996 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5421db8c-f81e-483e-ac15-c0b725a2c277-inventory" (OuterVolumeSpecName: "inventory") pod "5421db8c-f81e-483e-ac15-c0b725a2c277" (UID: "5421db8c-f81e-483e-ac15-c0b725a2c277"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:25:06 crc kubenswrapper[4754]: I1005 21:25:06.882325 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5421db8c-f81e-483e-ac15-c0b725a2c277-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "5421db8c-f81e-483e-ac15-c0b725a2c277" (UID: "5421db8c-f81e-483e-ac15-c0b725a2c277"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:25:06 crc kubenswrapper[4754]: I1005 21:25:06.944983 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-khz87\" (UniqueName: \"kubernetes.io/projected/5421db8c-f81e-483e-ac15-c0b725a2c277-kube-api-access-khz87\") on node \"crc\" DevicePath \"\"" Oct 05 21:25:06 crc kubenswrapper[4754]: I1005 21:25:06.945032 4754 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5421db8c-f81e-483e-ac15-c0b725a2c277-inventory\") on node \"crc\" DevicePath \"\"" Oct 05 21:25:06 crc kubenswrapper[4754]: I1005 21:25:06.945052 4754 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5421db8c-f81e-483e-ac15-c0b725a2c277-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 05 21:25:07 crc kubenswrapper[4754]: I1005 21:25:07.241075 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-7xfdc" event={"ID":"5421db8c-f81e-483e-ac15-c0b725a2c277","Type":"ContainerDied","Data":"f1e3f8fcd0cac274e04012db54d89f4d0669aaf142e413341b97cb9d94d8788a"} Oct 05 21:25:07 crc kubenswrapper[4754]: I1005 21:25:07.241120 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f1e3f8fcd0cac274e04012db54d89f4d0669aaf142e413341b97cb9d94d8788a" Oct 05 21:25:07 crc kubenswrapper[4754]: I1005 21:25:07.241120 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-7xfdc" Oct 05 21:25:07 crc kubenswrapper[4754]: I1005 21:25:07.352467 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-thlqz"] Oct 05 21:25:07 crc kubenswrapper[4754]: E1005 21:25:07.352968 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5421db8c-f81e-483e-ac15-c0b725a2c277" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Oct 05 21:25:07 crc kubenswrapper[4754]: I1005 21:25:07.352987 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="5421db8c-f81e-483e-ac15-c0b725a2c277" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Oct 05 21:25:07 crc kubenswrapper[4754]: I1005 21:25:07.353178 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="5421db8c-f81e-483e-ac15-c0b725a2c277" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Oct 05 21:25:07 crc kubenswrapper[4754]: I1005 21:25:07.353836 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-thlqz" Oct 05 21:25:07 crc kubenswrapper[4754]: I1005 21:25:07.370987 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-v54jd" Oct 05 21:25:07 crc kubenswrapper[4754]: I1005 21:25:07.371228 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 05 21:25:07 crc kubenswrapper[4754]: I1005 21:25:07.371396 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 05 21:25:07 crc kubenswrapper[4754]: I1005 21:25:07.379379 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 05 21:25:07 crc kubenswrapper[4754]: I1005 21:25:07.392970 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-thlqz"] Oct 05 21:25:07 crc kubenswrapper[4754]: I1005 21:25:07.455233 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-669ch\" (UniqueName: \"kubernetes.io/projected/773b7e70-b0ed-4b02-98fe-eeaf81925900-kube-api-access-669ch\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-thlqz\" (UID: \"773b7e70-b0ed-4b02-98fe-eeaf81925900\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-thlqz" Oct 05 21:25:07 crc kubenswrapper[4754]: I1005 21:25:07.455335 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/773b7e70-b0ed-4b02-98fe-eeaf81925900-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-thlqz\" (UID: \"773b7e70-b0ed-4b02-98fe-eeaf81925900\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-thlqz" Oct 05 21:25:07 crc kubenswrapper[4754]: I1005 21:25:07.455402 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/773b7e70-b0ed-4b02-98fe-eeaf81925900-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-thlqz\" (UID: \"773b7e70-b0ed-4b02-98fe-eeaf81925900\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-thlqz" Oct 05 21:25:07 crc kubenswrapper[4754]: I1005 21:25:07.557221 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-669ch\" (UniqueName: \"kubernetes.io/projected/773b7e70-b0ed-4b02-98fe-eeaf81925900-kube-api-access-669ch\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-thlqz\" (UID: \"773b7e70-b0ed-4b02-98fe-eeaf81925900\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-thlqz" Oct 05 21:25:07 crc kubenswrapper[4754]: I1005 21:25:07.557342 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/773b7e70-b0ed-4b02-98fe-eeaf81925900-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-thlqz\" (UID: \"773b7e70-b0ed-4b02-98fe-eeaf81925900\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-thlqz" Oct 05 21:25:07 crc kubenswrapper[4754]: I1005 21:25:07.557429 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/773b7e70-b0ed-4b02-98fe-eeaf81925900-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-thlqz\" (UID: \"773b7e70-b0ed-4b02-98fe-eeaf81925900\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-thlqz" Oct 05 21:25:07 crc kubenswrapper[4754]: I1005 21:25:07.563102 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/773b7e70-b0ed-4b02-98fe-eeaf81925900-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-thlqz\" (UID: \"773b7e70-b0ed-4b02-98fe-eeaf81925900\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-thlqz" Oct 05 21:25:07 crc kubenswrapper[4754]: I1005 21:25:07.592102 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/773b7e70-b0ed-4b02-98fe-eeaf81925900-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-thlqz\" (UID: \"773b7e70-b0ed-4b02-98fe-eeaf81925900\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-thlqz" Oct 05 21:25:07 crc kubenswrapper[4754]: I1005 21:25:07.608065 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-669ch\" (UniqueName: \"kubernetes.io/projected/773b7e70-b0ed-4b02-98fe-eeaf81925900-kube-api-access-669ch\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-thlqz\" (UID: \"773b7e70-b0ed-4b02-98fe-eeaf81925900\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-thlqz" Oct 05 21:25:07 crc kubenswrapper[4754]: I1005 21:25:07.672911 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-thlqz" Oct 05 21:25:08 crc kubenswrapper[4754]: I1005 21:25:08.061572 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-thlqz"] Oct 05 21:25:08 crc kubenswrapper[4754]: I1005 21:25:08.255803 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-thlqz" event={"ID":"773b7e70-b0ed-4b02-98fe-eeaf81925900","Type":"ContainerStarted","Data":"345b00ecd1c1eb62de5e04b274edfc945e69740b1b4f192add1d96ad18cda759"} Oct 05 21:25:10 crc kubenswrapper[4754]: I1005 21:25:10.279030 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-thlqz" event={"ID":"773b7e70-b0ed-4b02-98fe-eeaf81925900","Type":"ContainerStarted","Data":"14985fa280fbf9bd6c29be430d6aafa2e045e884eb2c88ea6a2522429c9a5893"} Oct 05 21:25:10 crc kubenswrapper[4754]: I1005 21:25:10.310991 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-thlqz" podStartSLOduration=2.555873597 podStartE2EDuration="3.310969766s" podCreationTimestamp="2025-10-05 21:25:07 +0000 UTC" firstStartedPulling="2025-10-05 21:25:08.079476983 +0000 UTC m=+1831.983595693" lastFinishedPulling="2025-10-05 21:25:08.834573152 +0000 UTC m=+1832.738691862" observedRunningTime="2025-10-05 21:25:10.300537374 +0000 UTC m=+1834.204656094" watchObservedRunningTime="2025-10-05 21:25:10.310969766 +0000 UTC m=+1834.215088486" Oct 05 21:25:10 crc kubenswrapper[4754]: I1005 21:25:10.837633 4754 scope.go:117] "RemoveContainer" containerID="e8573d2f52a359de8bfb0cb5d4d1a72d653442930ef95a60a1d6d2dd509ab1ac" Oct 05 21:25:10 crc kubenswrapper[4754]: E1005 21:25:10.838092 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 21:25:15 crc kubenswrapper[4754]: I1005 21:25:15.342851 4754 generic.go:334] "Generic (PLEG): container finished" podID="773b7e70-b0ed-4b02-98fe-eeaf81925900" containerID="14985fa280fbf9bd6c29be430d6aafa2e045e884eb2c88ea6a2522429c9a5893" exitCode=0 Oct 05 21:25:15 crc kubenswrapper[4754]: I1005 21:25:15.342994 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-thlqz" event={"ID":"773b7e70-b0ed-4b02-98fe-eeaf81925900","Type":"ContainerDied","Data":"14985fa280fbf9bd6c29be430d6aafa2e045e884eb2c88ea6a2522429c9a5893"} Oct 05 21:25:16 crc kubenswrapper[4754]: I1005 21:25:16.884175 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-thlqz" Oct 05 21:25:16 crc kubenswrapper[4754]: I1005 21:25:16.984392 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/773b7e70-b0ed-4b02-98fe-eeaf81925900-inventory\") pod \"773b7e70-b0ed-4b02-98fe-eeaf81925900\" (UID: \"773b7e70-b0ed-4b02-98fe-eeaf81925900\") " Oct 05 21:25:16 crc kubenswrapper[4754]: I1005 21:25:16.984892 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/773b7e70-b0ed-4b02-98fe-eeaf81925900-ssh-key\") pod \"773b7e70-b0ed-4b02-98fe-eeaf81925900\" (UID: \"773b7e70-b0ed-4b02-98fe-eeaf81925900\") " Oct 05 21:25:16 crc kubenswrapper[4754]: I1005 21:25:16.985555 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-669ch\" (UniqueName: \"kubernetes.io/projected/773b7e70-b0ed-4b02-98fe-eeaf81925900-kube-api-access-669ch\") pod \"773b7e70-b0ed-4b02-98fe-eeaf81925900\" (UID: \"773b7e70-b0ed-4b02-98fe-eeaf81925900\") " Oct 05 21:25:16 crc kubenswrapper[4754]: I1005 21:25:16.991824 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/773b7e70-b0ed-4b02-98fe-eeaf81925900-kube-api-access-669ch" (OuterVolumeSpecName: "kube-api-access-669ch") pod "773b7e70-b0ed-4b02-98fe-eeaf81925900" (UID: "773b7e70-b0ed-4b02-98fe-eeaf81925900"). InnerVolumeSpecName "kube-api-access-669ch". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:25:17 crc kubenswrapper[4754]: I1005 21:25:17.037686 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/773b7e70-b0ed-4b02-98fe-eeaf81925900-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "773b7e70-b0ed-4b02-98fe-eeaf81925900" (UID: "773b7e70-b0ed-4b02-98fe-eeaf81925900"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:25:17 crc kubenswrapper[4754]: I1005 21:25:17.045628 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/773b7e70-b0ed-4b02-98fe-eeaf81925900-inventory" (OuterVolumeSpecName: "inventory") pod "773b7e70-b0ed-4b02-98fe-eeaf81925900" (UID: "773b7e70-b0ed-4b02-98fe-eeaf81925900"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:25:17 crc kubenswrapper[4754]: I1005 21:25:17.093719 4754 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/773b7e70-b0ed-4b02-98fe-eeaf81925900-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 05 21:25:17 crc kubenswrapper[4754]: I1005 21:25:17.093779 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-669ch\" (UniqueName: \"kubernetes.io/projected/773b7e70-b0ed-4b02-98fe-eeaf81925900-kube-api-access-669ch\") on node \"crc\" DevicePath \"\"" Oct 05 21:25:17 crc kubenswrapper[4754]: I1005 21:25:17.093805 4754 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/773b7e70-b0ed-4b02-98fe-eeaf81925900-inventory\") on node \"crc\" DevicePath \"\"" Oct 05 21:25:17 crc kubenswrapper[4754]: I1005 21:25:17.368087 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-thlqz" event={"ID":"773b7e70-b0ed-4b02-98fe-eeaf81925900","Type":"ContainerDied","Data":"345b00ecd1c1eb62de5e04b274edfc945e69740b1b4f192add1d96ad18cda759"} Oct 05 21:25:17 crc kubenswrapper[4754]: I1005 21:25:17.368379 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="345b00ecd1c1eb62de5e04b274edfc945e69740b1b4f192add1d96ad18cda759" Oct 05 21:25:17 crc kubenswrapper[4754]: I1005 21:25:17.368203 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-thlqz" Oct 05 21:25:17 crc kubenswrapper[4754]: I1005 21:25:17.523348 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-csfdg"] Oct 05 21:25:17 crc kubenswrapper[4754]: E1005 21:25:17.523787 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="773b7e70-b0ed-4b02-98fe-eeaf81925900" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Oct 05 21:25:17 crc kubenswrapper[4754]: I1005 21:25:17.523807 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="773b7e70-b0ed-4b02-98fe-eeaf81925900" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Oct 05 21:25:17 crc kubenswrapper[4754]: I1005 21:25:17.523983 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="773b7e70-b0ed-4b02-98fe-eeaf81925900" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Oct 05 21:25:17 crc kubenswrapper[4754]: I1005 21:25:17.524682 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-csfdg" Oct 05 21:25:17 crc kubenswrapper[4754]: I1005 21:25:17.527602 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-v54jd" Oct 05 21:25:17 crc kubenswrapper[4754]: I1005 21:25:17.527873 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 05 21:25:17 crc kubenswrapper[4754]: I1005 21:25:17.528027 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 05 21:25:17 crc kubenswrapper[4754]: I1005 21:25:17.528080 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 05 21:25:17 crc kubenswrapper[4754]: I1005 21:25:17.584767 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-csfdg"] Oct 05 21:25:17 crc kubenswrapper[4754]: I1005 21:25:17.609708 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/35d09fe3-0746-4e6c-82b2-cc5bef030483-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-csfdg\" (UID: \"35d09fe3-0746-4e6c-82b2-cc5bef030483\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-csfdg" Oct 05 21:25:17 crc kubenswrapper[4754]: I1005 21:25:17.609789 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hj88s\" (UniqueName: \"kubernetes.io/projected/35d09fe3-0746-4e6c-82b2-cc5bef030483-kube-api-access-hj88s\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-csfdg\" (UID: \"35d09fe3-0746-4e6c-82b2-cc5bef030483\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-csfdg" Oct 05 21:25:17 crc kubenswrapper[4754]: I1005 21:25:17.609861 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/35d09fe3-0746-4e6c-82b2-cc5bef030483-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-csfdg\" (UID: \"35d09fe3-0746-4e6c-82b2-cc5bef030483\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-csfdg" Oct 05 21:25:17 crc kubenswrapper[4754]: I1005 21:25:17.711329 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/35d09fe3-0746-4e6c-82b2-cc5bef030483-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-csfdg\" (UID: \"35d09fe3-0746-4e6c-82b2-cc5bef030483\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-csfdg" Oct 05 21:25:17 crc kubenswrapper[4754]: I1005 21:25:17.711406 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hj88s\" (UniqueName: \"kubernetes.io/projected/35d09fe3-0746-4e6c-82b2-cc5bef030483-kube-api-access-hj88s\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-csfdg\" (UID: \"35d09fe3-0746-4e6c-82b2-cc5bef030483\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-csfdg" Oct 05 21:25:17 crc kubenswrapper[4754]: I1005 21:25:17.711482 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/35d09fe3-0746-4e6c-82b2-cc5bef030483-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-csfdg\" (UID: \"35d09fe3-0746-4e6c-82b2-cc5bef030483\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-csfdg" Oct 05 21:25:17 crc kubenswrapper[4754]: I1005 21:25:17.716187 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/35d09fe3-0746-4e6c-82b2-cc5bef030483-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-csfdg\" (UID: \"35d09fe3-0746-4e6c-82b2-cc5bef030483\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-csfdg" Oct 05 21:25:17 crc kubenswrapper[4754]: I1005 21:25:17.716520 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/35d09fe3-0746-4e6c-82b2-cc5bef030483-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-csfdg\" (UID: \"35d09fe3-0746-4e6c-82b2-cc5bef030483\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-csfdg" Oct 05 21:25:17 crc kubenswrapper[4754]: I1005 21:25:17.728302 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hj88s\" (UniqueName: \"kubernetes.io/projected/35d09fe3-0746-4e6c-82b2-cc5bef030483-kube-api-access-hj88s\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-csfdg\" (UID: \"35d09fe3-0746-4e6c-82b2-cc5bef030483\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-csfdg" Oct 05 21:25:17 crc kubenswrapper[4754]: I1005 21:25:17.874950 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-csfdg" Oct 05 21:25:18 crc kubenswrapper[4754]: I1005 21:25:18.502288 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-csfdg"] Oct 05 21:25:18 crc kubenswrapper[4754]: I1005 21:25:18.507097 4754 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 05 21:25:19 crc kubenswrapper[4754]: I1005 21:25:19.388428 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-csfdg" event={"ID":"35d09fe3-0746-4e6c-82b2-cc5bef030483","Type":"ContainerStarted","Data":"a4ca53f49741d8ce97a4d23d07f4ee3bee8d07d6dabdebb9b9d84110579cfc7d"} Oct 05 21:25:19 crc kubenswrapper[4754]: I1005 21:25:19.388879 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-csfdg" event={"ID":"35d09fe3-0746-4e6c-82b2-cc5bef030483","Type":"ContainerStarted","Data":"db03c97b1feeec0616f2e321849ed08f2eb3b331a5109f8cbd58e40b1107963e"} Oct 05 21:25:22 crc kubenswrapper[4754]: I1005 21:25:22.056070 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-csfdg" podStartSLOduration=4.525205903 podStartE2EDuration="5.056046477s" podCreationTimestamp="2025-10-05 21:25:17 +0000 UTC" firstStartedPulling="2025-10-05 21:25:18.506839393 +0000 UTC m=+1842.410958113" lastFinishedPulling="2025-10-05 21:25:19.037679947 +0000 UTC m=+1842.941798687" observedRunningTime="2025-10-05 21:25:19.412472825 +0000 UTC m=+1843.316591575" watchObservedRunningTime="2025-10-05 21:25:22.056046477 +0000 UTC m=+1845.960165197" Oct 05 21:25:22 crc kubenswrapper[4754]: I1005 21:25:22.064213 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-khclj"] Oct 05 21:25:22 crc kubenswrapper[4754]: I1005 21:25:22.079574 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-fpsrz"] Oct 05 21:25:22 crc kubenswrapper[4754]: I1005 21:25:22.093513 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-khclj"] Oct 05 21:25:22 crc kubenswrapper[4754]: I1005 21:25:22.110343 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-fpsrz"] Oct 05 21:25:22 crc kubenswrapper[4754]: I1005 21:25:22.849948 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5bc88d1c-ec79-415b-a441-86b68fdf064f" path="/var/lib/kubelet/pods/5bc88d1c-ec79-415b-a441-86b68fdf064f/volumes" Oct 05 21:25:22 crc kubenswrapper[4754]: I1005 21:25:22.851029 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d2683cf4-4f3b-4843-b859-75f9c38dfc4d" path="/var/lib/kubelet/pods/d2683cf4-4f3b-4843-b859-75f9c38dfc4d/volumes" Oct 05 21:25:25 crc kubenswrapper[4754]: I1005 21:25:25.838196 4754 scope.go:117] "RemoveContainer" containerID="e8573d2f52a359de8bfb0cb5d4d1a72d653442930ef95a60a1d6d2dd509ab1ac" Oct 05 21:25:25 crc kubenswrapper[4754]: E1005 21:25:25.839072 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 21:25:40 crc kubenswrapper[4754]: I1005 21:25:40.837903 4754 scope.go:117] "RemoveContainer" containerID="e8573d2f52a359de8bfb0cb5d4d1a72d653442930ef95a60a1d6d2dd509ab1ac" Oct 05 21:25:41 crc kubenswrapper[4754]: I1005 21:25:41.669483 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" event={"ID":"be95a413-6a59-45b4-84b7-b43cae694a26","Type":"ContainerStarted","Data":"fee409256f2bba9979dd69800f220c1257ab9bdb2e760315daae28edefc3f914"} Oct 05 21:25:55 crc kubenswrapper[4754]: I1005 21:25:55.476254 4754 scope.go:117] "RemoveContainer" containerID="b93f6db8a9b3fdf25a9dcf220ebb50e302b178a1bef18437f5f0cf17d7d6d3a4" Oct 05 21:25:55 crc kubenswrapper[4754]: I1005 21:25:55.519062 4754 scope.go:117] "RemoveContainer" containerID="9ead6f708e274a7942637dece32316d7cf2de9ff003c6fa19eb76d7bc9ad792b" Oct 05 21:26:05 crc kubenswrapper[4754]: I1005 21:26:05.066994 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-stkc2"] Oct 05 21:26:05 crc kubenswrapper[4754]: I1005 21:26:05.081707 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-stkc2"] Oct 05 21:26:05 crc kubenswrapper[4754]: I1005 21:26:05.940020 4754 generic.go:334] "Generic (PLEG): container finished" podID="35d09fe3-0746-4e6c-82b2-cc5bef030483" containerID="a4ca53f49741d8ce97a4d23d07f4ee3bee8d07d6dabdebb9b9d84110579cfc7d" exitCode=0 Oct 05 21:26:05 crc kubenswrapper[4754]: I1005 21:26:05.940107 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-csfdg" event={"ID":"35d09fe3-0746-4e6c-82b2-cc5bef030483","Type":"ContainerDied","Data":"a4ca53f49741d8ce97a4d23d07f4ee3bee8d07d6dabdebb9b9d84110579cfc7d"} Oct 05 21:26:06 crc kubenswrapper[4754]: I1005 21:26:06.857296 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ba3c8f4e-033c-48d4-b37e-7b2542be3bd1" path="/var/lib/kubelet/pods/ba3c8f4e-033c-48d4-b37e-7b2542be3bd1/volumes" Oct 05 21:26:07 crc kubenswrapper[4754]: I1005 21:26:07.436122 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-csfdg" Oct 05 21:26:07 crc kubenswrapper[4754]: I1005 21:26:07.450026 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/35d09fe3-0746-4e6c-82b2-cc5bef030483-ssh-key\") pod \"35d09fe3-0746-4e6c-82b2-cc5bef030483\" (UID: \"35d09fe3-0746-4e6c-82b2-cc5bef030483\") " Oct 05 21:26:07 crc kubenswrapper[4754]: I1005 21:26:07.450324 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hj88s\" (UniqueName: \"kubernetes.io/projected/35d09fe3-0746-4e6c-82b2-cc5bef030483-kube-api-access-hj88s\") pod \"35d09fe3-0746-4e6c-82b2-cc5bef030483\" (UID: \"35d09fe3-0746-4e6c-82b2-cc5bef030483\") " Oct 05 21:26:07 crc kubenswrapper[4754]: I1005 21:26:07.450443 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/35d09fe3-0746-4e6c-82b2-cc5bef030483-inventory\") pod \"35d09fe3-0746-4e6c-82b2-cc5bef030483\" (UID: \"35d09fe3-0746-4e6c-82b2-cc5bef030483\") " Oct 05 21:26:07 crc kubenswrapper[4754]: I1005 21:26:07.466255 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/35d09fe3-0746-4e6c-82b2-cc5bef030483-kube-api-access-hj88s" (OuterVolumeSpecName: "kube-api-access-hj88s") pod "35d09fe3-0746-4e6c-82b2-cc5bef030483" (UID: "35d09fe3-0746-4e6c-82b2-cc5bef030483"). InnerVolumeSpecName "kube-api-access-hj88s". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:26:07 crc kubenswrapper[4754]: I1005 21:26:07.483378 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35d09fe3-0746-4e6c-82b2-cc5bef030483-inventory" (OuterVolumeSpecName: "inventory") pod "35d09fe3-0746-4e6c-82b2-cc5bef030483" (UID: "35d09fe3-0746-4e6c-82b2-cc5bef030483"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:26:07 crc kubenswrapper[4754]: I1005 21:26:07.508673 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35d09fe3-0746-4e6c-82b2-cc5bef030483-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "35d09fe3-0746-4e6c-82b2-cc5bef030483" (UID: "35d09fe3-0746-4e6c-82b2-cc5bef030483"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:26:07 crc kubenswrapper[4754]: I1005 21:26:07.552759 4754 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/35d09fe3-0746-4e6c-82b2-cc5bef030483-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 05 21:26:07 crc kubenswrapper[4754]: I1005 21:26:07.552802 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hj88s\" (UniqueName: \"kubernetes.io/projected/35d09fe3-0746-4e6c-82b2-cc5bef030483-kube-api-access-hj88s\") on node \"crc\" DevicePath \"\"" Oct 05 21:26:07 crc kubenswrapper[4754]: I1005 21:26:07.552814 4754 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/35d09fe3-0746-4e6c-82b2-cc5bef030483-inventory\") on node \"crc\" DevicePath \"\"" Oct 05 21:26:07 crc kubenswrapper[4754]: I1005 21:26:07.968692 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-csfdg" event={"ID":"35d09fe3-0746-4e6c-82b2-cc5bef030483","Type":"ContainerDied","Data":"db03c97b1feeec0616f2e321849ed08f2eb3b331a5109f8cbd58e40b1107963e"} Oct 05 21:26:07 crc kubenswrapper[4754]: I1005 21:26:07.969648 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="db03c97b1feeec0616f2e321849ed08f2eb3b331a5109f8cbd58e40b1107963e" Oct 05 21:26:07 crc kubenswrapper[4754]: I1005 21:26:07.968981 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-csfdg" Oct 05 21:26:08 crc kubenswrapper[4754]: I1005 21:26:08.166136 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-p2gbk"] Oct 05 21:26:08 crc kubenswrapper[4754]: E1005 21:26:08.177974 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35d09fe3-0746-4e6c-82b2-cc5bef030483" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Oct 05 21:26:08 crc kubenswrapper[4754]: I1005 21:26:08.178006 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="35d09fe3-0746-4e6c-82b2-cc5bef030483" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Oct 05 21:26:08 crc kubenswrapper[4754]: I1005 21:26:08.178235 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="35d09fe3-0746-4e6c-82b2-cc5bef030483" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Oct 05 21:26:08 crc kubenswrapper[4754]: I1005 21:26:08.178999 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-p2gbk" Oct 05 21:26:08 crc kubenswrapper[4754]: I1005 21:26:08.183991 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-v54jd" Oct 05 21:26:08 crc kubenswrapper[4754]: I1005 21:26:08.184213 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 05 21:26:08 crc kubenswrapper[4754]: I1005 21:26:08.184245 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 05 21:26:08 crc kubenswrapper[4754]: I1005 21:26:08.184321 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 05 21:26:08 crc kubenswrapper[4754]: I1005 21:26:08.202242 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-p2gbk"] Oct 05 21:26:08 crc kubenswrapper[4754]: I1005 21:26:08.370310 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/50e8c589-a949-4193-94a1-022fea47e2ba-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-p2gbk\" (UID: \"50e8c589-a949-4193-94a1-022fea47e2ba\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-p2gbk" Oct 05 21:26:08 crc kubenswrapper[4754]: I1005 21:26:08.370689 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tff2m\" (UniqueName: \"kubernetes.io/projected/50e8c589-a949-4193-94a1-022fea47e2ba-kube-api-access-tff2m\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-p2gbk\" (UID: \"50e8c589-a949-4193-94a1-022fea47e2ba\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-p2gbk" Oct 05 21:26:08 crc kubenswrapper[4754]: I1005 21:26:08.370802 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/50e8c589-a949-4193-94a1-022fea47e2ba-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-p2gbk\" (UID: \"50e8c589-a949-4193-94a1-022fea47e2ba\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-p2gbk" Oct 05 21:26:08 crc kubenswrapper[4754]: I1005 21:26:08.473721 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/50e8c589-a949-4193-94a1-022fea47e2ba-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-p2gbk\" (UID: \"50e8c589-a949-4193-94a1-022fea47e2ba\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-p2gbk" Oct 05 21:26:08 crc kubenswrapper[4754]: I1005 21:26:08.473785 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tff2m\" (UniqueName: \"kubernetes.io/projected/50e8c589-a949-4193-94a1-022fea47e2ba-kube-api-access-tff2m\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-p2gbk\" (UID: \"50e8c589-a949-4193-94a1-022fea47e2ba\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-p2gbk" Oct 05 21:26:08 crc kubenswrapper[4754]: I1005 21:26:08.473878 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/50e8c589-a949-4193-94a1-022fea47e2ba-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-p2gbk\" (UID: \"50e8c589-a949-4193-94a1-022fea47e2ba\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-p2gbk" Oct 05 21:26:08 crc kubenswrapper[4754]: I1005 21:26:08.480410 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/50e8c589-a949-4193-94a1-022fea47e2ba-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-p2gbk\" (UID: \"50e8c589-a949-4193-94a1-022fea47e2ba\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-p2gbk" Oct 05 21:26:08 crc kubenswrapper[4754]: I1005 21:26:08.484901 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/50e8c589-a949-4193-94a1-022fea47e2ba-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-p2gbk\" (UID: \"50e8c589-a949-4193-94a1-022fea47e2ba\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-p2gbk" Oct 05 21:26:08 crc kubenswrapper[4754]: I1005 21:26:08.493817 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tff2m\" (UniqueName: \"kubernetes.io/projected/50e8c589-a949-4193-94a1-022fea47e2ba-kube-api-access-tff2m\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-p2gbk\" (UID: \"50e8c589-a949-4193-94a1-022fea47e2ba\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-p2gbk" Oct 05 21:26:08 crc kubenswrapper[4754]: I1005 21:26:08.525017 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-p2gbk" Oct 05 21:26:09 crc kubenswrapper[4754]: W1005 21:26:09.075673 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod50e8c589_a949_4193_94a1_022fea47e2ba.slice/crio-aa0f1e3c5a2a8fb0c92e563cfbdd9cd0c9a1e0128b801da05f882f4248e6bd05 WatchSource:0}: Error finding container aa0f1e3c5a2a8fb0c92e563cfbdd9cd0c9a1e0128b801da05f882f4248e6bd05: Status 404 returned error can't find the container with id aa0f1e3c5a2a8fb0c92e563cfbdd9cd0c9a1e0128b801da05f882f4248e6bd05 Oct 05 21:26:09 crc kubenswrapper[4754]: I1005 21:26:09.076355 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-p2gbk"] Oct 05 21:26:09 crc kubenswrapper[4754]: I1005 21:26:09.991069 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-p2gbk" event={"ID":"50e8c589-a949-4193-94a1-022fea47e2ba","Type":"ContainerStarted","Data":"d62ed854246ee9e7dde84dfddf0481401c0686e2544b6a0a5ce363afc46098fb"} Oct 05 21:26:09 crc kubenswrapper[4754]: I1005 21:26:09.991557 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-p2gbk" event={"ID":"50e8c589-a949-4193-94a1-022fea47e2ba","Type":"ContainerStarted","Data":"aa0f1e3c5a2a8fb0c92e563cfbdd9cd0c9a1e0128b801da05f882f4248e6bd05"} Oct 05 21:26:10 crc kubenswrapper[4754]: I1005 21:26:10.013637 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-p2gbk" podStartSLOduration=1.531387446 podStartE2EDuration="2.013616743s" podCreationTimestamp="2025-10-05 21:26:08 +0000 UTC" firstStartedPulling="2025-10-05 21:26:09.078721449 +0000 UTC m=+1892.982840179" lastFinishedPulling="2025-10-05 21:26:09.560950726 +0000 UTC m=+1893.465069476" observedRunningTime="2025-10-05 21:26:10.007143874 +0000 UTC m=+1893.911262584" watchObservedRunningTime="2025-10-05 21:26:10.013616743 +0000 UTC m=+1893.917735453" Oct 05 21:26:55 crc kubenswrapper[4754]: I1005 21:26:55.627371 4754 scope.go:117] "RemoveContainer" containerID="aff0a26aedb047df62a0a5b78a35b03c9c74f10020d36a1af66b2a6b57901861" Oct 05 21:27:10 crc kubenswrapper[4754]: I1005 21:27:10.682387 4754 generic.go:334] "Generic (PLEG): container finished" podID="50e8c589-a949-4193-94a1-022fea47e2ba" containerID="d62ed854246ee9e7dde84dfddf0481401c0686e2544b6a0a5ce363afc46098fb" exitCode=2 Oct 05 21:27:10 crc kubenswrapper[4754]: I1005 21:27:10.682670 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-p2gbk" event={"ID":"50e8c589-a949-4193-94a1-022fea47e2ba","Type":"ContainerDied","Data":"d62ed854246ee9e7dde84dfddf0481401c0686e2544b6a0a5ce363afc46098fb"} Oct 05 21:27:12 crc kubenswrapper[4754]: I1005 21:27:12.216107 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-p2gbk" Oct 05 21:27:12 crc kubenswrapper[4754]: I1005 21:27:12.297672 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/50e8c589-a949-4193-94a1-022fea47e2ba-ssh-key\") pod \"50e8c589-a949-4193-94a1-022fea47e2ba\" (UID: \"50e8c589-a949-4193-94a1-022fea47e2ba\") " Oct 05 21:27:12 crc kubenswrapper[4754]: I1005 21:27:12.297845 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tff2m\" (UniqueName: \"kubernetes.io/projected/50e8c589-a949-4193-94a1-022fea47e2ba-kube-api-access-tff2m\") pod \"50e8c589-a949-4193-94a1-022fea47e2ba\" (UID: \"50e8c589-a949-4193-94a1-022fea47e2ba\") " Oct 05 21:27:12 crc kubenswrapper[4754]: I1005 21:27:12.297920 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/50e8c589-a949-4193-94a1-022fea47e2ba-inventory\") pod \"50e8c589-a949-4193-94a1-022fea47e2ba\" (UID: \"50e8c589-a949-4193-94a1-022fea47e2ba\") " Oct 05 21:27:12 crc kubenswrapper[4754]: I1005 21:27:12.321953 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/50e8c589-a949-4193-94a1-022fea47e2ba-kube-api-access-tff2m" (OuterVolumeSpecName: "kube-api-access-tff2m") pod "50e8c589-a949-4193-94a1-022fea47e2ba" (UID: "50e8c589-a949-4193-94a1-022fea47e2ba"). InnerVolumeSpecName "kube-api-access-tff2m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:27:12 crc kubenswrapper[4754]: I1005 21:27:12.329129 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50e8c589-a949-4193-94a1-022fea47e2ba-inventory" (OuterVolumeSpecName: "inventory") pod "50e8c589-a949-4193-94a1-022fea47e2ba" (UID: "50e8c589-a949-4193-94a1-022fea47e2ba"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:27:12 crc kubenswrapper[4754]: I1005 21:27:12.353209 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50e8c589-a949-4193-94a1-022fea47e2ba-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "50e8c589-a949-4193-94a1-022fea47e2ba" (UID: "50e8c589-a949-4193-94a1-022fea47e2ba"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:27:12 crc kubenswrapper[4754]: I1005 21:27:12.400949 4754 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/50e8c589-a949-4193-94a1-022fea47e2ba-inventory\") on node \"crc\" DevicePath \"\"" Oct 05 21:27:12 crc kubenswrapper[4754]: I1005 21:27:12.400992 4754 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/50e8c589-a949-4193-94a1-022fea47e2ba-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 05 21:27:12 crc kubenswrapper[4754]: I1005 21:27:12.401005 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tff2m\" (UniqueName: \"kubernetes.io/projected/50e8c589-a949-4193-94a1-022fea47e2ba-kube-api-access-tff2m\") on node \"crc\" DevicePath \"\"" Oct 05 21:27:12 crc kubenswrapper[4754]: I1005 21:27:12.702276 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-p2gbk" event={"ID":"50e8c589-a949-4193-94a1-022fea47e2ba","Type":"ContainerDied","Data":"aa0f1e3c5a2a8fb0c92e563cfbdd9cd0c9a1e0128b801da05f882f4248e6bd05"} Oct 05 21:27:12 crc kubenswrapper[4754]: I1005 21:27:12.702335 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aa0f1e3c5a2a8fb0c92e563cfbdd9cd0c9a1e0128b801da05f882f4248e6bd05" Oct 05 21:27:12 crc kubenswrapper[4754]: I1005 21:27:12.702428 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-p2gbk" Oct 05 21:27:20 crc kubenswrapper[4754]: I1005 21:27:20.038779 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-k54sd"] Oct 05 21:27:20 crc kubenswrapper[4754]: E1005 21:27:20.039946 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50e8c589-a949-4193-94a1-022fea47e2ba" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 05 21:27:20 crc kubenswrapper[4754]: I1005 21:27:20.039964 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="50e8c589-a949-4193-94a1-022fea47e2ba" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 05 21:27:20 crc kubenswrapper[4754]: I1005 21:27:20.040224 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="50e8c589-a949-4193-94a1-022fea47e2ba" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 05 21:27:20 crc kubenswrapper[4754]: I1005 21:27:20.040875 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-k54sd" Oct 05 21:27:20 crc kubenswrapper[4754]: I1005 21:27:20.044248 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 05 21:27:20 crc kubenswrapper[4754]: I1005 21:27:20.044657 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 05 21:27:20 crc kubenswrapper[4754]: I1005 21:27:20.044672 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 05 21:27:20 crc kubenswrapper[4754]: I1005 21:27:20.048789 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-v54jd" Oct 05 21:27:20 crc kubenswrapper[4754]: I1005 21:27:20.066761 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-k54sd"] Oct 05 21:27:20 crc kubenswrapper[4754]: I1005 21:27:20.127739 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/77f69424-dcfc-4915-b704-3447e3cf31eb-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-k54sd\" (UID: \"77f69424-dcfc-4915-b704-3447e3cf31eb\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-k54sd" Oct 05 21:27:20 crc kubenswrapper[4754]: I1005 21:27:20.127995 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/77f69424-dcfc-4915-b704-3447e3cf31eb-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-k54sd\" (UID: \"77f69424-dcfc-4915-b704-3447e3cf31eb\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-k54sd" Oct 05 21:27:20 crc kubenswrapper[4754]: I1005 21:27:20.128101 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8xslh\" (UniqueName: \"kubernetes.io/projected/77f69424-dcfc-4915-b704-3447e3cf31eb-kube-api-access-8xslh\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-k54sd\" (UID: \"77f69424-dcfc-4915-b704-3447e3cf31eb\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-k54sd" Oct 05 21:27:20 crc kubenswrapper[4754]: I1005 21:27:20.230186 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/77f69424-dcfc-4915-b704-3447e3cf31eb-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-k54sd\" (UID: \"77f69424-dcfc-4915-b704-3447e3cf31eb\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-k54sd" Oct 05 21:27:20 crc kubenswrapper[4754]: I1005 21:27:20.230256 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8xslh\" (UniqueName: \"kubernetes.io/projected/77f69424-dcfc-4915-b704-3447e3cf31eb-kube-api-access-8xslh\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-k54sd\" (UID: \"77f69424-dcfc-4915-b704-3447e3cf31eb\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-k54sd" Oct 05 21:27:20 crc kubenswrapper[4754]: I1005 21:27:20.230335 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/77f69424-dcfc-4915-b704-3447e3cf31eb-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-k54sd\" (UID: \"77f69424-dcfc-4915-b704-3447e3cf31eb\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-k54sd" Oct 05 21:27:20 crc kubenswrapper[4754]: I1005 21:27:20.241843 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/77f69424-dcfc-4915-b704-3447e3cf31eb-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-k54sd\" (UID: \"77f69424-dcfc-4915-b704-3447e3cf31eb\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-k54sd" Oct 05 21:27:20 crc kubenswrapper[4754]: I1005 21:27:20.242284 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/77f69424-dcfc-4915-b704-3447e3cf31eb-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-k54sd\" (UID: \"77f69424-dcfc-4915-b704-3447e3cf31eb\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-k54sd" Oct 05 21:27:20 crc kubenswrapper[4754]: I1005 21:27:20.262473 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8xslh\" (UniqueName: \"kubernetes.io/projected/77f69424-dcfc-4915-b704-3447e3cf31eb-kube-api-access-8xslh\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-k54sd\" (UID: \"77f69424-dcfc-4915-b704-3447e3cf31eb\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-k54sd" Oct 05 21:27:20 crc kubenswrapper[4754]: I1005 21:27:20.377642 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-k54sd" Oct 05 21:27:21 crc kubenswrapper[4754]: I1005 21:27:21.015107 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-k54sd"] Oct 05 21:27:21 crc kubenswrapper[4754]: I1005 21:27:21.838021 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-k54sd" event={"ID":"77f69424-dcfc-4915-b704-3447e3cf31eb","Type":"ContainerStarted","Data":"c433607da97e5a0a2c6603fd4c78b74dc4d03b61b5b76b79c6b89f306355bef4"} Oct 05 21:27:22 crc kubenswrapper[4754]: I1005 21:27:22.862562 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-k54sd" event={"ID":"77f69424-dcfc-4915-b704-3447e3cf31eb","Type":"ContainerStarted","Data":"627bd541dbba96274dc486b777231db2dae67c7cf9bce778f8ac9b3da46d5c05"} Oct 05 21:27:22 crc kubenswrapper[4754]: I1005 21:27:22.883379 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-k54sd" podStartSLOduration=2.302371514 podStartE2EDuration="2.883302854s" podCreationTimestamp="2025-10-05 21:27:20 +0000 UTC" firstStartedPulling="2025-10-05 21:27:21.030760392 +0000 UTC m=+1964.934879102" lastFinishedPulling="2025-10-05 21:27:21.611691702 +0000 UTC m=+1965.515810442" observedRunningTime="2025-10-05 21:27:22.881390023 +0000 UTC m=+1966.785508743" watchObservedRunningTime="2025-10-05 21:27:22.883302854 +0000 UTC m=+1966.787421574" Oct 05 21:28:05 crc kubenswrapper[4754]: I1005 21:28:05.244868 4754 patch_prober.go:28] interesting pod/machine-config-daemon-b2h9k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 21:28:05 crc kubenswrapper[4754]: I1005 21:28:05.245570 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 21:28:21 crc kubenswrapper[4754]: I1005 21:28:21.507389 4754 generic.go:334] "Generic (PLEG): container finished" podID="77f69424-dcfc-4915-b704-3447e3cf31eb" containerID="627bd541dbba96274dc486b777231db2dae67c7cf9bce778f8ac9b3da46d5c05" exitCode=0 Oct 05 21:28:21 crc kubenswrapper[4754]: I1005 21:28:21.507613 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-k54sd" event={"ID":"77f69424-dcfc-4915-b704-3447e3cf31eb","Type":"ContainerDied","Data":"627bd541dbba96274dc486b777231db2dae67c7cf9bce778f8ac9b3da46d5c05"} Oct 05 21:28:22 crc kubenswrapper[4754]: I1005 21:28:22.976454 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-k54sd" Oct 05 21:28:23 crc kubenswrapper[4754]: I1005 21:28:23.064772 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/77f69424-dcfc-4915-b704-3447e3cf31eb-ssh-key\") pod \"77f69424-dcfc-4915-b704-3447e3cf31eb\" (UID: \"77f69424-dcfc-4915-b704-3447e3cf31eb\") " Oct 05 21:28:23 crc kubenswrapper[4754]: I1005 21:28:23.064907 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/77f69424-dcfc-4915-b704-3447e3cf31eb-inventory\") pod \"77f69424-dcfc-4915-b704-3447e3cf31eb\" (UID: \"77f69424-dcfc-4915-b704-3447e3cf31eb\") " Oct 05 21:28:23 crc kubenswrapper[4754]: I1005 21:28:23.064967 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8xslh\" (UniqueName: \"kubernetes.io/projected/77f69424-dcfc-4915-b704-3447e3cf31eb-kube-api-access-8xslh\") pod \"77f69424-dcfc-4915-b704-3447e3cf31eb\" (UID: \"77f69424-dcfc-4915-b704-3447e3cf31eb\") " Oct 05 21:28:23 crc kubenswrapper[4754]: I1005 21:28:23.071746 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/77f69424-dcfc-4915-b704-3447e3cf31eb-kube-api-access-8xslh" (OuterVolumeSpecName: "kube-api-access-8xslh") pod "77f69424-dcfc-4915-b704-3447e3cf31eb" (UID: "77f69424-dcfc-4915-b704-3447e3cf31eb"). InnerVolumeSpecName "kube-api-access-8xslh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:28:23 crc kubenswrapper[4754]: I1005 21:28:23.093537 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/77f69424-dcfc-4915-b704-3447e3cf31eb-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "77f69424-dcfc-4915-b704-3447e3cf31eb" (UID: "77f69424-dcfc-4915-b704-3447e3cf31eb"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:28:23 crc kubenswrapper[4754]: I1005 21:28:23.101412 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/77f69424-dcfc-4915-b704-3447e3cf31eb-inventory" (OuterVolumeSpecName: "inventory") pod "77f69424-dcfc-4915-b704-3447e3cf31eb" (UID: "77f69424-dcfc-4915-b704-3447e3cf31eb"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:28:23 crc kubenswrapper[4754]: I1005 21:28:23.167582 4754 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/77f69424-dcfc-4915-b704-3447e3cf31eb-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 05 21:28:23 crc kubenswrapper[4754]: I1005 21:28:23.167621 4754 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/77f69424-dcfc-4915-b704-3447e3cf31eb-inventory\") on node \"crc\" DevicePath \"\"" Oct 05 21:28:23 crc kubenswrapper[4754]: I1005 21:28:23.167632 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8xslh\" (UniqueName: \"kubernetes.io/projected/77f69424-dcfc-4915-b704-3447e3cf31eb-kube-api-access-8xslh\") on node \"crc\" DevicePath \"\"" Oct 05 21:28:23 crc kubenswrapper[4754]: I1005 21:28:23.536548 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-k54sd" event={"ID":"77f69424-dcfc-4915-b704-3447e3cf31eb","Type":"ContainerDied","Data":"c433607da97e5a0a2c6603fd4c78b74dc4d03b61b5b76b79c6b89f306355bef4"} Oct 05 21:28:23 crc kubenswrapper[4754]: I1005 21:28:23.537191 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c433607da97e5a0a2c6603fd4c78b74dc4d03b61b5b76b79c6b89f306355bef4" Oct 05 21:28:23 crc kubenswrapper[4754]: I1005 21:28:23.536755 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-k54sd" Oct 05 21:28:23 crc kubenswrapper[4754]: I1005 21:28:23.665429 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-mx4cz"] Oct 05 21:28:23 crc kubenswrapper[4754]: E1005 21:28:23.665838 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77f69424-dcfc-4915-b704-3447e3cf31eb" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 05 21:28:23 crc kubenswrapper[4754]: I1005 21:28:23.665858 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="77f69424-dcfc-4915-b704-3447e3cf31eb" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 05 21:28:23 crc kubenswrapper[4754]: I1005 21:28:23.670446 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="77f69424-dcfc-4915-b704-3447e3cf31eb" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 05 21:28:23 crc kubenswrapper[4754]: I1005 21:28:23.671156 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-mx4cz" Oct 05 21:28:23 crc kubenswrapper[4754]: I1005 21:28:23.675154 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 05 21:28:23 crc kubenswrapper[4754]: I1005 21:28:23.675215 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 05 21:28:23 crc kubenswrapper[4754]: I1005 21:28:23.675765 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-v54jd" Oct 05 21:28:23 crc kubenswrapper[4754]: I1005 21:28:23.675915 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 05 21:28:23 crc kubenswrapper[4754]: E1005 21:28:23.739410 4754 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod77f69424_dcfc_4915_b704_3447e3cf31eb.slice/crio-c433607da97e5a0a2c6603fd4c78b74dc4d03b61b5b76b79c6b89f306355bef4\": RecentStats: unable to find data in memory cache]" Oct 05 21:28:23 crc kubenswrapper[4754]: I1005 21:28:23.756455 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-mx4cz"] Oct 05 21:28:23 crc kubenswrapper[4754]: I1005 21:28:23.778447 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/50b5cacd-e284-4f29-acb5-1908246d11d6-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-mx4cz\" (UID: \"50b5cacd-e284-4f29-acb5-1908246d11d6\") " pod="openstack/ssh-known-hosts-edpm-deployment-mx4cz" Oct 05 21:28:23 crc kubenswrapper[4754]: I1005 21:28:23.778509 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/50b5cacd-e284-4f29-acb5-1908246d11d6-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-mx4cz\" (UID: \"50b5cacd-e284-4f29-acb5-1908246d11d6\") " pod="openstack/ssh-known-hosts-edpm-deployment-mx4cz" Oct 05 21:28:23 crc kubenswrapper[4754]: I1005 21:28:23.778595 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jjnxm\" (UniqueName: \"kubernetes.io/projected/50b5cacd-e284-4f29-acb5-1908246d11d6-kube-api-access-jjnxm\") pod \"ssh-known-hosts-edpm-deployment-mx4cz\" (UID: \"50b5cacd-e284-4f29-acb5-1908246d11d6\") " pod="openstack/ssh-known-hosts-edpm-deployment-mx4cz" Oct 05 21:28:23 crc kubenswrapper[4754]: I1005 21:28:23.881184 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/50b5cacd-e284-4f29-acb5-1908246d11d6-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-mx4cz\" (UID: \"50b5cacd-e284-4f29-acb5-1908246d11d6\") " pod="openstack/ssh-known-hosts-edpm-deployment-mx4cz" Oct 05 21:28:23 crc kubenswrapper[4754]: I1005 21:28:23.881237 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/50b5cacd-e284-4f29-acb5-1908246d11d6-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-mx4cz\" (UID: \"50b5cacd-e284-4f29-acb5-1908246d11d6\") " pod="openstack/ssh-known-hosts-edpm-deployment-mx4cz" Oct 05 21:28:23 crc kubenswrapper[4754]: I1005 21:28:23.881310 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jjnxm\" (UniqueName: \"kubernetes.io/projected/50b5cacd-e284-4f29-acb5-1908246d11d6-kube-api-access-jjnxm\") pod \"ssh-known-hosts-edpm-deployment-mx4cz\" (UID: \"50b5cacd-e284-4f29-acb5-1908246d11d6\") " pod="openstack/ssh-known-hosts-edpm-deployment-mx4cz" Oct 05 21:28:23 crc kubenswrapper[4754]: I1005 21:28:23.889409 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/50b5cacd-e284-4f29-acb5-1908246d11d6-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-mx4cz\" (UID: \"50b5cacd-e284-4f29-acb5-1908246d11d6\") " pod="openstack/ssh-known-hosts-edpm-deployment-mx4cz" Oct 05 21:28:23 crc kubenswrapper[4754]: I1005 21:28:23.889873 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/50b5cacd-e284-4f29-acb5-1908246d11d6-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-mx4cz\" (UID: \"50b5cacd-e284-4f29-acb5-1908246d11d6\") " pod="openstack/ssh-known-hosts-edpm-deployment-mx4cz" Oct 05 21:28:23 crc kubenswrapper[4754]: I1005 21:28:23.897532 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jjnxm\" (UniqueName: \"kubernetes.io/projected/50b5cacd-e284-4f29-acb5-1908246d11d6-kube-api-access-jjnxm\") pod \"ssh-known-hosts-edpm-deployment-mx4cz\" (UID: \"50b5cacd-e284-4f29-acb5-1908246d11d6\") " pod="openstack/ssh-known-hosts-edpm-deployment-mx4cz" Oct 05 21:28:24 crc kubenswrapper[4754]: I1005 21:28:24.052585 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-mx4cz" Oct 05 21:28:24 crc kubenswrapper[4754]: I1005 21:28:24.744098 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-mx4cz"] Oct 05 21:28:25 crc kubenswrapper[4754]: I1005 21:28:25.558734 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-mx4cz" event={"ID":"50b5cacd-e284-4f29-acb5-1908246d11d6","Type":"ContainerStarted","Data":"f7585a6b004722e3bb98d94281b9973ffd3b29842ae26ae74222e7dda97c8c18"} Oct 05 21:28:25 crc kubenswrapper[4754]: I1005 21:28:25.559215 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-mx4cz" event={"ID":"50b5cacd-e284-4f29-acb5-1908246d11d6","Type":"ContainerStarted","Data":"5a90a7a790d0b6cd9bf20cf8e8e2545fc066b36c525e608a0438f20cfa9e0356"} Oct 05 21:28:25 crc kubenswrapper[4754]: I1005 21:28:25.579195 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-mx4cz" podStartSLOduration=2.065809995 podStartE2EDuration="2.579167393s" podCreationTimestamp="2025-10-05 21:28:23 +0000 UTC" firstStartedPulling="2025-10-05 21:28:24.710193252 +0000 UTC m=+2028.614311972" lastFinishedPulling="2025-10-05 21:28:25.22355065 +0000 UTC m=+2029.127669370" observedRunningTime="2025-10-05 21:28:25.578424413 +0000 UTC m=+2029.482543143" watchObservedRunningTime="2025-10-05 21:28:25.579167393 +0000 UTC m=+2029.483286143" Oct 05 21:28:34 crc kubenswrapper[4754]: I1005 21:28:34.676591 4754 generic.go:334] "Generic (PLEG): container finished" podID="50b5cacd-e284-4f29-acb5-1908246d11d6" containerID="f7585a6b004722e3bb98d94281b9973ffd3b29842ae26ae74222e7dda97c8c18" exitCode=0 Oct 05 21:28:34 crc kubenswrapper[4754]: I1005 21:28:34.676722 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-mx4cz" event={"ID":"50b5cacd-e284-4f29-acb5-1908246d11d6","Type":"ContainerDied","Data":"f7585a6b004722e3bb98d94281b9973ffd3b29842ae26ae74222e7dda97c8c18"} Oct 05 21:28:35 crc kubenswrapper[4754]: I1005 21:28:35.244876 4754 patch_prober.go:28] interesting pod/machine-config-daemon-b2h9k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 21:28:35 crc kubenswrapper[4754]: I1005 21:28:35.244996 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 21:28:36 crc kubenswrapper[4754]: I1005 21:28:36.187839 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-mx4cz" Oct 05 21:28:36 crc kubenswrapper[4754]: I1005 21:28:36.315564 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/50b5cacd-e284-4f29-acb5-1908246d11d6-ssh-key-openstack-edpm-ipam\") pod \"50b5cacd-e284-4f29-acb5-1908246d11d6\" (UID: \"50b5cacd-e284-4f29-acb5-1908246d11d6\") " Oct 05 21:28:36 crc kubenswrapper[4754]: I1005 21:28:36.315700 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jjnxm\" (UniqueName: \"kubernetes.io/projected/50b5cacd-e284-4f29-acb5-1908246d11d6-kube-api-access-jjnxm\") pod \"50b5cacd-e284-4f29-acb5-1908246d11d6\" (UID: \"50b5cacd-e284-4f29-acb5-1908246d11d6\") " Oct 05 21:28:36 crc kubenswrapper[4754]: I1005 21:28:36.316056 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/50b5cacd-e284-4f29-acb5-1908246d11d6-inventory-0\") pod \"50b5cacd-e284-4f29-acb5-1908246d11d6\" (UID: \"50b5cacd-e284-4f29-acb5-1908246d11d6\") " Oct 05 21:28:36 crc kubenswrapper[4754]: I1005 21:28:36.325842 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/50b5cacd-e284-4f29-acb5-1908246d11d6-kube-api-access-jjnxm" (OuterVolumeSpecName: "kube-api-access-jjnxm") pod "50b5cacd-e284-4f29-acb5-1908246d11d6" (UID: "50b5cacd-e284-4f29-acb5-1908246d11d6"). InnerVolumeSpecName "kube-api-access-jjnxm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:28:36 crc kubenswrapper[4754]: I1005 21:28:36.347393 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50b5cacd-e284-4f29-acb5-1908246d11d6-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "50b5cacd-e284-4f29-acb5-1908246d11d6" (UID: "50b5cacd-e284-4f29-acb5-1908246d11d6"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:28:36 crc kubenswrapper[4754]: I1005 21:28:36.360841 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50b5cacd-e284-4f29-acb5-1908246d11d6-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "50b5cacd-e284-4f29-acb5-1908246d11d6" (UID: "50b5cacd-e284-4f29-acb5-1908246d11d6"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:28:36 crc kubenswrapper[4754]: I1005 21:28:36.419853 4754 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/50b5cacd-e284-4f29-acb5-1908246d11d6-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Oct 05 21:28:36 crc kubenswrapper[4754]: I1005 21:28:36.420979 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jjnxm\" (UniqueName: \"kubernetes.io/projected/50b5cacd-e284-4f29-acb5-1908246d11d6-kube-api-access-jjnxm\") on node \"crc\" DevicePath \"\"" Oct 05 21:28:36 crc kubenswrapper[4754]: I1005 21:28:36.421056 4754 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/50b5cacd-e284-4f29-acb5-1908246d11d6-inventory-0\") on node \"crc\" DevicePath \"\"" Oct 05 21:28:36 crc kubenswrapper[4754]: I1005 21:28:36.703339 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-mx4cz" event={"ID":"50b5cacd-e284-4f29-acb5-1908246d11d6","Type":"ContainerDied","Data":"5a90a7a790d0b6cd9bf20cf8e8e2545fc066b36c525e608a0438f20cfa9e0356"} Oct 05 21:28:36 crc kubenswrapper[4754]: I1005 21:28:36.703384 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5a90a7a790d0b6cd9bf20cf8e8e2545fc066b36c525e608a0438f20cfa9e0356" Oct 05 21:28:36 crc kubenswrapper[4754]: I1005 21:28:36.703454 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-mx4cz" Oct 05 21:28:36 crc kubenswrapper[4754]: I1005 21:28:36.832184 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-m5p76"] Oct 05 21:28:36 crc kubenswrapper[4754]: E1005 21:28:36.832650 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50b5cacd-e284-4f29-acb5-1908246d11d6" containerName="ssh-known-hosts-edpm-deployment" Oct 05 21:28:36 crc kubenswrapper[4754]: I1005 21:28:36.832668 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="50b5cacd-e284-4f29-acb5-1908246d11d6" containerName="ssh-known-hosts-edpm-deployment" Oct 05 21:28:36 crc kubenswrapper[4754]: I1005 21:28:36.832861 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="50b5cacd-e284-4f29-acb5-1908246d11d6" containerName="ssh-known-hosts-edpm-deployment" Oct 05 21:28:36 crc kubenswrapper[4754]: I1005 21:28:36.833548 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-m5p76" Oct 05 21:28:36 crc kubenswrapper[4754]: I1005 21:28:36.836968 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 05 21:28:36 crc kubenswrapper[4754]: I1005 21:28:36.839700 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-v54jd" Oct 05 21:28:36 crc kubenswrapper[4754]: I1005 21:28:36.839699 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 05 21:28:36 crc kubenswrapper[4754]: I1005 21:28:36.844311 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 05 21:28:36 crc kubenswrapper[4754]: I1005 21:28:36.885130 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b30dea9b-eb5b-46c2-88ed-e5b6df21356a-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-m5p76\" (UID: \"b30dea9b-eb5b-46c2-88ed-e5b6df21356a\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-m5p76" Oct 05 21:28:36 crc kubenswrapper[4754]: I1005 21:28:36.885198 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7bjlj\" (UniqueName: \"kubernetes.io/projected/b30dea9b-eb5b-46c2-88ed-e5b6df21356a-kube-api-access-7bjlj\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-m5p76\" (UID: \"b30dea9b-eb5b-46c2-88ed-e5b6df21356a\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-m5p76" Oct 05 21:28:36 crc kubenswrapper[4754]: I1005 21:28:36.885368 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b30dea9b-eb5b-46c2-88ed-e5b6df21356a-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-m5p76\" (UID: \"b30dea9b-eb5b-46c2-88ed-e5b6df21356a\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-m5p76" Oct 05 21:28:36 crc kubenswrapper[4754]: I1005 21:28:36.889112 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-m5p76"] Oct 05 21:28:36 crc kubenswrapper[4754]: I1005 21:28:36.986556 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b30dea9b-eb5b-46c2-88ed-e5b6df21356a-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-m5p76\" (UID: \"b30dea9b-eb5b-46c2-88ed-e5b6df21356a\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-m5p76" Oct 05 21:28:36 crc kubenswrapper[4754]: I1005 21:28:36.986667 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7bjlj\" (UniqueName: \"kubernetes.io/projected/b30dea9b-eb5b-46c2-88ed-e5b6df21356a-kube-api-access-7bjlj\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-m5p76\" (UID: \"b30dea9b-eb5b-46c2-88ed-e5b6df21356a\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-m5p76" Oct 05 21:28:36 crc kubenswrapper[4754]: I1005 21:28:36.986757 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b30dea9b-eb5b-46c2-88ed-e5b6df21356a-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-m5p76\" (UID: \"b30dea9b-eb5b-46c2-88ed-e5b6df21356a\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-m5p76" Oct 05 21:28:36 crc kubenswrapper[4754]: I1005 21:28:36.991123 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b30dea9b-eb5b-46c2-88ed-e5b6df21356a-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-m5p76\" (UID: \"b30dea9b-eb5b-46c2-88ed-e5b6df21356a\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-m5p76" Oct 05 21:28:36 crc kubenswrapper[4754]: I1005 21:28:36.991180 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b30dea9b-eb5b-46c2-88ed-e5b6df21356a-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-m5p76\" (UID: \"b30dea9b-eb5b-46c2-88ed-e5b6df21356a\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-m5p76" Oct 05 21:28:37 crc kubenswrapper[4754]: I1005 21:28:37.008635 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7bjlj\" (UniqueName: \"kubernetes.io/projected/b30dea9b-eb5b-46c2-88ed-e5b6df21356a-kube-api-access-7bjlj\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-m5p76\" (UID: \"b30dea9b-eb5b-46c2-88ed-e5b6df21356a\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-m5p76" Oct 05 21:28:37 crc kubenswrapper[4754]: I1005 21:28:37.158419 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-m5p76" Oct 05 21:28:37 crc kubenswrapper[4754]: I1005 21:28:37.741638 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-m5p76"] Oct 05 21:28:38 crc kubenswrapper[4754]: I1005 21:28:38.731839 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-m5p76" event={"ID":"b30dea9b-eb5b-46c2-88ed-e5b6df21356a","Type":"ContainerStarted","Data":"748c2d3c26ab649f3a3caf982647e39981f24dd65682b83445dc3bea7d9b3dc2"} Oct 05 21:28:38 crc kubenswrapper[4754]: I1005 21:28:38.732708 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-m5p76" event={"ID":"b30dea9b-eb5b-46c2-88ed-e5b6df21356a","Type":"ContainerStarted","Data":"226c63f69c7fe997f7a634b54e59f5e4ea6ec571660352aab7c59691298050d3"} Oct 05 21:28:38 crc kubenswrapper[4754]: I1005 21:28:38.818507 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-m5p76" podStartSLOduration=2.319051773 podStartE2EDuration="2.818454609s" podCreationTimestamp="2025-10-05 21:28:36 +0000 UTC" firstStartedPulling="2025-10-05 21:28:37.738680297 +0000 UTC m=+2041.642799017" lastFinishedPulling="2025-10-05 21:28:38.238083103 +0000 UTC m=+2042.142201853" observedRunningTime="2025-10-05 21:28:38.782862715 +0000 UTC m=+2042.686981425" watchObservedRunningTime="2025-10-05 21:28:38.818454609 +0000 UTC m=+2042.722573319" Oct 05 21:28:49 crc kubenswrapper[4754]: I1005 21:28:49.893719 4754 generic.go:334] "Generic (PLEG): container finished" podID="b30dea9b-eb5b-46c2-88ed-e5b6df21356a" containerID="748c2d3c26ab649f3a3caf982647e39981f24dd65682b83445dc3bea7d9b3dc2" exitCode=0 Oct 05 21:28:49 crc kubenswrapper[4754]: I1005 21:28:49.893828 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-m5p76" event={"ID":"b30dea9b-eb5b-46c2-88ed-e5b6df21356a","Type":"ContainerDied","Data":"748c2d3c26ab649f3a3caf982647e39981f24dd65682b83445dc3bea7d9b3dc2"} Oct 05 21:28:51 crc kubenswrapper[4754]: I1005 21:28:51.499655 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-m5p76" Oct 05 21:28:51 crc kubenswrapper[4754]: I1005 21:28:51.580514 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7bjlj\" (UniqueName: \"kubernetes.io/projected/b30dea9b-eb5b-46c2-88ed-e5b6df21356a-kube-api-access-7bjlj\") pod \"b30dea9b-eb5b-46c2-88ed-e5b6df21356a\" (UID: \"b30dea9b-eb5b-46c2-88ed-e5b6df21356a\") " Oct 05 21:28:51 crc kubenswrapper[4754]: I1005 21:28:51.580579 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b30dea9b-eb5b-46c2-88ed-e5b6df21356a-inventory\") pod \"b30dea9b-eb5b-46c2-88ed-e5b6df21356a\" (UID: \"b30dea9b-eb5b-46c2-88ed-e5b6df21356a\") " Oct 05 21:28:51 crc kubenswrapper[4754]: I1005 21:28:51.580774 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b30dea9b-eb5b-46c2-88ed-e5b6df21356a-ssh-key\") pod \"b30dea9b-eb5b-46c2-88ed-e5b6df21356a\" (UID: \"b30dea9b-eb5b-46c2-88ed-e5b6df21356a\") " Oct 05 21:28:51 crc kubenswrapper[4754]: I1005 21:28:51.594191 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b30dea9b-eb5b-46c2-88ed-e5b6df21356a-kube-api-access-7bjlj" (OuterVolumeSpecName: "kube-api-access-7bjlj") pod "b30dea9b-eb5b-46c2-88ed-e5b6df21356a" (UID: "b30dea9b-eb5b-46c2-88ed-e5b6df21356a"). InnerVolumeSpecName "kube-api-access-7bjlj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:28:51 crc kubenswrapper[4754]: I1005 21:28:51.613529 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b30dea9b-eb5b-46c2-88ed-e5b6df21356a-inventory" (OuterVolumeSpecName: "inventory") pod "b30dea9b-eb5b-46c2-88ed-e5b6df21356a" (UID: "b30dea9b-eb5b-46c2-88ed-e5b6df21356a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:28:51 crc kubenswrapper[4754]: I1005 21:28:51.614273 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b30dea9b-eb5b-46c2-88ed-e5b6df21356a-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "b30dea9b-eb5b-46c2-88ed-e5b6df21356a" (UID: "b30dea9b-eb5b-46c2-88ed-e5b6df21356a"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:28:51 crc kubenswrapper[4754]: I1005 21:28:51.682922 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7bjlj\" (UniqueName: \"kubernetes.io/projected/b30dea9b-eb5b-46c2-88ed-e5b6df21356a-kube-api-access-7bjlj\") on node \"crc\" DevicePath \"\"" Oct 05 21:28:51 crc kubenswrapper[4754]: I1005 21:28:51.682978 4754 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b30dea9b-eb5b-46c2-88ed-e5b6df21356a-inventory\") on node \"crc\" DevicePath \"\"" Oct 05 21:28:51 crc kubenswrapper[4754]: I1005 21:28:51.682991 4754 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b30dea9b-eb5b-46c2-88ed-e5b6df21356a-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 05 21:28:51 crc kubenswrapper[4754]: I1005 21:28:51.941777 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-m5p76" event={"ID":"b30dea9b-eb5b-46c2-88ed-e5b6df21356a","Type":"ContainerDied","Data":"226c63f69c7fe997f7a634b54e59f5e4ea6ec571660352aab7c59691298050d3"} Oct 05 21:28:51 crc kubenswrapper[4754]: I1005 21:28:51.941831 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="226c63f69c7fe997f7a634b54e59f5e4ea6ec571660352aab7c59691298050d3" Oct 05 21:28:51 crc kubenswrapper[4754]: I1005 21:28:51.941929 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-m5p76" Oct 05 21:28:52 crc kubenswrapper[4754]: I1005 21:28:52.044569 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-4srnb"] Oct 05 21:28:52 crc kubenswrapper[4754]: E1005 21:28:52.045074 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b30dea9b-eb5b-46c2-88ed-e5b6df21356a" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Oct 05 21:28:52 crc kubenswrapper[4754]: I1005 21:28:52.045095 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="b30dea9b-eb5b-46c2-88ed-e5b6df21356a" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Oct 05 21:28:52 crc kubenswrapper[4754]: I1005 21:28:52.045306 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="b30dea9b-eb5b-46c2-88ed-e5b6df21356a" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Oct 05 21:28:52 crc kubenswrapper[4754]: I1005 21:28:52.046054 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-4srnb" Oct 05 21:28:52 crc kubenswrapper[4754]: I1005 21:28:52.049765 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 05 21:28:52 crc kubenswrapper[4754]: I1005 21:28:52.049881 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-v54jd" Oct 05 21:28:52 crc kubenswrapper[4754]: I1005 21:28:52.050003 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 05 21:28:52 crc kubenswrapper[4754]: I1005 21:28:52.050392 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 05 21:28:52 crc kubenswrapper[4754]: I1005 21:28:52.071940 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-4srnb"] Oct 05 21:28:52 crc kubenswrapper[4754]: I1005 21:28:52.193489 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/66ef090a-4d9c-41f2-b225-a3bce5bb58c1-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-4srnb\" (UID: \"66ef090a-4d9c-41f2-b225-a3bce5bb58c1\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-4srnb" Oct 05 21:28:52 crc kubenswrapper[4754]: I1005 21:28:52.193690 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n6twg\" (UniqueName: \"kubernetes.io/projected/66ef090a-4d9c-41f2-b225-a3bce5bb58c1-kube-api-access-n6twg\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-4srnb\" (UID: \"66ef090a-4d9c-41f2-b225-a3bce5bb58c1\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-4srnb" Oct 05 21:28:52 crc kubenswrapper[4754]: I1005 21:28:52.193800 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/66ef090a-4d9c-41f2-b225-a3bce5bb58c1-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-4srnb\" (UID: \"66ef090a-4d9c-41f2-b225-a3bce5bb58c1\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-4srnb" Oct 05 21:28:52 crc kubenswrapper[4754]: I1005 21:28:52.296272 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/66ef090a-4d9c-41f2-b225-a3bce5bb58c1-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-4srnb\" (UID: \"66ef090a-4d9c-41f2-b225-a3bce5bb58c1\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-4srnb" Oct 05 21:28:52 crc kubenswrapper[4754]: I1005 21:28:52.296425 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/66ef090a-4d9c-41f2-b225-a3bce5bb58c1-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-4srnb\" (UID: \"66ef090a-4d9c-41f2-b225-a3bce5bb58c1\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-4srnb" Oct 05 21:28:52 crc kubenswrapper[4754]: I1005 21:28:52.296668 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n6twg\" (UniqueName: \"kubernetes.io/projected/66ef090a-4d9c-41f2-b225-a3bce5bb58c1-kube-api-access-n6twg\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-4srnb\" (UID: \"66ef090a-4d9c-41f2-b225-a3bce5bb58c1\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-4srnb" Oct 05 21:28:52 crc kubenswrapper[4754]: I1005 21:28:52.311165 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/66ef090a-4d9c-41f2-b225-a3bce5bb58c1-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-4srnb\" (UID: \"66ef090a-4d9c-41f2-b225-a3bce5bb58c1\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-4srnb" Oct 05 21:28:52 crc kubenswrapper[4754]: I1005 21:28:52.311670 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/66ef090a-4d9c-41f2-b225-a3bce5bb58c1-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-4srnb\" (UID: \"66ef090a-4d9c-41f2-b225-a3bce5bb58c1\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-4srnb" Oct 05 21:28:52 crc kubenswrapper[4754]: I1005 21:28:52.329369 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n6twg\" (UniqueName: \"kubernetes.io/projected/66ef090a-4d9c-41f2-b225-a3bce5bb58c1-kube-api-access-n6twg\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-4srnb\" (UID: \"66ef090a-4d9c-41f2-b225-a3bce5bb58c1\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-4srnb" Oct 05 21:28:52 crc kubenswrapper[4754]: I1005 21:28:52.383885 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-4srnb" Oct 05 21:28:53 crc kubenswrapper[4754]: I1005 21:28:53.059885 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-4srnb"] Oct 05 21:28:53 crc kubenswrapper[4754]: I1005 21:28:53.967259 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-4srnb" event={"ID":"66ef090a-4d9c-41f2-b225-a3bce5bb58c1","Type":"ContainerStarted","Data":"835ec39964084b74c861462437932556044f55064b36a70d2fcd70d60327f52f"} Oct 05 21:28:53 crc kubenswrapper[4754]: I1005 21:28:53.967330 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-4srnb" event={"ID":"66ef090a-4d9c-41f2-b225-a3bce5bb58c1","Type":"ContainerStarted","Data":"524a362c2d86ca92c44a7a224bcbc6c0247bbea5f96e2b8d5c2805ca67e75d67"} Oct 05 21:28:53 crc kubenswrapper[4754]: I1005 21:28:53.994024 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-4srnb" podStartSLOduration=1.570141214 podStartE2EDuration="1.993996818s" podCreationTimestamp="2025-10-05 21:28:52 +0000 UTC" firstStartedPulling="2025-10-05 21:28:53.079328494 +0000 UTC m=+2056.983447244" lastFinishedPulling="2025-10-05 21:28:53.503184098 +0000 UTC m=+2057.407302848" observedRunningTime="2025-10-05 21:28:53.986995952 +0000 UTC m=+2057.891114682" watchObservedRunningTime="2025-10-05 21:28:53.993996818 +0000 UTC m=+2057.898115538" Oct 05 21:29:05 crc kubenswrapper[4754]: I1005 21:29:05.088563 4754 generic.go:334] "Generic (PLEG): container finished" podID="66ef090a-4d9c-41f2-b225-a3bce5bb58c1" containerID="835ec39964084b74c861462437932556044f55064b36a70d2fcd70d60327f52f" exitCode=0 Oct 05 21:29:05 crc kubenswrapper[4754]: I1005 21:29:05.089738 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-4srnb" event={"ID":"66ef090a-4d9c-41f2-b225-a3bce5bb58c1","Type":"ContainerDied","Data":"835ec39964084b74c861462437932556044f55064b36a70d2fcd70d60327f52f"} Oct 05 21:29:05 crc kubenswrapper[4754]: I1005 21:29:05.244740 4754 patch_prober.go:28] interesting pod/machine-config-daemon-b2h9k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 21:29:05 crc kubenswrapper[4754]: I1005 21:29:05.244835 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 21:29:05 crc kubenswrapper[4754]: I1005 21:29:05.244897 4754 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" Oct 05 21:29:05 crc kubenswrapper[4754]: I1005 21:29:05.246127 4754 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"fee409256f2bba9979dd69800f220c1257ab9bdb2e760315daae28edefc3f914"} pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 05 21:29:05 crc kubenswrapper[4754]: I1005 21:29:05.246217 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" containerName="machine-config-daemon" containerID="cri-o://fee409256f2bba9979dd69800f220c1257ab9bdb2e760315daae28edefc3f914" gracePeriod=600 Oct 05 21:29:06 crc kubenswrapper[4754]: I1005 21:29:06.102564 4754 generic.go:334] "Generic (PLEG): container finished" podID="be95a413-6a59-45b4-84b7-b43cae694a26" containerID="fee409256f2bba9979dd69800f220c1257ab9bdb2e760315daae28edefc3f914" exitCode=0 Oct 05 21:29:06 crc kubenswrapper[4754]: I1005 21:29:06.102665 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" event={"ID":"be95a413-6a59-45b4-84b7-b43cae694a26","Type":"ContainerDied","Data":"fee409256f2bba9979dd69800f220c1257ab9bdb2e760315daae28edefc3f914"} Oct 05 21:29:06 crc kubenswrapper[4754]: I1005 21:29:06.103615 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" event={"ID":"be95a413-6a59-45b4-84b7-b43cae694a26","Type":"ContainerStarted","Data":"0df70ace610d977d6937bca9a99a0d188993dfa6fcd8015008efc0870430fdb3"} Oct 05 21:29:06 crc kubenswrapper[4754]: I1005 21:29:06.103660 4754 scope.go:117] "RemoveContainer" containerID="e8573d2f52a359de8bfb0cb5d4d1a72d653442930ef95a60a1d6d2dd509ab1ac" Oct 05 21:29:06 crc kubenswrapper[4754]: I1005 21:29:06.626464 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-4srnb" Oct 05 21:29:06 crc kubenswrapper[4754]: I1005 21:29:06.785516 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/66ef090a-4d9c-41f2-b225-a3bce5bb58c1-ssh-key\") pod \"66ef090a-4d9c-41f2-b225-a3bce5bb58c1\" (UID: \"66ef090a-4d9c-41f2-b225-a3bce5bb58c1\") " Oct 05 21:29:06 crc kubenswrapper[4754]: I1005 21:29:06.786738 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n6twg\" (UniqueName: \"kubernetes.io/projected/66ef090a-4d9c-41f2-b225-a3bce5bb58c1-kube-api-access-n6twg\") pod \"66ef090a-4d9c-41f2-b225-a3bce5bb58c1\" (UID: \"66ef090a-4d9c-41f2-b225-a3bce5bb58c1\") " Oct 05 21:29:06 crc kubenswrapper[4754]: I1005 21:29:06.786883 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/66ef090a-4d9c-41f2-b225-a3bce5bb58c1-inventory\") pod \"66ef090a-4d9c-41f2-b225-a3bce5bb58c1\" (UID: \"66ef090a-4d9c-41f2-b225-a3bce5bb58c1\") " Oct 05 21:29:06 crc kubenswrapper[4754]: I1005 21:29:06.794349 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/66ef090a-4d9c-41f2-b225-a3bce5bb58c1-kube-api-access-n6twg" (OuterVolumeSpecName: "kube-api-access-n6twg") pod "66ef090a-4d9c-41f2-b225-a3bce5bb58c1" (UID: "66ef090a-4d9c-41f2-b225-a3bce5bb58c1"). InnerVolumeSpecName "kube-api-access-n6twg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:29:06 crc kubenswrapper[4754]: I1005 21:29:06.830697 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66ef090a-4d9c-41f2-b225-a3bce5bb58c1-inventory" (OuterVolumeSpecName: "inventory") pod "66ef090a-4d9c-41f2-b225-a3bce5bb58c1" (UID: "66ef090a-4d9c-41f2-b225-a3bce5bb58c1"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:29:06 crc kubenswrapper[4754]: I1005 21:29:06.834300 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66ef090a-4d9c-41f2-b225-a3bce5bb58c1-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "66ef090a-4d9c-41f2-b225-a3bce5bb58c1" (UID: "66ef090a-4d9c-41f2-b225-a3bce5bb58c1"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:29:06 crc kubenswrapper[4754]: I1005 21:29:06.891409 4754 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/66ef090a-4d9c-41f2-b225-a3bce5bb58c1-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 05 21:29:06 crc kubenswrapper[4754]: I1005 21:29:06.891570 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n6twg\" (UniqueName: \"kubernetes.io/projected/66ef090a-4d9c-41f2-b225-a3bce5bb58c1-kube-api-access-n6twg\") on node \"crc\" DevicePath \"\"" Oct 05 21:29:06 crc kubenswrapper[4754]: I1005 21:29:06.891610 4754 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/66ef090a-4d9c-41f2-b225-a3bce5bb58c1-inventory\") on node \"crc\" DevicePath \"\"" Oct 05 21:29:07 crc kubenswrapper[4754]: I1005 21:29:07.136590 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-4srnb" event={"ID":"66ef090a-4d9c-41f2-b225-a3bce5bb58c1","Type":"ContainerDied","Data":"524a362c2d86ca92c44a7a224bcbc6c0247bbea5f96e2b8d5c2805ca67e75d67"} Oct 05 21:29:07 crc kubenswrapper[4754]: I1005 21:29:07.137043 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="524a362c2d86ca92c44a7a224bcbc6c0247bbea5f96e2b8d5c2805ca67e75d67" Oct 05 21:29:07 crc kubenswrapper[4754]: I1005 21:29:07.136975 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-4srnb" Oct 05 21:29:07 crc kubenswrapper[4754]: I1005 21:29:07.226283 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ch8jw"] Oct 05 21:29:07 crc kubenswrapper[4754]: E1005 21:29:07.226786 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66ef090a-4d9c-41f2-b225-a3bce5bb58c1" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Oct 05 21:29:07 crc kubenswrapper[4754]: I1005 21:29:07.226809 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="66ef090a-4d9c-41f2-b225-a3bce5bb58c1" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Oct 05 21:29:07 crc kubenswrapper[4754]: I1005 21:29:07.227012 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="66ef090a-4d9c-41f2-b225-a3bce5bb58c1" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Oct 05 21:29:07 crc kubenswrapper[4754]: I1005 21:29:07.229408 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ch8jw" Oct 05 21:29:07 crc kubenswrapper[4754]: I1005 21:29:07.235209 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 05 21:29:07 crc kubenswrapper[4754]: I1005 21:29:07.236863 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Oct 05 21:29:07 crc kubenswrapper[4754]: I1005 21:29:07.237412 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-default-certs-0" Oct 05 21:29:07 crc kubenswrapper[4754]: I1005 21:29:07.238084 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Oct 05 21:29:07 crc kubenswrapper[4754]: I1005 21:29:07.238217 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-v54jd" Oct 05 21:29:07 crc kubenswrapper[4754]: I1005 21:29:07.238316 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 05 21:29:07 crc kubenswrapper[4754]: I1005 21:29:07.238425 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 05 21:29:07 crc kubenswrapper[4754]: I1005 21:29:07.238996 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Oct 05 21:29:07 crc kubenswrapper[4754]: I1005 21:29:07.252514 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ch8jw"] Oct 05 21:29:07 crc kubenswrapper[4754]: I1005 21:29:07.405397 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zjcwh\" (UniqueName: \"kubernetes.io/projected/27e32907-804b-43ba-ad6b-741e84eee1ab-kube-api-access-zjcwh\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ch8jw\" (UID: \"27e32907-804b-43ba-ad6b-741e84eee1ab\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ch8jw" Oct 05 21:29:07 crc kubenswrapper[4754]: I1005 21:29:07.408224 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27e32907-804b-43ba-ad6b-741e84eee1ab-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ch8jw\" (UID: \"27e32907-804b-43ba-ad6b-741e84eee1ab\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ch8jw" Oct 05 21:29:07 crc kubenswrapper[4754]: I1005 21:29:07.408323 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/27e32907-804b-43ba-ad6b-741e84eee1ab-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ch8jw\" (UID: \"27e32907-804b-43ba-ad6b-741e84eee1ab\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ch8jw" Oct 05 21:29:07 crc kubenswrapper[4754]: I1005 21:29:07.408383 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/27e32907-804b-43ba-ad6b-741e84eee1ab-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ch8jw\" (UID: \"27e32907-804b-43ba-ad6b-741e84eee1ab\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ch8jw" Oct 05 21:29:07 crc kubenswrapper[4754]: I1005 21:29:07.408521 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/27e32907-804b-43ba-ad6b-741e84eee1ab-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ch8jw\" (UID: \"27e32907-804b-43ba-ad6b-741e84eee1ab\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ch8jw" Oct 05 21:29:07 crc kubenswrapper[4754]: I1005 21:29:07.408669 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/27e32907-804b-43ba-ad6b-741e84eee1ab-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ch8jw\" (UID: \"27e32907-804b-43ba-ad6b-741e84eee1ab\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ch8jw" Oct 05 21:29:07 crc kubenswrapper[4754]: I1005 21:29:07.408725 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27e32907-804b-43ba-ad6b-741e84eee1ab-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ch8jw\" (UID: \"27e32907-804b-43ba-ad6b-741e84eee1ab\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ch8jw" Oct 05 21:29:07 crc kubenswrapper[4754]: I1005 21:29:07.408786 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27e32907-804b-43ba-ad6b-741e84eee1ab-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ch8jw\" (UID: \"27e32907-804b-43ba-ad6b-741e84eee1ab\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ch8jw" Oct 05 21:29:07 crc kubenswrapper[4754]: I1005 21:29:07.408868 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/27e32907-804b-43ba-ad6b-741e84eee1ab-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ch8jw\" (UID: \"27e32907-804b-43ba-ad6b-741e84eee1ab\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ch8jw" Oct 05 21:29:07 crc kubenswrapper[4754]: I1005 21:29:07.408970 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/27e32907-804b-43ba-ad6b-741e84eee1ab-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ch8jw\" (UID: \"27e32907-804b-43ba-ad6b-741e84eee1ab\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ch8jw" Oct 05 21:29:07 crc kubenswrapper[4754]: I1005 21:29:07.409030 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27e32907-804b-43ba-ad6b-741e84eee1ab-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ch8jw\" (UID: \"27e32907-804b-43ba-ad6b-741e84eee1ab\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ch8jw" Oct 05 21:29:07 crc kubenswrapper[4754]: I1005 21:29:07.409075 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27e32907-804b-43ba-ad6b-741e84eee1ab-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ch8jw\" (UID: \"27e32907-804b-43ba-ad6b-741e84eee1ab\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ch8jw" Oct 05 21:29:07 crc kubenswrapper[4754]: I1005 21:29:07.409210 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27e32907-804b-43ba-ad6b-741e84eee1ab-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ch8jw\" (UID: \"27e32907-804b-43ba-ad6b-741e84eee1ab\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ch8jw" Oct 05 21:29:07 crc kubenswrapper[4754]: I1005 21:29:07.409248 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27e32907-804b-43ba-ad6b-741e84eee1ab-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ch8jw\" (UID: \"27e32907-804b-43ba-ad6b-741e84eee1ab\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ch8jw" Oct 05 21:29:07 crc kubenswrapper[4754]: I1005 21:29:07.511758 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27e32907-804b-43ba-ad6b-741e84eee1ab-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ch8jw\" (UID: \"27e32907-804b-43ba-ad6b-741e84eee1ab\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ch8jw" Oct 05 21:29:07 crc kubenswrapper[4754]: I1005 21:29:07.511921 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/27e32907-804b-43ba-ad6b-741e84eee1ab-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ch8jw\" (UID: \"27e32907-804b-43ba-ad6b-741e84eee1ab\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ch8jw" Oct 05 21:29:07 crc kubenswrapper[4754]: I1005 21:29:07.512000 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/27e32907-804b-43ba-ad6b-741e84eee1ab-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ch8jw\" (UID: \"27e32907-804b-43ba-ad6b-741e84eee1ab\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ch8jw" Oct 05 21:29:07 crc kubenswrapper[4754]: I1005 21:29:07.512315 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27e32907-804b-43ba-ad6b-741e84eee1ab-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ch8jw\" (UID: \"27e32907-804b-43ba-ad6b-741e84eee1ab\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ch8jw" Oct 05 21:29:07 crc kubenswrapper[4754]: I1005 21:29:07.514092 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27e32907-804b-43ba-ad6b-741e84eee1ab-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ch8jw\" (UID: \"27e32907-804b-43ba-ad6b-741e84eee1ab\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ch8jw" Oct 05 21:29:07 crc kubenswrapper[4754]: I1005 21:29:07.514238 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27e32907-804b-43ba-ad6b-741e84eee1ab-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ch8jw\" (UID: \"27e32907-804b-43ba-ad6b-741e84eee1ab\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ch8jw" Oct 05 21:29:07 crc kubenswrapper[4754]: I1005 21:29:07.514363 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27e32907-804b-43ba-ad6b-741e84eee1ab-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ch8jw\" (UID: \"27e32907-804b-43ba-ad6b-741e84eee1ab\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ch8jw" Oct 05 21:29:07 crc kubenswrapper[4754]: I1005 21:29:07.514415 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zjcwh\" (UniqueName: \"kubernetes.io/projected/27e32907-804b-43ba-ad6b-741e84eee1ab-kube-api-access-zjcwh\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ch8jw\" (UID: \"27e32907-804b-43ba-ad6b-741e84eee1ab\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ch8jw" Oct 05 21:29:07 crc kubenswrapper[4754]: I1005 21:29:07.514633 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27e32907-804b-43ba-ad6b-741e84eee1ab-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ch8jw\" (UID: \"27e32907-804b-43ba-ad6b-741e84eee1ab\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ch8jw" Oct 05 21:29:07 crc kubenswrapper[4754]: I1005 21:29:07.514705 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/27e32907-804b-43ba-ad6b-741e84eee1ab-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ch8jw\" (UID: \"27e32907-804b-43ba-ad6b-741e84eee1ab\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ch8jw" Oct 05 21:29:07 crc kubenswrapper[4754]: I1005 21:29:07.514748 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/27e32907-804b-43ba-ad6b-741e84eee1ab-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ch8jw\" (UID: \"27e32907-804b-43ba-ad6b-741e84eee1ab\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ch8jw" Oct 05 21:29:07 crc kubenswrapper[4754]: I1005 21:29:07.514801 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/27e32907-804b-43ba-ad6b-741e84eee1ab-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ch8jw\" (UID: \"27e32907-804b-43ba-ad6b-741e84eee1ab\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ch8jw" Oct 05 21:29:07 crc kubenswrapper[4754]: I1005 21:29:07.514868 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/27e32907-804b-43ba-ad6b-741e84eee1ab-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ch8jw\" (UID: \"27e32907-804b-43ba-ad6b-741e84eee1ab\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ch8jw" Oct 05 21:29:07 crc kubenswrapper[4754]: I1005 21:29:07.514910 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27e32907-804b-43ba-ad6b-741e84eee1ab-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ch8jw\" (UID: \"27e32907-804b-43ba-ad6b-741e84eee1ab\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ch8jw" Oct 05 21:29:07 crc kubenswrapper[4754]: I1005 21:29:07.522940 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/27e32907-804b-43ba-ad6b-741e84eee1ab-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ch8jw\" (UID: \"27e32907-804b-43ba-ad6b-741e84eee1ab\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ch8jw" Oct 05 21:29:07 crc kubenswrapper[4754]: I1005 21:29:07.523341 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/27e32907-804b-43ba-ad6b-741e84eee1ab-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ch8jw\" (UID: \"27e32907-804b-43ba-ad6b-741e84eee1ab\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ch8jw" Oct 05 21:29:07 crc kubenswrapper[4754]: I1005 21:29:07.523766 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/27e32907-804b-43ba-ad6b-741e84eee1ab-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ch8jw\" (UID: \"27e32907-804b-43ba-ad6b-741e84eee1ab\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ch8jw" Oct 05 21:29:07 crc kubenswrapper[4754]: I1005 21:29:07.524618 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27e32907-804b-43ba-ad6b-741e84eee1ab-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ch8jw\" (UID: \"27e32907-804b-43ba-ad6b-741e84eee1ab\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ch8jw" Oct 05 21:29:07 crc kubenswrapper[4754]: I1005 21:29:07.525180 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27e32907-804b-43ba-ad6b-741e84eee1ab-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ch8jw\" (UID: \"27e32907-804b-43ba-ad6b-741e84eee1ab\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ch8jw" Oct 05 21:29:07 crc kubenswrapper[4754]: I1005 21:29:07.528727 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/27e32907-804b-43ba-ad6b-741e84eee1ab-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ch8jw\" (UID: \"27e32907-804b-43ba-ad6b-741e84eee1ab\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ch8jw" Oct 05 21:29:07 crc kubenswrapper[4754]: I1005 21:29:07.529304 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27e32907-804b-43ba-ad6b-741e84eee1ab-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ch8jw\" (UID: \"27e32907-804b-43ba-ad6b-741e84eee1ab\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ch8jw" Oct 05 21:29:07 crc kubenswrapper[4754]: I1005 21:29:07.530974 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/27e32907-804b-43ba-ad6b-741e84eee1ab-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ch8jw\" (UID: \"27e32907-804b-43ba-ad6b-741e84eee1ab\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ch8jw" Oct 05 21:29:07 crc kubenswrapper[4754]: I1005 21:29:07.531476 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27e32907-804b-43ba-ad6b-741e84eee1ab-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ch8jw\" (UID: \"27e32907-804b-43ba-ad6b-741e84eee1ab\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ch8jw" Oct 05 21:29:07 crc kubenswrapper[4754]: I1005 21:29:07.534557 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27e32907-804b-43ba-ad6b-741e84eee1ab-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ch8jw\" (UID: \"27e32907-804b-43ba-ad6b-741e84eee1ab\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ch8jw" Oct 05 21:29:07 crc kubenswrapper[4754]: I1005 21:29:07.537637 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/27e32907-804b-43ba-ad6b-741e84eee1ab-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ch8jw\" (UID: \"27e32907-804b-43ba-ad6b-741e84eee1ab\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ch8jw" Oct 05 21:29:07 crc kubenswrapper[4754]: I1005 21:29:07.540194 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27e32907-804b-43ba-ad6b-741e84eee1ab-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ch8jw\" (UID: \"27e32907-804b-43ba-ad6b-741e84eee1ab\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ch8jw" Oct 05 21:29:07 crc kubenswrapper[4754]: I1005 21:29:07.542941 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27e32907-804b-43ba-ad6b-741e84eee1ab-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ch8jw\" (UID: \"27e32907-804b-43ba-ad6b-741e84eee1ab\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ch8jw" Oct 05 21:29:07 crc kubenswrapper[4754]: I1005 21:29:07.543734 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zjcwh\" (UniqueName: \"kubernetes.io/projected/27e32907-804b-43ba-ad6b-741e84eee1ab-kube-api-access-zjcwh\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ch8jw\" (UID: \"27e32907-804b-43ba-ad6b-741e84eee1ab\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ch8jw" Oct 05 21:29:07 crc kubenswrapper[4754]: I1005 21:29:07.552854 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ch8jw" Oct 05 21:29:08 crc kubenswrapper[4754]: I1005 21:29:08.170268 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ch8jw"] Oct 05 21:29:08 crc kubenswrapper[4754]: W1005 21:29:08.172407 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod27e32907_804b_43ba_ad6b_741e84eee1ab.slice/crio-77845ac9c4621b26b95e9a0871937920df2709bdbf96c9f9f100cfb5134d90de WatchSource:0}: Error finding container 77845ac9c4621b26b95e9a0871937920df2709bdbf96c9f9f100cfb5134d90de: Status 404 returned error can't find the container with id 77845ac9c4621b26b95e9a0871937920df2709bdbf96c9f9f100cfb5134d90de Oct 05 21:29:09 crc kubenswrapper[4754]: I1005 21:29:09.157550 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ch8jw" event={"ID":"27e32907-804b-43ba-ad6b-741e84eee1ab","Type":"ContainerStarted","Data":"14ca46829d4409abb8e2c1b8cda227a9379ee297145eda876786044547d36fd0"} Oct 05 21:29:09 crc kubenswrapper[4754]: I1005 21:29:09.158088 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ch8jw" event={"ID":"27e32907-804b-43ba-ad6b-741e84eee1ab","Type":"ContainerStarted","Data":"77845ac9c4621b26b95e9a0871937920df2709bdbf96c9f9f100cfb5134d90de"} Oct 05 21:29:09 crc kubenswrapper[4754]: I1005 21:29:09.190251 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ch8jw" podStartSLOduration=1.8173357 podStartE2EDuration="2.190219642s" podCreationTimestamp="2025-10-05 21:29:07 +0000 UTC" firstStartedPulling="2025-10-05 21:29:08.178800972 +0000 UTC m=+2072.082919682" lastFinishedPulling="2025-10-05 21:29:08.551684914 +0000 UTC m=+2072.455803624" observedRunningTime="2025-10-05 21:29:09.177264019 +0000 UTC m=+2073.081382739" watchObservedRunningTime="2025-10-05 21:29:09.190219642 +0000 UTC m=+2073.094338392" Oct 05 21:29:57 crc kubenswrapper[4754]: I1005 21:29:57.692681 4754 generic.go:334] "Generic (PLEG): container finished" podID="27e32907-804b-43ba-ad6b-741e84eee1ab" containerID="14ca46829d4409abb8e2c1b8cda227a9379ee297145eda876786044547d36fd0" exitCode=0 Oct 05 21:29:57 crc kubenswrapper[4754]: I1005 21:29:57.692820 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ch8jw" event={"ID":"27e32907-804b-43ba-ad6b-741e84eee1ab","Type":"ContainerDied","Data":"14ca46829d4409abb8e2c1b8cda227a9379ee297145eda876786044547d36fd0"} Oct 05 21:29:59 crc kubenswrapper[4754]: I1005 21:29:59.257248 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ch8jw" Oct 05 21:29:59 crc kubenswrapper[4754]: I1005 21:29:59.307099 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27e32907-804b-43ba-ad6b-741e84eee1ab-repo-setup-combined-ca-bundle\") pod \"27e32907-804b-43ba-ad6b-741e84eee1ab\" (UID: \"27e32907-804b-43ba-ad6b-741e84eee1ab\") " Oct 05 21:29:59 crc kubenswrapper[4754]: I1005 21:29:59.307218 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/27e32907-804b-43ba-ad6b-741e84eee1ab-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"27e32907-804b-43ba-ad6b-741e84eee1ab\" (UID: \"27e32907-804b-43ba-ad6b-741e84eee1ab\") " Oct 05 21:29:59 crc kubenswrapper[4754]: I1005 21:29:59.307258 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27e32907-804b-43ba-ad6b-741e84eee1ab-libvirt-combined-ca-bundle\") pod \"27e32907-804b-43ba-ad6b-741e84eee1ab\" (UID: \"27e32907-804b-43ba-ad6b-741e84eee1ab\") " Oct 05 21:29:59 crc kubenswrapper[4754]: I1005 21:29:59.307281 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27e32907-804b-43ba-ad6b-741e84eee1ab-nova-combined-ca-bundle\") pod \"27e32907-804b-43ba-ad6b-741e84eee1ab\" (UID: \"27e32907-804b-43ba-ad6b-741e84eee1ab\") " Oct 05 21:29:59 crc kubenswrapper[4754]: I1005 21:29:59.307339 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zjcwh\" (UniqueName: \"kubernetes.io/projected/27e32907-804b-43ba-ad6b-741e84eee1ab-kube-api-access-zjcwh\") pod \"27e32907-804b-43ba-ad6b-741e84eee1ab\" (UID: \"27e32907-804b-43ba-ad6b-741e84eee1ab\") " Oct 05 21:29:59 crc kubenswrapper[4754]: I1005 21:29:59.307391 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27e32907-804b-43ba-ad6b-741e84eee1ab-neutron-metadata-combined-ca-bundle\") pod \"27e32907-804b-43ba-ad6b-741e84eee1ab\" (UID: \"27e32907-804b-43ba-ad6b-741e84eee1ab\") " Oct 05 21:29:59 crc kubenswrapper[4754]: I1005 21:29:59.307539 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/27e32907-804b-43ba-ad6b-741e84eee1ab-inventory\") pod \"27e32907-804b-43ba-ad6b-741e84eee1ab\" (UID: \"27e32907-804b-43ba-ad6b-741e84eee1ab\") " Oct 05 21:29:59 crc kubenswrapper[4754]: I1005 21:29:59.307634 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27e32907-804b-43ba-ad6b-741e84eee1ab-telemetry-combined-ca-bundle\") pod \"27e32907-804b-43ba-ad6b-741e84eee1ab\" (UID: \"27e32907-804b-43ba-ad6b-741e84eee1ab\") " Oct 05 21:29:59 crc kubenswrapper[4754]: I1005 21:29:59.307690 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/27e32907-804b-43ba-ad6b-741e84eee1ab-ssh-key\") pod \"27e32907-804b-43ba-ad6b-741e84eee1ab\" (UID: \"27e32907-804b-43ba-ad6b-741e84eee1ab\") " Oct 05 21:29:59 crc kubenswrapper[4754]: I1005 21:29:59.307751 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/27e32907-804b-43ba-ad6b-741e84eee1ab-openstack-edpm-ipam-ovn-default-certs-0\") pod \"27e32907-804b-43ba-ad6b-741e84eee1ab\" (UID: \"27e32907-804b-43ba-ad6b-741e84eee1ab\") " Oct 05 21:29:59 crc kubenswrapper[4754]: I1005 21:29:59.307811 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27e32907-804b-43ba-ad6b-741e84eee1ab-ovn-combined-ca-bundle\") pod \"27e32907-804b-43ba-ad6b-741e84eee1ab\" (UID: \"27e32907-804b-43ba-ad6b-741e84eee1ab\") " Oct 05 21:29:59 crc kubenswrapper[4754]: I1005 21:29:59.307847 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27e32907-804b-43ba-ad6b-741e84eee1ab-bootstrap-combined-ca-bundle\") pod \"27e32907-804b-43ba-ad6b-741e84eee1ab\" (UID: \"27e32907-804b-43ba-ad6b-741e84eee1ab\") " Oct 05 21:29:59 crc kubenswrapper[4754]: I1005 21:29:59.307909 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/27e32907-804b-43ba-ad6b-741e84eee1ab-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"27e32907-804b-43ba-ad6b-741e84eee1ab\" (UID: \"27e32907-804b-43ba-ad6b-741e84eee1ab\") " Oct 05 21:29:59 crc kubenswrapper[4754]: I1005 21:29:59.307959 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/27e32907-804b-43ba-ad6b-741e84eee1ab-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"27e32907-804b-43ba-ad6b-741e84eee1ab\" (UID: \"27e32907-804b-43ba-ad6b-741e84eee1ab\") " Oct 05 21:29:59 crc kubenswrapper[4754]: I1005 21:29:59.317162 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/27e32907-804b-43ba-ad6b-741e84eee1ab-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "27e32907-804b-43ba-ad6b-741e84eee1ab" (UID: "27e32907-804b-43ba-ad6b-741e84eee1ab"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:29:59 crc kubenswrapper[4754]: I1005 21:29:59.330318 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27e32907-804b-43ba-ad6b-741e84eee1ab-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "27e32907-804b-43ba-ad6b-741e84eee1ab" (UID: "27e32907-804b-43ba-ad6b-741e84eee1ab"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:29:59 crc kubenswrapper[4754]: I1005 21:29:59.330429 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27e32907-804b-43ba-ad6b-741e84eee1ab-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "27e32907-804b-43ba-ad6b-741e84eee1ab" (UID: "27e32907-804b-43ba-ad6b-741e84eee1ab"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:29:59 crc kubenswrapper[4754]: I1005 21:29:59.330916 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27e32907-804b-43ba-ad6b-741e84eee1ab-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "27e32907-804b-43ba-ad6b-741e84eee1ab" (UID: "27e32907-804b-43ba-ad6b-741e84eee1ab"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:29:59 crc kubenswrapper[4754]: I1005 21:29:59.331323 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/27e32907-804b-43ba-ad6b-741e84eee1ab-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "27e32907-804b-43ba-ad6b-741e84eee1ab" (UID: "27e32907-804b-43ba-ad6b-741e84eee1ab"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:29:59 crc kubenswrapper[4754]: I1005 21:29:59.334870 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27e32907-804b-43ba-ad6b-741e84eee1ab-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "27e32907-804b-43ba-ad6b-741e84eee1ab" (UID: "27e32907-804b-43ba-ad6b-741e84eee1ab"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:29:59 crc kubenswrapper[4754]: I1005 21:29:59.335018 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/27e32907-804b-43ba-ad6b-741e84eee1ab-kube-api-access-zjcwh" (OuterVolumeSpecName: "kube-api-access-zjcwh") pod "27e32907-804b-43ba-ad6b-741e84eee1ab" (UID: "27e32907-804b-43ba-ad6b-741e84eee1ab"). InnerVolumeSpecName "kube-api-access-zjcwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:29:59 crc kubenswrapper[4754]: I1005 21:29:59.336216 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27e32907-804b-43ba-ad6b-741e84eee1ab-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "27e32907-804b-43ba-ad6b-741e84eee1ab" (UID: "27e32907-804b-43ba-ad6b-741e84eee1ab"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:29:59 crc kubenswrapper[4754]: I1005 21:29:59.336958 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/27e32907-804b-43ba-ad6b-741e84eee1ab-openstack-edpm-ipam-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-default-certs-0") pod "27e32907-804b-43ba-ad6b-741e84eee1ab" (UID: "27e32907-804b-43ba-ad6b-741e84eee1ab"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:29:59 crc kubenswrapper[4754]: I1005 21:29:59.344247 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/27e32907-804b-43ba-ad6b-741e84eee1ab-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "27e32907-804b-43ba-ad6b-741e84eee1ab" (UID: "27e32907-804b-43ba-ad6b-741e84eee1ab"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:29:59 crc kubenswrapper[4754]: I1005 21:29:59.345166 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27e32907-804b-43ba-ad6b-741e84eee1ab-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "27e32907-804b-43ba-ad6b-741e84eee1ab" (UID: "27e32907-804b-43ba-ad6b-741e84eee1ab"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:29:59 crc kubenswrapper[4754]: I1005 21:29:59.345403 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27e32907-804b-43ba-ad6b-741e84eee1ab-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "27e32907-804b-43ba-ad6b-741e84eee1ab" (UID: "27e32907-804b-43ba-ad6b-741e84eee1ab"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:29:59 crc kubenswrapper[4754]: I1005 21:29:59.357371 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27e32907-804b-43ba-ad6b-741e84eee1ab-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "27e32907-804b-43ba-ad6b-741e84eee1ab" (UID: "27e32907-804b-43ba-ad6b-741e84eee1ab"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:29:59 crc kubenswrapper[4754]: I1005 21:29:59.360423 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27e32907-804b-43ba-ad6b-741e84eee1ab-inventory" (OuterVolumeSpecName: "inventory") pod "27e32907-804b-43ba-ad6b-741e84eee1ab" (UID: "27e32907-804b-43ba-ad6b-741e84eee1ab"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:29:59 crc kubenswrapper[4754]: I1005 21:29:59.409944 4754 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/27e32907-804b-43ba-ad6b-741e84eee1ab-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 05 21:29:59 crc kubenswrapper[4754]: I1005 21:29:59.410013 4754 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/27e32907-804b-43ba-ad6b-741e84eee1ab-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 05 21:29:59 crc kubenswrapper[4754]: I1005 21:29:59.410438 4754 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27e32907-804b-43ba-ad6b-741e84eee1ab-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 21:29:59 crc kubenswrapper[4754]: I1005 21:29:59.410458 4754 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27e32907-804b-43ba-ad6b-741e84eee1ab-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 21:29:59 crc kubenswrapper[4754]: I1005 21:29:59.410479 4754 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/27e32907-804b-43ba-ad6b-741e84eee1ab-openstack-edpm-ipam-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 05 21:29:59 crc kubenswrapper[4754]: I1005 21:29:59.410566 4754 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/27e32907-804b-43ba-ad6b-741e84eee1ab-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 05 21:29:59 crc kubenswrapper[4754]: I1005 21:29:59.410581 4754 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27e32907-804b-43ba-ad6b-741e84eee1ab-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 21:29:59 crc kubenswrapper[4754]: I1005 21:29:59.410599 4754 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/27e32907-804b-43ba-ad6b-741e84eee1ab-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 05 21:29:59 crc kubenswrapper[4754]: I1005 21:29:59.410615 4754 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27e32907-804b-43ba-ad6b-741e84eee1ab-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 21:29:59 crc kubenswrapper[4754]: I1005 21:29:59.410628 4754 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27e32907-804b-43ba-ad6b-741e84eee1ab-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 21:29:59 crc kubenswrapper[4754]: I1005 21:29:59.410642 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zjcwh\" (UniqueName: \"kubernetes.io/projected/27e32907-804b-43ba-ad6b-741e84eee1ab-kube-api-access-zjcwh\") on node \"crc\" DevicePath \"\"" Oct 05 21:29:59 crc kubenswrapper[4754]: I1005 21:29:59.410656 4754 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27e32907-804b-43ba-ad6b-741e84eee1ab-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 21:29:59 crc kubenswrapper[4754]: I1005 21:29:59.410670 4754 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/27e32907-804b-43ba-ad6b-741e84eee1ab-inventory\") on node \"crc\" DevicePath \"\"" Oct 05 21:29:59 crc kubenswrapper[4754]: I1005 21:29:59.410683 4754 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27e32907-804b-43ba-ad6b-741e84eee1ab-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 21:29:59 crc kubenswrapper[4754]: I1005 21:29:59.731557 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ch8jw" event={"ID":"27e32907-804b-43ba-ad6b-741e84eee1ab","Type":"ContainerDied","Data":"77845ac9c4621b26b95e9a0871937920df2709bdbf96c9f9f100cfb5134d90de"} Oct 05 21:29:59 crc kubenswrapper[4754]: I1005 21:29:59.731630 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="77845ac9c4621b26b95e9a0871937920df2709bdbf96c9f9f100cfb5134d90de" Oct 05 21:29:59 crc kubenswrapper[4754]: I1005 21:29:59.731648 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ch8jw" Oct 05 21:29:59 crc kubenswrapper[4754]: I1005 21:29:59.880067 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-cwth8"] Oct 05 21:29:59 crc kubenswrapper[4754]: E1005 21:29:59.880830 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27e32907-804b-43ba-ad6b-741e84eee1ab" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Oct 05 21:29:59 crc kubenswrapper[4754]: I1005 21:29:59.880945 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="27e32907-804b-43ba-ad6b-741e84eee1ab" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Oct 05 21:29:59 crc kubenswrapper[4754]: I1005 21:29:59.881230 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="27e32907-804b-43ba-ad6b-741e84eee1ab" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Oct 05 21:29:59 crc kubenswrapper[4754]: I1005 21:29:59.882231 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-cwth8" Oct 05 21:29:59 crc kubenswrapper[4754]: I1005 21:29:59.885301 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 05 21:29:59 crc kubenswrapper[4754]: I1005 21:29:59.885692 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-v54jd" Oct 05 21:29:59 crc kubenswrapper[4754]: I1005 21:29:59.887522 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 05 21:29:59 crc kubenswrapper[4754]: I1005 21:29:59.887710 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Oct 05 21:29:59 crc kubenswrapper[4754]: I1005 21:29:59.893757 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 05 21:29:59 crc kubenswrapper[4754]: I1005 21:29:59.924259 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c26e6366-e15d-4c7b-a370-a601e2cee56f-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-cwth8\" (UID: \"c26e6366-e15d-4c7b-a370-a601e2cee56f\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-cwth8" Oct 05 21:29:59 crc kubenswrapper[4754]: I1005 21:29:59.924653 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qxk2d\" (UniqueName: \"kubernetes.io/projected/c26e6366-e15d-4c7b-a370-a601e2cee56f-kube-api-access-qxk2d\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-cwth8\" (UID: \"c26e6366-e15d-4c7b-a370-a601e2cee56f\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-cwth8" Oct 05 21:29:59 crc kubenswrapper[4754]: I1005 21:29:59.924899 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c26e6366-e15d-4c7b-a370-a601e2cee56f-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-cwth8\" (UID: \"c26e6366-e15d-4c7b-a370-a601e2cee56f\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-cwth8" Oct 05 21:29:59 crc kubenswrapper[4754]: I1005 21:29:59.925077 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/c26e6366-e15d-4c7b-a370-a601e2cee56f-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-cwth8\" (UID: \"c26e6366-e15d-4c7b-a370-a601e2cee56f\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-cwth8" Oct 05 21:29:59 crc kubenswrapper[4754]: I1005 21:29:59.925148 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c26e6366-e15d-4c7b-a370-a601e2cee56f-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-cwth8\" (UID: \"c26e6366-e15d-4c7b-a370-a601e2cee56f\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-cwth8" Oct 05 21:29:59 crc kubenswrapper[4754]: I1005 21:29:59.932195 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-cwth8"] Oct 05 21:30:00 crc kubenswrapper[4754]: I1005 21:30:00.026801 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c26e6366-e15d-4c7b-a370-a601e2cee56f-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-cwth8\" (UID: \"c26e6366-e15d-4c7b-a370-a601e2cee56f\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-cwth8" Oct 05 21:30:00 crc kubenswrapper[4754]: I1005 21:30:00.027176 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/c26e6366-e15d-4c7b-a370-a601e2cee56f-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-cwth8\" (UID: \"c26e6366-e15d-4c7b-a370-a601e2cee56f\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-cwth8" Oct 05 21:30:00 crc kubenswrapper[4754]: I1005 21:30:00.027211 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c26e6366-e15d-4c7b-a370-a601e2cee56f-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-cwth8\" (UID: \"c26e6366-e15d-4c7b-a370-a601e2cee56f\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-cwth8" Oct 05 21:30:00 crc kubenswrapper[4754]: I1005 21:30:00.027252 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c26e6366-e15d-4c7b-a370-a601e2cee56f-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-cwth8\" (UID: \"c26e6366-e15d-4c7b-a370-a601e2cee56f\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-cwth8" Oct 05 21:30:00 crc kubenswrapper[4754]: I1005 21:30:00.027314 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qxk2d\" (UniqueName: \"kubernetes.io/projected/c26e6366-e15d-4c7b-a370-a601e2cee56f-kube-api-access-qxk2d\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-cwth8\" (UID: \"c26e6366-e15d-4c7b-a370-a601e2cee56f\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-cwth8" Oct 05 21:30:00 crc kubenswrapper[4754]: I1005 21:30:00.028363 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/c26e6366-e15d-4c7b-a370-a601e2cee56f-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-cwth8\" (UID: \"c26e6366-e15d-4c7b-a370-a601e2cee56f\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-cwth8" Oct 05 21:30:00 crc kubenswrapper[4754]: I1005 21:30:00.034062 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c26e6366-e15d-4c7b-a370-a601e2cee56f-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-cwth8\" (UID: \"c26e6366-e15d-4c7b-a370-a601e2cee56f\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-cwth8" Oct 05 21:30:00 crc kubenswrapper[4754]: I1005 21:30:00.034341 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c26e6366-e15d-4c7b-a370-a601e2cee56f-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-cwth8\" (UID: \"c26e6366-e15d-4c7b-a370-a601e2cee56f\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-cwth8" Oct 05 21:30:00 crc kubenswrapper[4754]: I1005 21:30:00.035965 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c26e6366-e15d-4c7b-a370-a601e2cee56f-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-cwth8\" (UID: \"c26e6366-e15d-4c7b-a370-a601e2cee56f\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-cwth8" Oct 05 21:30:00 crc kubenswrapper[4754]: I1005 21:30:00.044781 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qxk2d\" (UniqueName: \"kubernetes.io/projected/c26e6366-e15d-4c7b-a370-a601e2cee56f-kube-api-access-qxk2d\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-cwth8\" (UID: \"c26e6366-e15d-4c7b-a370-a601e2cee56f\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-cwth8" Oct 05 21:30:00 crc kubenswrapper[4754]: I1005 21:30:00.146795 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29328330-b7xbb"] Oct 05 21:30:00 crc kubenswrapper[4754]: I1005 21:30:00.148677 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29328330-b7xbb" Oct 05 21:30:00 crc kubenswrapper[4754]: I1005 21:30:00.151887 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 05 21:30:00 crc kubenswrapper[4754]: I1005 21:30:00.152614 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 05 21:30:00 crc kubenswrapper[4754]: I1005 21:30:00.160695 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29328330-b7xbb"] Oct 05 21:30:00 crc kubenswrapper[4754]: I1005 21:30:00.206647 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-cwth8" Oct 05 21:30:00 crc kubenswrapper[4754]: I1005 21:30:00.232972 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7h6f7\" (UniqueName: \"kubernetes.io/projected/8d619d93-19a1-4ecb-8008-64b1d4dc931a-kube-api-access-7h6f7\") pod \"collect-profiles-29328330-b7xbb\" (UID: \"8d619d93-19a1-4ecb-8008-64b1d4dc931a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29328330-b7xbb" Oct 05 21:30:00 crc kubenswrapper[4754]: I1005 21:30:00.233132 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8d619d93-19a1-4ecb-8008-64b1d4dc931a-config-volume\") pod \"collect-profiles-29328330-b7xbb\" (UID: \"8d619d93-19a1-4ecb-8008-64b1d4dc931a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29328330-b7xbb" Oct 05 21:30:00 crc kubenswrapper[4754]: I1005 21:30:00.233174 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8d619d93-19a1-4ecb-8008-64b1d4dc931a-secret-volume\") pod \"collect-profiles-29328330-b7xbb\" (UID: \"8d619d93-19a1-4ecb-8008-64b1d4dc931a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29328330-b7xbb" Oct 05 21:30:00 crc kubenswrapper[4754]: I1005 21:30:00.337443 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8d619d93-19a1-4ecb-8008-64b1d4dc931a-config-volume\") pod \"collect-profiles-29328330-b7xbb\" (UID: \"8d619d93-19a1-4ecb-8008-64b1d4dc931a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29328330-b7xbb" Oct 05 21:30:00 crc kubenswrapper[4754]: I1005 21:30:00.337539 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8d619d93-19a1-4ecb-8008-64b1d4dc931a-secret-volume\") pod \"collect-profiles-29328330-b7xbb\" (UID: \"8d619d93-19a1-4ecb-8008-64b1d4dc931a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29328330-b7xbb" Oct 05 21:30:00 crc kubenswrapper[4754]: I1005 21:30:00.337599 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7h6f7\" (UniqueName: \"kubernetes.io/projected/8d619d93-19a1-4ecb-8008-64b1d4dc931a-kube-api-access-7h6f7\") pod \"collect-profiles-29328330-b7xbb\" (UID: \"8d619d93-19a1-4ecb-8008-64b1d4dc931a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29328330-b7xbb" Oct 05 21:30:00 crc kubenswrapper[4754]: I1005 21:30:00.340151 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8d619d93-19a1-4ecb-8008-64b1d4dc931a-config-volume\") pod \"collect-profiles-29328330-b7xbb\" (UID: \"8d619d93-19a1-4ecb-8008-64b1d4dc931a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29328330-b7xbb" Oct 05 21:30:00 crc kubenswrapper[4754]: I1005 21:30:00.360344 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7h6f7\" (UniqueName: \"kubernetes.io/projected/8d619d93-19a1-4ecb-8008-64b1d4dc931a-kube-api-access-7h6f7\") pod \"collect-profiles-29328330-b7xbb\" (UID: \"8d619d93-19a1-4ecb-8008-64b1d4dc931a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29328330-b7xbb" Oct 05 21:30:00 crc kubenswrapper[4754]: I1005 21:30:00.361970 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8d619d93-19a1-4ecb-8008-64b1d4dc931a-secret-volume\") pod \"collect-profiles-29328330-b7xbb\" (UID: \"8d619d93-19a1-4ecb-8008-64b1d4dc931a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29328330-b7xbb" Oct 05 21:30:00 crc kubenswrapper[4754]: I1005 21:30:00.528449 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29328330-b7xbb" Oct 05 21:30:00 crc kubenswrapper[4754]: I1005 21:30:00.852978 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-cwth8"] Oct 05 21:30:01 crc kubenswrapper[4754]: I1005 21:30:01.006455 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29328330-b7xbb"] Oct 05 21:30:01 crc kubenswrapper[4754]: W1005 21:30:01.007184 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8d619d93_19a1_4ecb_8008_64b1d4dc931a.slice/crio-252de5485ca4e87b728d456981c41808ccc3b8bf7f4689232f04576104823762 WatchSource:0}: Error finding container 252de5485ca4e87b728d456981c41808ccc3b8bf7f4689232f04576104823762: Status 404 returned error can't find the container with id 252de5485ca4e87b728d456981c41808ccc3b8bf7f4689232f04576104823762 Oct 05 21:30:01 crc kubenswrapper[4754]: I1005 21:30:01.768143 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-cwth8" event={"ID":"c26e6366-e15d-4c7b-a370-a601e2cee56f","Type":"ContainerStarted","Data":"92d6390e8d2fadbf9896fab63fe584db05fca63a9ca44a70123ae4bfe54d6503"} Oct 05 21:30:01 crc kubenswrapper[4754]: I1005 21:30:01.769637 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-cwth8" event={"ID":"c26e6366-e15d-4c7b-a370-a601e2cee56f","Type":"ContainerStarted","Data":"f375256c7648bf626d64c7750878170fee5c47f4df7468df0671f9591987b728"} Oct 05 21:30:01 crc kubenswrapper[4754]: I1005 21:30:01.774788 4754 generic.go:334] "Generic (PLEG): container finished" podID="8d619d93-19a1-4ecb-8008-64b1d4dc931a" containerID="b85dbda630293941d96625b3609eacbc989e66ec844fb6f26d28401d7f031a57" exitCode=0 Oct 05 21:30:01 crc kubenswrapper[4754]: I1005 21:30:01.774856 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29328330-b7xbb" event={"ID":"8d619d93-19a1-4ecb-8008-64b1d4dc931a","Type":"ContainerDied","Data":"b85dbda630293941d96625b3609eacbc989e66ec844fb6f26d28401d7f031a57"} Oct 05 21:30:01 crc kubenswrapper[4754]: I1005 21:30:01.774883 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29328330-b7xbb" event={"ID":"8d619d93-19a1-4ecb-8008-64b1d4dc931a","Type":"ContainerStarted","Data":"252de5485ca4e87b728d456981c41808ccc3b8bf7f4689232f04576104823762"} Oct 05 21:30:01 crc kubenswrapper[4754]: I1005 21:30:01.788040 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-cwth8" podStartSLOduration=2.192794302 podStartE2EDuration="2.788027082s" podCreationTimestamp="2025-10-05 21:29:59 +0000 UTC" firstStartedPulling="2025-10-05 21:30:00.858386961 +0000 UTC m=+2124.762505671" lastFinishedPulling="2025-10-05 21:30:01.453619741 +0000 UTC m=+2125.357738451" observedRunningTime="2025-10-05 21:30:01.786172373 +0000 UTC m=+2125.690291083" watchObservedRunningTime="2025-10-05 21:30:01.788027082 +0000 UTC m=+2125.692145802" Oct 05 21:30:03 crc kubenswrapper[4754]: I1005 21:30:03.165822 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29328330-b7xbb" Oct 05 21:30:03 crc kubenswrapper[4754]: I1005 21:30:03.223641 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7h6f7\" (UniqueName: \"kubernetes.io/projected/8d619d93-19a1-4ecb-8008-64b1d4dc931a-kube-api-access-7h6f7\") pod \"8d619d93-19a1-4ecb-8008-64b1d4dc931a\" (UID: \"8d619d93-19a1-4ecb-8008-64b1d4dc931a\") " Oct 05 21:30:03 crc kubenswrapper[4754]: I1005 21:30:03.223725 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8d619d93-19a1-4ecb-8008-64b1d4dc931a-config-volume\") pod \"8d619d93-19a1-4ecb-8008-64b1d4dc931a\" (UID: \"8d619d93-19a1-4ecb-8008-64b1d4dc931a\") " Oct 05 21:30:03 crc kubenswrapper[4754]: I1005 21:30:03.223867 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8d619d93-19a1-4ecb-8008-64b1d4dc931a-secret-volume\") pod \"8d619d93-19a1-4ecb-8008-64b1d4dc931a\" (UID: \"8d619d93-19a1-4ecb-8008-64b1d4dc931a\") " Oct 05 21:30:03 crc kubenswrapper[4754]: I1005 21:30:03.225681 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8d619d93-19a1-4ecb-8008-64b1d4dc931a-config-volume" (OuterVolumeSpecName: "config-volume") pod "8d619d93-19a1-4ecb-8008-64b1d4dc931a" (UID: "8d619d93-19a1-4ecb-8008-64b1d4dc931a"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 21:30:03 crc kubenswrapper[4754]: I1005 21:30:03.232958 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d619d93-19a1-4ecb-8008-64b1d4dc931a-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "8d619d93-19a1-4ecb-8008-64b1d4dc931a" (UID: "8d619d93-19a1-4ecb-8008-64b1d4dc931a"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:30:03 crc kubenswrapper[4754]: I1005 21:30:03.233642 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8d619d93-19a1-4ecb-8008-64b1d4dc931a-kube-api-access-7h6f7" (OuterVolumeSpecName: "kube-api-access-7h6f7") pod "8d619d93-19a1-4ecb-8008-64b1d4dc931a" (UID: "8d619d93-19a1-4ecb-8008-64b1d4dc931a"). InnerVolumeSpecName "kube-api-access-7h6f7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:30:03 crc kubenswrapper[4754]: I1005 21:30:03.327390 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7h6f7\" (UniqueName: \"kubernetes.io/projected/8d619d93-19a1-4ecb-8008-64b1d4dc931a-kube-api-access-7h6f7\") on node \"crc\" DevicePath \"\"" Oct 05 21:30:03 crc kubenswrapper[4754]: I1005 21:30:03.327481 4754 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8d619d93-19a1-4ecb-8008-64b1d4dc931a-config-volume\") on node \"crc\" DevicePath \"\"" Oct 05 21:30:03 crc kubenswrapper[4754]: I1005 21:30:03.327506 4754 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8d619d93-19a1-4ecb-8008-64b1d4dc931a-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 05 21:30:03 crc kubenswrapper[4754]: I1005 21:30:03.808906 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29328330-b7xbb" event={"ID":"8d619d93-19a1-4ecb-8008-64b1d4dc931a","Type":"ContainerDied","Data":"252de5485ca4e87b728d456981c41808ccc3b8bf7f4689232f04576104823762"} Oct 05 21:30:03 crc kubenswrapper[4754]: I1005 21:30:03.808980 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="252de5485ca4e87b728d456981c41808ccc3b8bf7f4689232f04576104823762" Oct 05 21:30:03 crc kubenswrapper[4754]: I1005 21:30:03.808999 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29328330-b7xbb" Oct 05 21:30:04 crc kubenswrapper[4754]: I1005 21:30:04.248184 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29328285-mrswx"] Oct 05 21:30:04 crc kubenswrapper[4754]: I1005 21:30:04.255988 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29328285-mrswx"] Oct 05 21:30:04 crc kubenswrapper[4754]: I1005 21:30:04.856897 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="60ff4c5c-9e45-47d3-b45d-a365e988342e" path="/var/lib/kubelet/pods/60ff4c5c-9e45-47d3-b45d-a365e988342e/volumes" Oct 05 21:30:06 crc kubenswrapper[4754]: I1005 21:30:06.570036 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-4mw5m"] Oct 05 21:30:06 crc kubenswrapper[4754]: E1005 21:30:06.573582 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d619d93-19a1-4ecb-8008-64b1d4dc931a" containerName="collect-profiles" Oct 05 21:30:06 crc kubenswrapper[4754]: I1005 21:30:06.573678 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d619d93-19a1-4ecb-8008-64b1d4dc931a" containerName="collect-profiles" Oct 05 21:30:06 crc kubenswrapper[4754]: I1005 21:30:06.573983 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d619d93-19a1-4ecb-8008-64b1d4dc931a" containerName="collect-profiles" Oct 05 21:30:06 crc kubenswrapper[4754]: I1005 21:30:06.576306 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4mw5m" Oct 05 21:30:06 crc kubenswrapper[4754]: I1005 21:30:06.594670 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4mw5m"] Oct 05 21:30:06 crc kubenswrapper[4754]: I1005 21:30:06.724421 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/afa6a6bd-d9cf-4379-b4ed-452bff7149ae-catalog-content\") pod \"redhat-marketplace-4mw5m\" (UID: \"afa6a6bd-d9cf-4379-b4ed-452bff7149ae\") " pod="openshift-marketplace/redhat-marketplace-4mw5m" Oct 05 21:30:06 crc kubenswrapper[4754]: I1005 21:30:06.724676 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/afa6a6bd-d9cf-4379-b4ed-452bff7149ae-utilities\") pod \"redhat-marketplace-4mw5m\" (UID: \"afa6a6bd-d9cf-4379-b4ed-452bff7149ae\") " pod="openshift-marketplace/redhat-marketplace-4mw5m" Oct 05 21:30:06 crc kubenswrapper[4754]: I1005 21:30:06.724760 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k7p2q\" (UniqueName: \"kubernetes.io/projected/afa6a6bd-d9cf-4379-b4ed-452bff7149ae-kube-api-access-k7p2q\") pod \"redhat-marketplace-4mw5m\" (UID: \"afa6a6bd-d9cf-4379-b4ed-452bff7149ae\") " pod="openshift-marketplace/redhat-marketplace-4mw5m" Oct 05 21:30:06 crc kubenswrapper[4754]: I1005 21:30:06.826439 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/afa6a6bd-d9cf-4379-b4ed-452bff7149ae-utilities\") pod \"redhat-marketplace-4mw5m\" (UID: \"afa6a6bd-d9cf-4379-b4ed-452bff7149ae\") " pod="openshift-marketplace/redhat-marketplace-4mw5m" Oct 05 21:30:06 crc kubenswrapper[4754]: I1005 21:30:06.826552 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k7p2q\" (UniqueName: \"kubernetes.io/projected/afa6a6bd-d9cf-4379-b4ed-452bff7149ae-kube-api-access-k7p2q\") pod \"redhat-marketplace-4mw5m\" (UID: \"afa6a6bd-d9cf-4379-b4ed-452bff7149ae\") " pod="openshift-marketplace/redhat-marketplace-4mw5m" Oct 05 21:30:06 crc kubenswrapper[4754]: I1005 21:30:06.826578 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/afa6a6bd-d9cf-4379-b4ed-452bff7149ae-catalog-content\") pod \"redhat-marketplace-4mw5m\" (UID: \"afa6a6bd-d9cf-4379-b4ed-452bff7149ae\") " pod="openshift-marketplace/redhat-marketplace-4mw5m" Oct 05 21:30:06 crc kubenswrapper[4754]: I1005 21:30:06.827095 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/afa6a6bd-d9cf-4379-b4ed-452bff7149ae-utilities\") pod \"redhat-marketplace-4mw5m\" (UID: \"afa6a6bd-d9cf-4379-b4ed-452bff7149ae\") " pod="openshift-marketplace/redhat-marketplace-4mw5m" Oct 05 21:30:06 crc kubenswrapper[4754]: I1005 21:30:06.827141 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/afa6a6bd-d9cf-4379-b4ed-452bff7149ae-catalog-content\") pod \"redhat-marketplace-4mw5m\" (UID: \"afa6a6bd-d9cf-4379-b4ed-452bff7149ae\") " pod="openshift-marketplace/redhat-marketplace-4mw5m" Oct 05 21:30:06 crc kubenswrapper[4754]: I1005 21:30:06.848780 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k7p2q\" (UniqueName: \"kubernetes.io/projected/afa6a6bd-d9cf-4379-b4ed-452bff7149ae-kube-api-access-k7p2q\") pod \"redhat-marketplace-4mw5m\" (UID: \"afa6a6bd-d9cf-4379-b4ed-452bff7149ae\") " pod="openshift-marketplace/redhat-marketplace-4mw5m" Oct 05 21:30:06 crc kubenswrapper[4754]: I1005 21:30:06.955421 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4mw5m" Oct 05 21:30:07 crc kubenswrapper[4754]: I1005 21:30:07.469275 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4mw5m"] Oct 05 21:30:07 crc kubenswrapper[4754]: I1005 21:30:07.861057 4754 generic.go:334] "Generic (PLEG): container finished" podID="afa6a6bd-d9cf-4379-b4ed-452bff7149ae" containerID="87bc79876d6a0b07484ecdae6175897b43680ca3775b79030856a03dcbcae10e" exitCode=0 Oct 05 21:30:07 crc kubenswrapper[4754]: I1005 21:30:07.861795 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4mw5m" event={"ID":"afa6a6bd-d9cf-4379-b4ed-452bff7149ae","Type":"ContainerDied","Data":"87bc79876d6a0b07484ecdae6175897b43680ca3775b79030856a03dcbcae10e"} Oct 05 21:30:07 crc kubenswrapper[4754]: I1005 21:30:07.862479 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4mw5m" event={"ID":"afa6a6bd-d9cf-4379-b4ed-452bff7149ae","Type":"ContainerStarted","Data":"4074e2c8ab02ea1f7a2a382221e98d16123b505d332aae9196943f6c49d3effb"} Oct 05 21:30:09 crc kubenswrapper[4754]: I1005 21:30:09.885673 4754 generic.go:334] "Generic (PLEG): container finished" podID="afa6a6bd-d9cf-4379-b4ed-452bff7149ae" containerID="9cc96bd93a8fffe110b4b4b8f54493e3599b9cb81822493682ec500dfb32830b" exitCode=0 Oct 05 21:30:09 crc kubenswrapper[4754]: I1005 21:30:09.885804 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4mw5m" event={"ID":"afa6a6bd-d9cf-4379-b4ed-452bff7149ae","Type":"ContainerDied","Data":"9cc96bd93a8fffe110b4b4b8f54493e3599b9cb81822493682ec500dfb32830b"} Oct 05 21:30:10 crc kubenswrapper[4754]: I1005 21:30:10.904894 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4mw5m" event={"ID":"afa6a6bd-d9cf-4379-b4ed-452bff7149ae","Type":"ContainerStarted","Data":"61c236730f6787cf3e79717c8f9e99edc189740cbd58cdf8e32e468dd3938846"} Oct 05 21:30:10 crc kubenswrapper[4754]: I1005 21:30:10.937230 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-4mw5m" podStartSLOduration=2.398259859 podStartE2EDuration="4.93721016s" podCreationTimestamp="2025-10-05 21:30:06 +0000 UTC" firstStartedPulling="2025-10-05 21:30:07.863404842 +0000 UTC m=+2131.767523552" lastFinishedPulling="2025-10-05 21:30:10.402355133 +0000 UTC m=+2134.306473853" observedRunningTime="2025-10-05 21:30:10.933854981 +0000 UTC m=+2134.837973701" watchObservedRunningTime="2025-10-05 21:30:10.93721016 +0000 UTC m=+2134.841328880" Oct 05 21:30:11 crc kubenswrapper[4754]: I1005 21:30:11.742312 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-5nb8z"] Oct 05 21:30:11 crc kubenswrapper[4754]: I1005 21:30:11.750003 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5nb8z" Oct 05 21:30:11 crc kubenswrapper[4754]: I1005 21:30:11.763386 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-5nb8z"] Oct 05 21:30:11 crc kubenswrapper[4754]: I1005 21:30:11.862851 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/279f17f1-4a1a-44d1-9634-fb921e129a43-utilities\") pod \"community-operators-5nb8z\" (UID: \"279f17f1-4a1a-44d1-9634-fb921e129a43\") " pod="openshift-marketplace/community-operators-5nb8z" Oct 05 21:30:11 crc kubenswrapper[4754]: I1005 21:30:11.862992 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/279f17f1-4a1a-44d1-9634-fb921e129a43-catalog-content\") pod \"community-operators-5nb8z\" (UID: \"279f17f1-4a1a-44d1-9634-fb921e129a43\") " pod="openshift-marketplace/community-operators-5nb8z" Oct 05 21:30:11 crc kubenswrapper[4754]: I1005 21:30:11.863103 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6gtlk\" (UniqueName: \"kubernetes.io/projected/279f17f1-4a1a-44d1-9634-fb921e129a43-kube-api-access-6gtlk\") pod \"community-operators-5nb8z\" (UID: \"279f17f1-4a1a-44d1-9634-fb921e129a43\") " pod="openshift-marketplace/community-operators-5nb8z" Oct 05 21:30:11 crc kubenswrapper[4754]: I1005 21:30:11.967387 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6gtlk\" (UniqueName: \"kubernetes.io/projected/279f17f1-4a1a-44d1-9634-fb921e129a43-kube-api-access-6gtlk\") pod \"community-operators-5nb8z\" (UID: \"279f17f1-4a1a-44d1-9634-fb921e129a43\") " pod="openshift-marketplace/community-operators-5nb8z" Oct 05 21:30:11 crc kubenswrapper[4754]: I1005 21:30:11.968212 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/279f17f1-4a1a-44d1-9634-fb921e129a43-utilities\") pod \"community-operators-5nb8z\" (UID: \"279f17f1-4a1a-44d1-9634-fb921e129a43\") " pod="openshift-marketplace/community-operators-5nb8z" Oct 05 21:30:11 crc kubenswrapper[4754]: I1005 21:30:11.968301 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/279f17f1-4a1a-44d1-9634-fb921e129a43-catalog-content\") pod \"community-operators-5nb8z\" (UID: \"279f17f1-4a1a-44d1-9634-fb921e129a43\") " pod="openshift-marketplace/community-operators-5nb8z" Oct 05 21:30:11 crc kubenswrapper[4754]: I1005 21:30:11.968783 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/279f17f1-4a1a-44d1-9634-fb921e129a43-utilities\") pod \"community-operators-5nb8z\" (UID: \"279f17f1-4a1a-44d1-9634-fb921e129a43\") " pod="openshift-marketplace/community-operators-5nb8z" Oct 05 21:30:11 crc kubenswrapper[4754]: I1005 21:30:11.968908 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/279f17f1-4a1a-44d1-9634-fb921e129a43-catalog-content\") pod \"community-operators-5nb8z\" (UID: \"279f17f1-4a1a-44d1-9634-fb921e129a43\") " pod="openshift-marketplace/community-operators-5nb8z" Oct 05 21:30:11 crc kubenswrapper[4754]: I1005 21:30:11.992481 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6gtlk\" (UniqueName: \"kubernetes.io/projected/279f17f1-4a1a-44d1-9634-fb921e129a43-kube-api-access-6gtlk\") pod \"community-operators-5nb8z\" (UID: \"279f17f1-4a1a-44d1-9634-fb921e129a43\") " pod="openshift-marketplace/community-operators-5nb8z" Oct 05 21:30:12 crc kubenswrapper[4754]: I1005 21:30:12.074403 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5nb8z" Oct 05 21:30:12 crc kubenswrapper[4754]: I1005 21:30:12.558993 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-5nb8z"] Oct 05 21:30:12 crc kubenswrapper[4754]: I1005 21:30:12.934338 4754 generic.go:334] "Generic (PLEG): container finished" podID="279f17f1-4a1a-44d1-9634-fb921e129a43" containerID="5011a49fff738fc78415c7621a3d95bcb3d587b8f3d136b1c51f1edd0b0efe5a" exitCode=0 Oct 05 21:30:12 crc kubenswrapper[4754]: I1005 21:30:12.934416 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5nb8z" event={"ID":"279f17f1-4a1a-44d1-9634-fb921e129a43","Type":"ContainerDied","Data":"5011a49fff738fc78415c7621a3d95bcb3d587b8f3d136b1c51f1edd0b0efe5a"} Oct 05 21:30:12 crc kubenswrapper[4754]: I1005 21:30:12.934476 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5nb8z" event={"ID":"279f17f1-4a1a-44d1-9634-fb921e129a43","Type":"ContainerStarted","Data":"7c0e96738349cb471d038eedbe7a1ab077229b2adb60ea65caacbb22648aba0e"} Oct 05 21:30:14 crc kubenswrapper[4754]: I1005 21:30:14.978165 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5nb8z" event={"ID":"279f17f1-4a1a-44d1-9634-fb921e129a43","Type":"ContainerStarted","Data":"a154eb52a0e4eb3972a1a035441b64735dd56c0bab28645f3c7dde8c13eb81bb"} Oct 05 21:30:15 crc kubenswrapper[4754]: I1005 21:30:15.993067 4754 generic.go:334] "Generic (PLEG): container finished" podID="279f17f1-4a1a-44d1-9634-fb921e129a43" containerID="a154eb52a0e4eb3972a1a035441b64735dd56c0bab28645f3c7dde8c13eb81bb" exitCode=0 Oct 05 21:30:15 crc kubenswrapper[4754]: I1005 21:30:15.993182 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5nb8z" event={"ID":"279f17f1-4a1a-44d1-9634-fb921e129a43","Type":"ContainerDied","Data":"a154eb52a0e4eb3972a1a035441b64735dd56c0bab28645f3c7dde8c13eb81bb"} Oct 05 21:30:16 crc kubenswrapper[4754]: I1005 21:30:16.956024 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-4mw5m" Oct 05 21:30:16 crc kubenswrapper[4754]: I1005 21:30:16.956472 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-4mw5m" Oct 05 21:30:17 crc kubenswrapper[4754]: I1005 21:30:17.006746 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5nb8z" event={"ID":"279f17f1-4a1a-44d1-9634-fb921e129a43","Type":"ContainerStarted","Data":"0ed9bcc538b9e8162fe95581c1847aadcc1170e7b50e29683c314682df3bd65c"} Oct 05 21:30:17 crc kubenswrapper[4754]: I1005 21:30:17.014314 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-4mw5m" Oct 05 21:30:17 crc kubenswrapper[4754]: I1005 21:30:17.041135 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-5nb8z" podStartSLOduration=2.509894556 podStartE2EDuration="6.041112148s" podCreationTimestamp="2025-10-05 21:30:11 +0000 UTC" firstStartedPulling="2025-10-05 21:30:12.938769636 +0000 UTC m=+2136.842888386" lastFinishedPulling="2025-10-05 21:30:16.469987238 +0000 UTC m=+2140.374105978" observedRunningTime="2025-10-05 21:30:17.03590367 +0000 UTC m=+2140.940022390" watchObservedRunningTime="2025-10-05 21:30:17.041112148 +0000 UTC m=+2140.945230858" Oct 05 21:30:17 crc kubenswrapper[4754]: I1005 21:30:17.089179 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-4mw5m" Oct 05 21:30:19 crc kubenswrapper[4754]: I1005 21:30:19.128652 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-4mw5m"] Oct 05 21:30:19 crc kubenswrapper[4754]: I1005 21:30:19.129889 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-4mw5m" podUID="afa6a6bd-d9cf-4379-b4ed-452bff7149ae" containerName="registry-server" containerID="cri-o://61c236730f6787cf3e79717c8f9e99edc189740cbd58cdf8e32e468dd3938846" gracePeriod=2 Oct 05 21:30:19 crc kubenswrapper[4754]: I1005 21:30:19.696846 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4mw5m" Oct 05 21:30:19 crc kubenswrapper[4754]: I1005 21:30:19.876090 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/afa6a6bd-d9cf-4379-b4ed-452bff7149ae-utilities\") pod \"afa6a6bd-d9cf-4379-b4ed-452bff7149ae\" (UID: \"afa6a6bd-d9cf-4379-b4ed-452bff7149ae\") " Oct 05 21:30:19 crc kubenswrapper[4754]: I1005 21:30:19.876261 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k7p2q\" (UniqueName: \"kubernetes.io/projected/afa6a6bd-d9cf-4379-b4ed-452bff7149ae-kube-api-access-k7p2q\") pod \"afa6a6bd-d9cf-4379-b4ed-452bff7149ae\" (UID: \"afa6a6bd-d9cf-4379-b4ed-452bff7149ae\") " Oct 05 21:30:19 crc kubenswrapper[4754]: I1005 21:30:19.876394 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/afa6a6bd-d9cf-4379-b4ed-452bff7149ae-catalog-content\") pod \"afa6a6bd-d9cf-4379-b4ed-452bff7149ae\" (UID: \"afa6a6bd-d9cf-4379-b4ed-452bff7149ae\") " Oct 05 21:30:19 crc kubenswrapper[4754]: I1005 21:30:19.876829 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/afa6a6bd-d9cf-4379-b4ed-452bff7149ae-utilities" (OuterVolumeSpecName: "utilities") pod "afa6a6bd-d9cf-4379-b4ed-452bff7149ae" (UID: "afa6a6bd-d9cf-4379-b4ed-452bff7149ae"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 21:30:19 crc kubenswrapper[4754]: I1005 21:30:19.877166 4754 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/afa6a6bd-d9cf-4379-b4ed-452bff7149ae-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 21:30:19 crc kubenswrapper[4754]: I1005 21:30:19.885955 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/afa6a6bd-d9cf-4379-b4ed-452bff7149ae-kube-api-access-k7p2q" (OuterVolumeSpecName: "kube-api-access-k7p2q") pod "afa6a6bd-d9cf-4379-b4ed-452bff7149ae" (UID: "afa6a6bd-d9cf-4379-b4ed-452bff7149ae"). InnerVolumeSpecName "kube-api-access-k7p2q". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:30:19 crc kubenswrapper[4754]: I1005 21:30:19.893076 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/afa6a6bd-d9cf-4379-b4ed-452bff7149ae-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "afa6a6bd-d9cf-4379-b4ed-452bff7149ae" (UID: "afa6a6bd-d9cf-4379-b4ed-452bff7149ae"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 21:30:19 crc kubenswrapper[4754]: I1005 21:30:19.980343 4754 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/afa6a6bd-d9cf-4379-b4ed-452bff7149ae-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 21:30:19 crc kubenswrapper[4754]: I1005 21:30:19.980375 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k7p2q\" (UniqueName: \"kubernetes.io/projected/afa6a6bd-d9cf-4379-b4ed-452bff7149ae-kube-api-access-k7p2q\") on node \"crc\" DevicePath \"\"" Oct 05 21:30:20 crc kubenswrapper[4754]: I1005 21:30:20.040960 4754 generic.go:334] "Generic (PLEG): container finished" podID="afa6a6bd-d9cf-4379-b4ed-452bff7149ae" containerID="61c236730f6787cf3e79717c8f9e99edc189740cbd58cdf8e32e468dd3938846" exitCode=0 Oct 05 21:30:20 crc kubenswrapper[4754]: I1005 21:30:20.041035 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4mw5m" event={"ID":"afa6a6bd-d9cf-4379-b4ed-452bff7149ae","Type":"ContainerDied","Data":"61c236730f6787cf3e79717c8f9e99edc189740cbd58cdf8e32e468dd3938846"} Oct 05 21:30:20 crc kubenswrapper[4754]: I1005 21:30:20.041080 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4mw5m" event={"ID":"afa6a6bd-d9cf-4379-b4ed-452bff7149ae","Type":"ContainerDied","Data":"4074e2c8ab02ea1f7a2a382221e98d16123b505d332aae9196943f6c49d3effb"} Oct 05 21:30:20 crc kubenswrapper[4754]: I1005 21:30:20.041111 4754 scope.go:117] "RemoveContainer" containerID="61c236730f6787cf3e79717c8f9e99edc189740cbd58cdf8e32e468dd3938846" Oct 05 21:30:20 crc kubenswrapper[4754]: I1005 21:30:20.041349 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4mw5m" Oct 05 21:30:20 crc kubenswrapper[4754]: I1005 21:30:20.061819 4754 scope.go:117] "RemoveContainer" containerID="9cc96bd93a8fffe110b4b4b8f54493e3599b9cb81822493682ec500dfb32830b" Oct 05 21:30:20 crc kubenswrapper[4754]: I1005 21:30:20.103144 4754 scope.go:117] "RemoveContainer" containerID="87bc79876d6a0b07484ecdae6175897b43680ca3775b79030856a03dcbcae10e" Oct 05 21:30:20 crc kubenswrapper[4754]: I1005 21:30:20.108412 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-4mw5m"] Oct 05 21:30:20 crc kubenswrapper[4754]: I1005 21:30:20.117929 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-4mw5m"] Oct 05 21:30:20 crc kubenswrapper[4754]: I1005 21:30:20.172956 4754 scope.go:117] "RemoveContainer" containerID="61c236730f6787cf3e79717c8f9e99edc189740cbd58cdf8e32e468dd3938846" Oct 05 21:30:20 crc kubenswrapper[4754]: E1005 21:30:20.173713 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"61c236730f6787cf3e79717c8f9e99edc189740cbd58cdf8e32e468dd3938846\": container with ID starting with 61c236730f6787cf3e79717c8f9e99edc189740cbd58cdf8e32e468dd3938846 not found: ID does not exist" containerID="61c236730f6787cf3e79717c8f9e99edc189740cbd58cdf8e32e468dd3938846" Oct 05 21:30:20 crc kubenswrapper[4754]: I1005 21:30:20.173764 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"61c236730f6787cf3e79717c8f9e99edc189740cbd58cdf8e32e468dd3938846"} err="failed to get container status \"61c236730f6787cf3e79717c8f9e99edc189740cbd58cdf8e32e468dd3938846\": rpc error: code = NotFound desc = could not find container \"61c236730f6787cf3e79717c8f9e99edc189740cbd58cdf8e32e468dd3938846\": container with ID starting with 61c236730f6787cf3e79717c8f9e99edc189740cbd58cdf8e32e468dd3938846 not found: ID does not exist" Oct 05 21:30:20 crc kubenswrapper[4754]: I1005 21:30:20.173798 4754 scope.go:117] "RemoveContainer" containerID="9cc96bd93a8fffe110b4b4b8f54493e3599b9cb81822493682ec500dfb32830b" Oct 05 21:30:20 crc kubenswrapper[4754]: E1005 21:30:20.174275 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9cc96bd93a8fffe110b4b4b8f54493e3599b9cb81822493682ec500dfb32830b\": container with ID starting with 9cc96bd93a8fffe110b4b4b8f54493e3599b9cb81822493682ec500dfb32830b not found: ID does not exist" containerID="9cc96bd93a8fffe110b4b4b8f54493e3599b9cb81822493682ec500dfb32830b" Oct 05 21:30:20 crc kubenswrapper[4754]: I1005 21:30:20.174298 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9cc96bd93a8fffe110b4b4b8f54493e3599b9cb81822493682ec500dfb32830b"} err="failed to get container status \"9cc96bd93a8fffe110b4b4b8f54493e3599b9cb81822493682ec500dfb32830b\": rpc error: code = NotFound desc = could not find container \"9cc96bd93a8fffe110b4b4b8f54493e3599b9cb81822493682ec500dfb32830b\": container with ID starting with 9cc96bd93a8fffe110b4b4b8f54493e3599b9cb81822493682ec500dfb32830b not found: ID does not exist" Oct 05 21:30:20 crc kubenswrapper[4754]: I1005 21:30:20.174317 4754 scope.go:117] "RemoveContainer" containerID="87bc79876d6a0b07484ecdae6175897b43680ca3775b79030856a03dcbcae10e" Oct 05 21:30:20 crc kubenswrapper[4754]: E1005 21:30:20.174616 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"87bc79876d6a0b07484ecdae6175897b43680ca3775b79030856a03dcbcae10e\": container with ID starting with 87bc79876d6a0b07484ecdae6175897b43680ca3775b79030856a03dcbcae10e not found: ID does not exist" containerID="87bc79876d6a0b07484ecdae6175897b43680ca3775b79030856a03dcbcae10e" Oct 05 21:30:20 crc kubenswrapper[4754]: I1005 21:30:20.174641 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"87bc79876d6a0b07484ecdae6175897b43680ca3775b79030856a03dcbcae10e"} err="failed to get container status \"87bc79876d6a0b07484ecdae6175897b43680ca3775b79030856a03dcbcae10e\": rpc error: code = NotFound desc = could not find container \"87bc79876d6a0b07484ecdae6175897b43680ca3775b79030856a03dcbcae10e\": container with ID starting with 87bc79876d6a0b07484ecdae6175897b43680ca3775b79030856a03dcbcae10e not found: ID does not exist" Oct 05 21:30:20 crc kubenswrapper[4754]: I1005 21:30:20.857140 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="afa6a6bd-d9cf-4379-b4ed-452bff7149ae" path="/var/lib/kubelet/pods/afa6a6bd-d9cf-4379-b4ed-452bff7149ae/volumes" Oct 05 21:30:22 crc kubenswrapper[4754]: I1005 21:30:22.074903 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-5nb8z" Oct 05 21:30:22 crc kubenswrapper[4754]: I1005 21:30:22.075575 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-5nb8z" Oct 05 21:30:22 crc kubenswrapper[4754]: I1005 21:30:22.156067 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-5nb8z" Oct 05 21:30:23 crc kubenswrapper[4754]: I1005 21:30:23.166817 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-5nb8z" Oct 05 21:30:23 crc kubenswrapper[4754]: I1005 21:30:23.326946 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-5nb8z"] Oct 05 21:30:25 crc kubenswrapper[4754]: I1005 21:30:25.113541 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-5nb8z" podUID="279f17f1-4a1a-44d1-9634-fb921e129a43" containerName="registry-server" containerID="cri-o://0ed9bcc538b9e8162fe95581c1847aadcc1170e7b50e29683c314682df3bd65c" gracePeriod=2 Oct 05 21:30:25 crc kubenswrapper[4754]: I1005 21:30:25.776145 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5nb8z" Oct 05 21:30:25 crc kubenswrapper[4754]: I1005 21:30:25.926869 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/279f17f1-4a1a-44d1-9634-fb921e129a43-catalog-content\") pod \"279f17f1-4a1a-44d1-9634-fb921e129a43\" (UID: \"279f17f1-4a1a-44d1-9634-fb921e129a43\") " Oct 05 21:30:25 crc kubenswrapper[4754]: I1005 21:30:25.926954 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6gtlk\" (UniqueName: \"kubernetes.io/projected/279f17f1-4a1a-44d1-9634-fb921e129a43-kube-api-access-6gtlk\") pod \"279f17f1-4a1a-44d1-9634-fb921e129a43\" (UID: \"279f17f1-4a1a-44d1-9634-fb921e129a43\") " Oct 05 21:30:25 crc kubenswrapper[4754]: I1005 21:30:25.927149 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/279f17f1-4a1a-44d1-9634-fb921e129a43-utilities\") pod \"279f17f1-4a1a-44d1-9634-fb921e129a43\" (UID: \"279f17f1-4a1a-44d1-9634-fb921e129a43\") " Oct 05 21:30:25 crc kubenswrapper[4754]: I1005 21:30:25.929237 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/279f17f1-4a1a-44d1-9634-fb921e129a43-utilities" (OuterVolumeSpecName: "utilities") pod "279f17f1-4a1a-44d1-9634-fb921e129a43" (UID: "279f17f1-4a1a-44d1-9634-fb921e129a43"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 21:30:25 crc kubenswrapper[4754]: I1005 21:30:25.948898 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/279f17f1-4a1a-44d1-9634-fb921e129a43-kube-api-access-6gtlk" (OuterVolumeSpecName: "kube-api-access-6gtlk") pod "279f17f1-4a1a-44d1-9634-fb921e129a43" (UID: "279f17f1-4a1a-44d1-9634-fb921e129a43"). InnerVolumeSpecName "kube-api-access-6gtlk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:30:25 crc kubenswrapper[4754]: I1005 21:30:25.989434 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/279f17f1-4a1a-44d1-9634-fb921e129a43-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "279f17f1-4a1a-44d1-9634-fb921e129a43" (UID: "279f17f1-4a1a-44d1-9634-fb921e129a43"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 21:30:26 crc kubenswrapper[4754]: I1005 21:30:26.030287 4754 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/279f17f1-4a1a-44d1-9634-fb921e129a43-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 21:30:26 crc kubenswrapper[4754]: I1005 21:30:26.030625 4754 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/279f17f1-4a1a-44d1-9634-fb921e129a43-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 21:30:26 crc kubenswrapper[4754]: I1005 21:30:26.030722 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6gtlk\" (UniqueName: \"kubernetes.io/projected/279f17f1-4a1a-44d1-9634-fb921e129a43-kube-api-access-6gtlk\") on node \"crc\" DevicePath \"\"" Oct 05 21:30:26 crc kubenswrapper[4754]: I1005 21:30:26.126272 4754 generic.go:334] "Generic (PLEG): container finished" podID="279f17f1-4a1a-44d1-9634-fb921e129a43" containerID="0ed9bcc538b9e8162fe95581c1847aadcc1170e7b50e29683c314682df3bd65c" exitCode=0 Oct 05 21:30:26 crc kubenswrapper[4754]: I1005 21:30:26.126366 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5nb8z" event={"ID":"279f17f1-4a1a-44d1-9634-fb921e129a43","Type":"ContainerDied","Data":"0ed9bcc538b9e8162fe95581c1847aadcc1170e7b50e29683c314682df3bd65c"} Oct 05 21:30:26 crc kubenswrapper[4754]: I1005 21:30:26.126427 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5nb8z" event={"ID":"279f17f1-4a1a-44d1-9634-fb921e129a43","Type":"ContainerDied","Data":"7c0e96738349cb471d038eedbe7a1ab077229b2adb60ea65caacbb22648aba0e"} Oct 05 21:30:26 crc kubenswrapper[4754]: I1005 21:30:26.126461 4754 scope.go:117] "RemoveContainer" containerID="0ed9bcc538b9e8162fe95581c1847aadcc1170e7b50e29683c314682df3bd65c" Oct 05 21:30:26 crc kubenswrapper[4754]: I1005 21:30:26.126787 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5nb8z" Oct 05 21:30:26 crc kubenswrapper[4754]: I1005 21:30:26.167141 4754 scope.go:117] "RemoveContainer" containerID="a154eb52a0e4eb3972a1a035441b64735dd56c0bab28645f3c7dde8c13eb81bb" Oct 05 21:30:26 crc kubenswrapper[4754]: I1005 21:30:26.189632 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-5nb8z"] Oct 05 21:30:26 crc kubenswrapper[4754]: I1005 21:30:26.197027 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-5nb8z"] Oct 05 21:30:26 crc kubenswrapper[4754]: I1005 21:30:26.205422 4754 scope.go:117] "RemoveContainer" containerID="5011a49fff738fc78415c7621a3d95bcb3d587b8f3d136b1c51f1edd0b0efe5a" Oct 05 21:30:26 crc kubenswrapper[4754]: I1005 21:30:26.245145 4754 scope.go:117] "RemoveContainer" containerID="0ed9bcc538b9e8162fe95581c1847aadcc1170e7b50e29683c314682df3bd65c" Oct 05 21:30:26 crc kubenswrapper[4754]: E1005 21:30:26.245977 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0ed9bcc538b9e8162fe95581c1847aadcc1170e7b50e29683c314682df3bd65c\": container with ID starting with 0ed9bcc538b9e8162fe95581c1847aadcc1170e7b50e29683c314682df3bd65c not found: ID does not exist" containerID="0ed9bcc538b9e8162fe95581c1847aadcc1170e7b50e29683c314682df3bd65c" Oct 05 21:30:26 crc kubenswrapper[4754]: I1005 21:30:26.246075 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0ed9bcc538b9e8162fe95581c1847aadcc1170e7b50e29683c314682df3bd65c"} err="failed to get container status \"0ed9bcc538b9e8162fe95581c1847aadcc1170e7b50e29683c314682df3bd65c\": rpc error: code = NotFound desc = could not find container \"0ed9bcc538b9e8162fe95581c1847aadcc1170e7b50e29683c314682df3bd65c\": container with ID starting with 0ed9bcc538b9e8162fe95581c1847aadcc1170e7b50e29683c314682df3bd65c not found: ID does not exist" Oct 05 21:30:26 crc kubenswrapper[4754]: I1005 21:30:26.246117 4754 scope.go:117] "RemoveContainer" containerID="a154eb52a0e4eb3972a1a035441b64735dd56c0bab28645f3c7dde8c13eb81bb" Oct 05 21:30:26 crc kubenswrapper[4754]: E1005 21:30:26.246666 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a154eb52a0e4eb3972a1a035441b64735dd56c0bab28645f3c7dde8c13eb81bb\": container with ID starting with a154eb52a0e4eb3972a1a035441b64735dd56c0bab28645f3c7dde8c13eb81bb not found: ID does not exist" containerID="a154eb52a0e4eb3972a1a035441b64735dd56c0bab28645f3c7dde8c13eb81bb" Oct 05 21:30:26 crc kubenswrapper[4754]: I1005 21:30:26.246713 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a154eb52a0e4eb3972a1a035441b64735dd56c0bab28645f3c7dde8c13eb81bb"} err="failed to get container status \"a154eb52a0e4eb3972a1a035441b64735dd56c0bab28645f3c7dde8c13eb81bb\": rpc error: code = NotFound desc = could not find container \"a154eb52a0e4eb3972a1a035441b64735dd56c0bab28645f3c7dde8c13eb81bb\": container with ID starting with a154eb52a0e4eb3972a1a035441b64735dd56c0bab28645f3c7dde8c13eb81bb not found: ID does not exist" Oct 05 21:30:26 crc kubenswrapper[4754]: I1005 21:30:26.246745 4754 scope.go:117] "RemoveContainer" containerID="5011a49fff738fc78415c7621a3d95bcb3d587b8f3d136b1c51f1edd0b0efe5a" Oct 05 21:30:26 crc kubenswrapper[4754]: E1005 21:30:26.247104 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5011a49fff738fc78415c7621a3d95bcb3d587b8f3d136b1c51f1edd0b0efe5a\": container with ID starting with 5011a49fff738fc78415c7621a3d95bcb3d587b8f3d136b1c51f1edd0b0efe5a not found: ID does not exist" containerID="5011a49fff738fc78415c7621a3d95bcb3d587b8f3d136b1c51f1edd0b0efe5a" Oct 05 21:30:26 crc kubenswrapper[4754]: I1005 21:30:26.247132 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5011a49fff738fc78415c7621a3d95bcb3d587b8f3d136b1c51f1edd0b0efe5a"} err="failed to get container status \"5011a49fff738fc78415c7621a3d95bcb3d587b8f3d136b1c51f1edd0b0efe5a\": rpc error: code = NotFound desc = could not find container \"5011a49fff738fc78415c7621a3d95bcb3d587b8f3d136b1c51f1edd0b0efe5a\": container with ID starting with 5011a49fff738fc78415c7621a3d95bcb3d587b8f3d136b1c51f1edd0b0efe5a not found: ID does not exist" Oct 05 21:30:26 crc kubenswrapper[4754]: I1005 21:30:26.865117 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="279f17f1-4a1a-44d1-9634-fb921e129a43" path="/var/lib/kubelet/pods/279f17f1-4a1a-44d1-9634-fb921e129a43/volumes" Oct 05 21:30:41 crc kubenswrapper[4754]: I1005 21:30:41.207908 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-bfl4v"] Oct 05 21:30:41 crc kubenswrapper[4754]: E1005 21:30:41.209030 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="279f17f1-4a1a-44d1-9634-fb921e129a43" containerName="extract-utilities" Oct 05 21:30:41 crc kubenswrapper[4754]: I1005 21:30:41.209048 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="279f17f1-4a1a-44d1-9634-fb921e129a43" containerName="extract-utilities" Oct 05 21:30:41 crc kubenswrapper[4754]: E1005 21:30:41.209065 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="279f17f1-4a1a-44d1-9634-fb921e129a43" containerName="extract-content" Oct 05 21:30:41 crc kubenswrapper[4754]: I1005 21:30:41.209075 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="279f17f1-4a1a-44d1-9634-fb921e129a43" containerName="extract-content" Oct 05 21:30:41 crc kubenswrapper[4754]: E1005 21:30:41.209111 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="afa6a6bd-d9cf-4379-b4ed-452bff7149ae" containerName="registry-server" Oct 05 21:30:41 crc kubenswrapper[4754]: I1005 21:30:41.209121 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="afa6a6bd-d9cf-4379-b4ed-452bff7149ae" containerName="registry-server" Oct 05 21:30:41 crc kubenswrapper[4754]: E1005 21:30:41.209137 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="afa6a6bd-d9cf-4379-b4ed-452bff7149ae" containerName="extract-utilities" Oct 05 21:30:41 crc kubenswrapper[4754]: I1005 21:30:41.209146 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="afa6a6bd-d9cf-4379-b4ed-452bff7149ae" containerName="extract-utilities" Oct 05 21:30:41 crc kubenswrapper[4754]: E1005 21:30:41.209171 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="afa6a6bd-d9cf-4379-b4ed-452bff7149ae" containerName="extract-content" Oct 05 21:30:41 crc kubenswrapper[4754]: I1005 21:30:41.209182 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="afa6a6bd-d9cf-4379-b4ed-452bff7149ae" containerName="extract-content" Oct 05 21:30:41 crc kubenswrapper[4754]: E1005 21:30:41.209202 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="279f17f1-4a1a-44d1-9634-fb921e129a43" containerName="registry-server" Oct 05 21:30:41 crc kubenswrapper[4754]: I1005 21:30:41.209211 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="279f17f1-4a1a-44d1-9634-fb921e129a43" containerName="registry-server" Oct 05 21:30:41 crc kubenswrapper[4754]: I1005 21:30:41.209457 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="afa6a6bd-d9cf-4379-b4ed-452bff7149ae" containerName="registry-server" Oct 05 21:30:41 crc kubenswrapper[4754]: I1005 21:30:41.209541 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="279f17f1-4a1a-44d1-9634-fb921e129a43" containerName="registry-server" Oct 05 21:30:41 crc kubenswrapper[4754]: I1005 21:30:41.211350 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bfl4v" Oct 05 21:30:41 crc kubenswrapper[4754]: I1005 21:30:41.227328 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bfl4v"] Oct 05 21:30:41 crc kubenswrapper[4754]: I1005 21:30:41.259975 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0fffea5a-ebbf-4457-b966-0fb29dce3c46-utilities\") pod \"redhat-operators-bfl4v\" (UID: \"0fffea5a-ebbf-4457-b966-0fb29dce3c46\") " pod="openshift-marketplace/redhat-operators-bfl4v" Oct 05 21:30:41 crc kubenswrapper[4754]: I1005 21:30:41.260130 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7j4wr\" (UniqueName: \"kubernetes.io/projected/0fffea5a-ebbf-4457-b966-0fb29dce3c46-kube-api-access-7j4wr\") pod \"redhat-operators-bfl4v\" (UID: \"0fffea5a-ebbf-4457-b966-0fb29dce3c46\") " pod="openshift-marketplace/redhat-operators-bfl4v" Oct 05 21:30:41 crc kubenswrapper[4754]: I1005 21:30:41.260189 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0fffea5a-ebbf-4457-b966-0fb29dce3c46-catalog-content\") pod \"redhat-operators-bfl4v\" (UID: \"0fffea5a-ebbf-4457-b966-0fb29dce3c46\") " pod="openshift-marketplace/redhat-operators-bfl4v" Oct 05 21:30:41 crc kubenswrapper[4754]: I1005 21:30:41.361723 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7j4wr\" (UniqueName: \"kubernetes.io/projected/0fffea5a-ebbf-4457-b966-0fb29dce3c46-kube-api-access-7j4wr\") pod \"redhat-operators-bfl4v\" (UID: \"0fffea5a-ebbf-4457-b966-0fb29dce3c46\") " pod="openshift-marketplace/redhat-operators-bfl4v" Oct 05 21:30:41 crc kubenswrapper[4754]: I1005 21:30:41.361802 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0fffea5a-ebbf-4457-b966-0fb29dce3c46-catalog-content\") pod \"redhat-operators-bfl4v\" (UID: \"0fffea5a-ebbf-4457-b966-0fb29dce3c46\") " pod="openshift-marketplace/redhat-operators-bfl4v" Oct 05 21:30:41 crc kubenswrapper[4754]: I1005 21:30:41.361825 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0fffea5a-ebbf-4457-b966-0fb29dce3c46-utilities\") pod \"redhat-operators-bfl4v\" (UID: \"0fffea5a-ebbf-4457-b966-0fb29dce3c46\") " pod="openshift-marketplace/redhat-operators-bfl4v" Oct 05 21:30:41 crc kubenswrapper[4754]: I1005 21:30:41.362375 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0fffea5a-ebbf-4457-b966-0fb29dce3c46-utilities\") pod \"redhat-operators-bfl4v\" (UID: \"0fffea5a-ebbf-4457-b966-0fb29dce3c46\") " pod="openshift-marketplace/redhat-operators-bfl4v" Oct 05 21:30:41 crc kubenswrapper[4754]: I1005 21:30:41.362516 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0fffea5a-ebbf-4457-b966-0fb29dce3c46-catalog-content\") pod \"redhat-operators-bfl4v\" (UID: \"0fffea5a-ebbf-4457-b966-0fb29dce3c46\") " pod="openshift-marketplace/redhat-operators-bfl4v" Oct 05 21:30:41 crc kubenswrapper[4754]: I1005 21:30:41.397419 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7j4wr\" (UniqueName: \"kubernetes.io/projected/0fffea5a-ebbf-4457-b966-0fb29dce3c46-kube-api-access-7j4wr\") pod \"redhat-operators-bfl4v\" (UID: \"0fffea5a-ebbf-4457-b966-0fb29dce3c46\") " pod="openshift-marketplace/redhat-operators-bfl4v" Oct 05 21:30:41 crc kubenswrapper[4754]: I1005 21:30:41.534813 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bfl4v" Oct 05 21:30:42 crc kubenswrapper[4754]: I1005 21:30:42.085647 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bfl4v"] Oct 05 21:30:42 crc kubenswrapper[4754]: I1005 21:30:42.314024 4754 generic.go:334] "Generic (PLEG): container finished" podID="0fffea5a-ebbf-4457-b966-0fb29dce3c46" containerID="57218cb167c3e02b7535eadb43f94f24ad216dcd020cd3c6c49e62a929fe4a69" exitCode=0 Oct 05 21:30:42 crc kubenswrapper[4754]: I1005 21:30:42.314079 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bfl4v" event={"ID":"0fffea5a-ebbf-4457-b966-0fb29dce3c46","Type":"ContainerDied","Data":"57218cb167c3e02b7535eadb43f94f24ad216dcd020cd3c6c49e62a929fe4a69"} Oct 05 21:30:42 crc kubenswrapper[4754]: I1005 21:30:42.314113 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bfl4v" event={"ID":"0fffea5a-ebbf-4457-b966-0fb29dce3c46","Type":"ContainerStarted","Data":"3ef0ba6854d433eff25d645b07cb29c037dbb8b595730beb87cad1ff19e6fda6"} Oct 05 21:30:42 crc kubenswrapper[4754]: I1005 21:30:42.317076 4754 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 05 21:30:44 crc kubenswrapper[4754]: I1005 21:30:44.348475 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bfl4v" event={"ID":"0fffea5a-ebbf-4457-b966-0fb29dce3c46","Type":"ContainerStarted","Data":"dbfd5728591129fa7d2bba9eaf9de33ec3ca768d1bd68f415b8294059982e93f"} Oct 05 21:30:47 crc kubenswrapper[4754]: I1005 21:30:47.388575 4754 generic.go:334] "Generic (PLEG): container finished" podID="0fffea5a-ebbf-4457-b966-0fb29dce3c46" containerID="dbfd5728591129fa7d2bba9eaf9de33ec3ca768d1bd68f415b8294059982e93f" exitCode=0 Oct 05 21:30:47 crc kubenswrapper[4754]: I1005 21:30:47.388701 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bfl4v" event={"ID":"0fffea5a-ebbf-4457-b966-0fb29dce3c46","Type":"ContainerDied","Data":"dbfd5728591129fa7d2bba9eaf9de33ec3ca768d1bd68f415b8294059982e93f"} Oct 05 21:30:48 crc kubenswrapper[4754]: I1005 21:30:48.405741 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bfl4v" event={"ID":"0fffea5a-ebbf-4457-b966-0fb29dce3c46","Type":"ContainerStarted","Data":"4f98230eb701d9daba3183c224ab435d6051f86e006460cc034a5cc75da36a49"} Oct 05 21:30:51 crc kubenswrapper[4754]: I1005 21:30:51.535765 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-bfl4v" Oct 05 21:30:51 crc kubenswrapper[4754]: I1005 21:30:51.536313 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-bfl4v" Oct 05 21:30:52 crc kubenswrapper[4754]: I1005 21:30:52.614753 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-bfl4v" podUID="0fffea5a-ebbf-4457-b966-0fb29dce3c46" containerName="registry-server" probeResult="failure" output=< Oct 05 21:30:52 crc kubenswrapper[4754]: timeout: failed to connect service ":50051" within 1s Oct 05 21:30:52 crc kubenswrapper[4754]: > Oct 05 21:30:55 crc kubenswrapper[4754]: I1005 21:30:55.796368 4754 scope.go:117] "RemoveContainer" containerID="ff732b3e3d89105355ef4324b72c3db67106face7c8a9fd05f5ee00e8b2ceb24" Oct 05 21:31:01 crc kubenswrapper[4754]: I1005 21:31:01.600247 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-bfl4v" Oct 05 21:31:01 crc kubenswrapper[4754]: I1005 21:31:01.647669 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-bfl4v" podStartSLOduration=15.114928549 podStartE2EDuration="20.647633464s" podCreationTimestamp="2025-10-05 21:30:41 +0000 UTC" firstStartedPulling="2025-10-05 21:30:42.316827203 +0000 UTC m=+2166.220945913" lastFinishedPulling="2025-10-05 21:30:47.849532108 +0000 UTC m=+2171.753650828" observedRunningTime="2025-10-05 21:30:48.439008585 +0000 UTC m=+2172.343127305" watchObservedRunningTime="2025-10-05 21:31:01.647633464 +0000 UTC m=+2185.551752184" Oct 05 21:31:01 crc kubenswrapper[4754]: I1005 21:31:01.682085 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-bfl4v" Oct 05 21:31:01 crc kubenswrapper[4754]: I1005 21:31:01.857862 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-bfl4v"] Oct 05 21:31:03 crc kubenswrapper[4754]: I1005 21:31:03.606158 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-bfl4v" podUID="0fffea5a-ebbf-4457-b966-0fb29dce3c46" containerName="registry-server" containerID="cri-o://4f98230eb701d9daba3183c224ab435d6051f86e006460cc034a5cc75da36a49" gracePeriod=2 Oct 05 21:31:04 crc kubenswrapper[4754]: I1005 21:31:04.077343 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bfl4v" Oct 05 21:31:04 crc kubenswrapper[4754]: I1005 21:31:04.175164 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7j4wr\" (UniqueName: \"kubernetes.io/projected/0fffea5a-ebbf-4457-b966-0fb29dce3c46-kube-api-access-7j4wr\") pod \"0fffea5a-ebbf-4457-b966-0fb29dce3c46\" (UID: \"0fffea5a-ebbf-4457-b966-0fb29dce3c46\") " Oct 05 21:31:04 crc kubenswrapper[4754]: I1005 21:31:04.175216 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0fffea5a-ebbf-4457-b966-0fb29dce3c46-utilities\") pod \"0fffea5a-ebbf-4457-b966-0fb29dce3c46\" (UID: \"0fffea5a-ebbf-4457-b966-0fb29dce3c46\") " Oct 05 21:31:04 crc kubenswrapper[4754]: I1005 21:31:04.175277 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0fffea5a-ebbf-4457-b966-0fb29dce3c46-catalog-content\") pod \"0fffea5a-ebbf-4457-b966-0fb29dce3c46\" (UID: \"0fffea5a-ebbf-4457-b966-0fb29dce3c46\") " Oct 05 21:31:04 crc kubenswrapper[4754]: I1005 21:31:04.176100 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0fffea5a-ebbf-4457-b966-0fb29dce3c46-utilities" (OuterVolumeSpecName: "utilities") pod "0fffea5a-ebbf-4457-b966-0fb29dce3c46" (UID: "0fffea5a-ebbf-4457-b966-0fb29dce3c46"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 21:31:04 crc kubenswrapper[4754]: I1005 21:31:04.183539 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0fffea5a-ebbf-4457-b966-0fb29dce3c46-kube-api-access-7j4wr" (OuterVolumeSpecName: "kube-api-access-7j4wr") pod "0fffea5a-ebbf-4457-b966-0fb29dce3c46" (UID: "0fffea5a-ebbf-4457-b966-0fb29dce3c46"). InnerVolumeSpecName "kube-api-access-7j4wr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:31:04 crc kubenswrapper[4754]: I1005 21:31:04.274227 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0fffea5a-ebbf-4457-b966-0fb29dce3c46-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0fffea5a-ebbf-4457-b966-0fb29dce3c46" (UID: "0fffea5a-ebbf-4457-b966-0fb29dce3c46"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 21:31:04 crc kubenswrapper[4754]: I1005 21:31:04.278083 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7j4wr\" (UniqueName: \"kubernetes.io/projected/0fffea5a-ebbf-4457-b966-0fb29dce3c46-kube-api-access-7j4wr\") on node \"crc\" DevicePath \"\"" Oct 05 21:31:04 crc kubenswrapper[4754]: I1005 21:31:04.278190 4754 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0fffea5a-ebbf-4457-b966-0fb29dce3c46-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 21:31:04 crc kubenswrapper[4754]: I1005 21:31:04.278268 4754 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0fffea5a-ebbf-4457-b966-0fb29dce3c46-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 21:31:04 crc kubenswrapper[4754]: I1005 21:31:04.619157 4754 generic.go:334] "Generic (PLEG): container finished" podID="0fffea5a-ebbf-4457-b966-0fb29dce3c46" containerID="4f98230eb701d9daba3183c224ab435d6051f86e006460cc034a5cc75da36a49" exitCode=0 Oct 05 21:31:04 crc kubenswrapper[4754]: I1005 21:31:04.619208 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bfl4v" event={"ID":"0fffea5a-ebbf-4457-b966-0fb29dce3c46","Type":"ContainerDied","Data":"4f98230eb701d9daba3183c224ab435d6051f86e006460cc034a5cc75da36a49"} Oct 05 21:31:04 crc kubenswrapper[4754]: I1005 21:31:04.619243 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bfl4v" event={"ID":"0fffea5a-ebbf-4457-b966-0fb29dce3c46","Type":"ContainerDied","Data":"3ef0ba6854d433eff25d645b07cb29c037dbb8b595730beb87cad1ff19e6fda6"} Oct 05 21:31:04 crc kubenswrapper[4754]: I1005 21:31:04.619261 4754 scope.go:117] "RemoveContainer" containerID="4f98230eb701d9daba3183c224ab435d6051f86e006460cc034a5cc75da36a49" Oct 05 21:31:04 crc kubenswrapper[4754]: I1005 21:31:04.620573 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bfl4v" Oct 05 21:31:04 crc kubenswrapper[4754]: I1005 21:31:04.690221 4754 scope.go:117] "RemoveContainer" containerID="dbfd5728591129fa7d2bba9eaf9de33ec3ca768d1bd68f415b8294059982e93f" Oct 05 21:31:04 crc kubenswrapper[4754]: I1005 21:31:04.718392 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-bfl4v"] Oct 05 21:31:04 crc kubenswrapper[4754]: I1005 21:31:04.725194 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-bfl4v"] Oct 05 21:31:04 crc kubenswrapper[4754]: I1005 21:31:04.730949 4754 scope.go:117] "RemoveContainer" containerID="57218cb167c3e02b7535eadb43f94f24ad216dcd020cd3c6c49e62a929fe4a69" Oct 05 21:31:04 crc kubenswrapper[4754]: I1005 21:31:04.770903 4754 scope.go:117] "RemoveContainer" containerID="4f98230eb701d9daba3183c224ab435d6051f86e006460cc034a5cc75da36a49" Oct 05 21:31:04 crc kubenswrapper[4754]: E1005 21:31:04.772167 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4f98230eb701d9daba3183c224ab435d6051f86e006460cc034a5cc75da36a49\": container with ID starting with 4f98230eb701d9daba3183c224ab435d6051f86e006460cc034a5cc75da36a49 not found: ID does not exist" containerID="4f98230eb701d9daba3183c224ab435d6051f86e006460cc034a5cc75da36a49" Oct 05 21:31:04 crc kubenswrapper[4754]: I1005 21:31:04.772275 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4f98230eb701d9daba3183c224ab435d6051f86e006460cc034a5cc75da36a49"} err="failed to get container status \"4f98230eb701d9daba3183c224ab435d6051f86e006460cc034a5cc75da36a49\": rpc error: code = NotFound desc = could not find container \"4f98230eb701d9daba3183c224ab435d6051f86e006460cc034a5cc75da36a49\": container with ID starting with 4f98230eb701d9daba3183c224ab435d6051f86e006460cc034a5cc75da36a49 not found: ID does not exist" Oct 05 21:31:04 crc kubenswrapper[4754]: I1005 21:31:04.772364 4754 scope.go:117] "RemoveContainer" containerID="dbfd5728591129fa7d2bba9eaf9de33ec3ca768d1bd68f415b8294059982e93f" Oct 05 21:31:04 crc kubenswrapper[4754]: E1005 21:31:04.772763 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dbfd5728591129fa7d2bba9eaf9de33ec3ca768d1bd68f415b8294059982e93f\": container with ID starting with dbfd5728591129fa7d2bba9eaf9de33ec3ca768d1bd68f415b8294059982e93f not found: ID does not exist" containerID="dbfd5728591129fa7d2bba9eaf9de33ec3ca768d1bd68f415b8294059982e93f" Oct 05 21:31:04 crc kubenswrapper[4754]: I1005 21:31:04.772857 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dbfd5728591129fa7d2bba9eaf9de33ec3ca768d1bd68f415b8294059982e93f"} err="failed to get container status \"dbfd5728591129fa7d2bba9eaf9de33ec3ca768d1bd68f415b8294059982e93f\": rpc error: code = NotFound desc = could not find container \"dbfd5728591129fa7d2bba9eaf9de33ec3ca768d1bd68f415b8294059982e93f\": container with ID starting with dbfd5728591129fa7d2bba9eaf9de33ec3ca768d1bd68f415b8294059982e93f not found: ID does not exist" Oct 05 21:31:04 crc kubenswrapper[4754]: I1005 21:31:04.772938 4754 scope.go:117] "RemoveContainer" containerID="57218cb167c3e02b7535eadb43f94f24ad216dcd020cd3c6c49e62a929fe4a69" Oct 05 21:31:04 crc kubenswrapper[4754]: E1005 21:31:04.773577 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"57218cb167c3e02b7535eadb43f94f24ad216dcd020cd3c6c49e62a929fe4a69\": container with ID starting with 57218cb167c3e02b7535eadb43f94f24ad216dcd020cd3c6c49e62a929fe4a69 not found: ID does not exist" containerID="57218cb167c3e02b7535eadb43f94f24ad216dcd020cd3c6c49e62a929fe4a69" Oct 05 21:31:04 crc kubenswrapper[4754]: I1005 21:31:04.773667 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"57218cb167c3e02b7535eadb43f94f24ad216dcd020cd3c6c49e62a929fe4a69"} err="failed to get container status \"57218cb167c3e02b7535eadb43f94f24ad216dcd020cd3c6c49e62a929fe4a69\": rpc error: code = NotFound desc = could not find container \"57218cb167c3e02b7535eadb43f94f24ad216dcd020cd3c6c49e62a929fe4a69\": container with ID starting with 57218cb167c3e02b7535eadb43f94f24ad216dcd020cd3c6c49e62a929fe4a69 not found: ID does not exist" Oct 05 21:31:04 crc kubenswrapper[4754]: I1005 21:31:04.852548 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0fffea5a-ebbf-4457-b966-0fb29dce3c46" path="/var/lib/kubelet/pods/0fffea5a-ebbf-4457-b966-0fb29dce3c46/volumes" Oct 05 21:31:05 crc kubenswrapper[4754]: I1005 21:31:05.245290 4754 patch_prober.go:28] interesting pod/machine-config-daemon-b2h9k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 21:31:05 crc kubenswrapper[4754]: I1005 21:31:05.245378 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 21:31:22 crc kubenswrapper[4754]: I1005 21:31:22.838283 4754 generic.go:334] "Generic (PLEG): container finished" podID="c26e6366-e15d-4c7b-a370-a601e2cee56f" containerID="92d6390e8d2fadbf9896fab63fe584db05fca63a9ca44a70123ae4bfe54d6503" exitCode=0 Oct 05 21:31:22 crc kubenswrapper[4754]: I1005 21:31:22.852749 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-cwth8" event={"ID":"c26e6366-e15d-4c7b-a370-a601e2cee56f","Type":"ContainerDied","Data":"92d6390e8d2fadbf9896fab63fe584db05fca63a9ca44a70123ae4bfe54d6503"} Oct 05 21:31:24 crc kubenswrapper[4754]: I1005 21:31:24.389227 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-cwth8" Oct 05 21:31:24 crc kubenswrapper[4754]: I1005 21:31:24.519252 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c26e6366-e15d-4c7b-a370-a601e2cee56f-ovn-combined-ca-bundle\") pod \"c26e6366-e15d-4c7b-a370-a601e2cee56f\" (UID: \"c26e6366-e15d-4c7b-a370-a601e2cee56f\") " Oct 05 21:31:24 crc kubenswrapper[4754]: I1005 21:31:24.519379 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qxk2d\" (UniqueName: \"kubernetes.io/projected/c26e6366-e15d-4c7b-a370-a601e2cee56f-kube-api-access-qxk2d\") pod \"c26e6366-e15d-4c7b-a370-a601e2cee56f\" (UID: \"c26e6366-e15d-4c7b-a370-a601e2cee56f\") " Oct 05 21:31:24 crc kubenswrapper[4754]: I1005 21:31:24.519469 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c26e6366-e15d-4c7b-a370-a601e2cee56f-inventory\") pod \"c26e6366-e15d-4c7b-a370-a601e2cee56f\" (UID: \"c26e6366-e15d-4c7b-a370-a601e2cee56f\") " Oct 05 21:31:24 crc kubenswrapper[4754]: I1005 21:31:24.519563 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/c26e6366-e15d-4c7b-a370-a601e2cee56f-ovncontroller-config-0\") pod \"c26e6366-e15d-4c7b-a370-a601e2cee56f\" (UID: \"c26e6366-e15d-4c7b-a370-a601e2cee56f\") " Oct 05 21:31:24 crc kubenswrapper[4754]: I1005 21:31:24.519678 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c26e6366-e15d-4c7b-a370-a601e2cee56f-ssh-key\") pod \"c26e6366-e15d-4c7b-a370-a601e2cee56f\" (UID: \"c26e6366-e15d-4c7b-a370-a601e2cee56f\") " Oct 05 21:31:24 crc kubenswrapper[4754]: I1005 21:31:24.527795 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c26e6366-e15d-4c7b-a370-a601e2cee56f-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "c26e6366-e15d-4c7b-a370-a601e2cee56f" (UID: "c26e6366-e15d-4c7b-a370-a601e2cee56f"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:31:24 crc kubenswrapper[4754]: I1005 21:31:24.530742 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c26e6366-e15d-4c7b-a370-a601e2cee56f-kube-api-access-qxk2d" (OuterVolumeSpecName: "kube-api-access-qxk2d") pod "c26e6366-e15d-4c7b-a370-a601e2cee56f" (UID: "c26e6366-e15d-4c7b-a370-a601e2cee56f"). InnerVolumeSpecName "kube-api-access-qxk2d". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:31:24 crc kubenswrapper[4754]: I1005 21:31:24.556679 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c26e6366-e15d-4c7b-a370-a601e2cee56f-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "c26e6366-e15d-4c7b-a370-a601e2cee56f" (UID: "c26e6366-e15d-4c7b-a370-a601e2cee56f"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:31:24 crc kubenswrapper[4754]: I1005 21:31:24.558975 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c26e6366-e15d-4c7b-a370-a601e2cee56f-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "c26e6366-e15d-4c7b-a370-a601e2cee56f" (UID: "c26e6366-e15d-4c7b-a370-a601e2cee56f"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 21:31:24 crc kubenswrapper[4754]: I1005 21:31:24.584422 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c26e6366-e15d-4c7b-a370-a601e2cee56f-inventory" (OuterVolumeSpecName: "inventory") pod "c26e6366-e15d-4c7b-a370-a601e2cee56f" (UID: "c26e6366-e15d-4c7b-a370-a601e2cee56f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:31:24 crc kubenswrapper[4754]: I1005 21:31:24.622719 4754 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c26e6366-e15d-4c7b-a370-a601e2cee56f-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 05 21:31:24 crc kubenswrapper[4754]: I1005 21:31:24.622918 4754 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c26e6366-e15d-4c7b-a370-a601e2cee56f-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 21:31:24 crc kubenswrapper[4754]: I1005 21:31:24.623006 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qxk2d\" (UniqueName: \"kubernetes.io/projected/c26e6366-e15d-4c7b-a370-a601e2cee56f-kube-api-access-qxk2d\") on node \"crc\" DevicePath \"\"" Oct 05 21:31:24 crc kubenswrapper[4754]: I1005 21:31:24.623081 4754 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c26e6366-e15d-4c7b-a370-a601e2cee56f-inventory\") on node \"crc\" DevicePath \"\"" Oct 05 21:31:24 crc kubenswrapper[4754]: I1005 21:31:24.623170 4754 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/c26e6366-e15d-4c7b-a370-a601e2cee56f-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Oct 05 21:31:24 crc kubenswrapper[4754]: I1005 21:31:24.871990 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-cwth8" event={"ID":"c26e6366-e15d-4c7b-a370-a601e2cee56f","Type":"ContainerDied","Data":"f375256c7648bf626d64c7750878170fee5c47f4df7468df0671f9591987b728"} Oct 05 21:31:24 crc kubenswrapper[4754]: I1005 21:31:24.872211 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f375256c7648bf626d64c7750878170fee5c47f4df7468df0671f9591987b728" Oct 05 21:31:24 crc kubenswrapper[4754]: I1005 21:31:24.872126 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-cwth8" Oct 05 21:31:24 crc kubenswrapper[4754]: I1005 21:31:24.997667 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6rhpl"] Oct 05 21:31:24 crc kubenswrapper[4754]: E1005 21:31:24.998095 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0fffea5a-ebbf-4457-b966-0fb29dce3c46" containerName="extract-content" Oct 05 21:31:24 crc kubenswrapper[4754]: I1005 21:31:24.998111 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="0fffea5a-ebbf-4457-b966-0fb29dce3c46" containerName="extract-content" Oct 05 21:31:24 crc kubenswrapper[4754]: E1005 21:31:24.998132 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0fffea5a-ebbf-4457-b966-0fb29dce3c46" containerName="registry-server" Oct 05 21:31:24 crc kubenswrapper[4754]: I1005 21:31:24.998138 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="0fffea5a-ebbf-4457-b966-0fb29dce3c46" containerName="registry-server" Oct 05 21:31:24 crc kubenswrapper[4754]: E1005 21:31:24.998166 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c26e6366-e15d-4c7b-a370-a601e2cee56f" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Oct 05 21:31:24 crc kubenswrapper[4754]: I1005 21:31:24.998172 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="c26e6366-e15d-4c7b-a370-a601e2cee56f" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Oct 05 21:31:24 crc kubenswrapper[4754]: E1005 21:31:24.998187 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0fffea5a-ebbf-4457-b966-0fb29dce3c46" containerName="extract-utilities" Oct 05 21:31:24 crc kubenswrapper[4754]: I1005 21:31:24.998193 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="0fffea5a-ebbf-4457-b966-0fb29dce3c46" containerName="extract-utilities" Oct 05 21:31:24 crc kubenswrapper[4754]: I1005 21:31:24.998375 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="0fffea5a-ebbf-4457-b966-0fb29dce3c46" containerName="registry-server" Oct 05 21:31:24 crc kubenswrapper[4754]: I1005 21:31:24.998389 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="c26e6366-e15d-4c7b-a370-a601e2cee56f" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Oct 05 21:31:24 crc kubenswrapper[4754]: I1005 21:31:24.999083 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6rhpl" Oct 05 21:31:25 crc kubenswrapper[4754]: I1005 21:31:25.003583 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 05 21:31:25 crc kubenswrapper[4754]: I1005 21:31:25.004652 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 05 21:31:25 crc kubenswrapper[4754]: I1005 21:31:25.014591 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Oct 05 21:31:25 crc kubenswrapper[4754]: I1005 21:31:25.014756 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Oct 05 21:31:25 crc kubenswrapper[4754]: I1005 21:31:25.016975 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-v54jd" Oct 05 21:31:25 crc kubenswrapper[4754]: I1005 21:31:25.017279 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 05 21:31:25 crc kubenswrapper[4754]: I1005 21:31:25.017292 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6rhpl"] Oct 05 21:31:25 crc kubenswrapper[4754]: I1005 21:31:25.140190 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dmcfw\" (UniqueName: \"kubernetes.io/projected/61d4efe0-4a19-4b36-9509-67354a6fd537-kube-api-access-dmcfw\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-6rhpl\" (UID: \"61d4efe0-4a19-4b36-9509-67354a6fd537\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6rhpl" Oct 05 21:31:25 crc kubenswrapper[4754]: I1005 21:31:25.140482 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61d4efe0-4a19-4b36-9509-67354a6fd537-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-6rhpl\" (UID: \"61d4efe0-4a19-4b36-9509-67354a6fd537\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6rhpl" Oct 05 21:31:25 crc kubenswrapper[4754]: I1005 21:31:25.140554 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/61d4efe0-4a19-4b36-9509-67354a6fd537-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-6rhpl\" (UID: \"61d4efe0-4a19-4b36-9509-67354a6fd537\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6rhpl" Oct 05 21:31:25 crc kubenswrapper[4754]: I1005 21:31:25.140592 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/61d4efe0-4a19-4b36-9509-67354a6fd537-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-6rhpl\" (UID: \"61d4efe0-4a19-4b36-9509-67354a6fd537\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6rhpl" Oct 05 21:31:25 crc kubenswrapper[4754]: I1005 21:31:25.140672 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/61d4efe0-4a19-4b36-9509-67354a6fd537-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-6rhpl\" (UID: \"61d4efe0-4a19-4b36-9509-67354a6fd537\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6rhpl" Oct 05 21:31:25 crc kubenswrapper[4754]: I1005 21:31:25.140714 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/61d4efe0-4a19-4b36-9509-67354a6fd537-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-6rhpl\" (UID: \"61d4efe0-4a19-4b36-9509-67354a6fd537\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6rhpl" Oct 05 21:31:25 crc kubenswrapper[4754]: I1005 21:31:25.242877 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dmcfw\" (UniqueName: \"kubernetes.io/projected/61d4efe0-4a19-4b36-9509-67354a6fd537-kube-api-access-dmcfw\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-6rhpl\" (UID: \"61d4efe0-4a19-4b36-9509-67354a6fd537\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6rhpl" Oct 05 21:31:25 crc kubenswrapper[4754]: I1005 21:31:25.242954 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61d4efe0-4a19-4b36-9509-67354a6fd537-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-6rhpl\" (UID: \"61d4efe0-4a19-4b36-9509-67354a6fd537\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6rhpl" Oct 05 21:31:25 crc kubenswrapper[4754]: I1005 21:31:25.242997 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/61d4efe0-4a19-4b36-9509-67354a6fd537-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-6rhpl\" (UID: \"61d4efe0-4a19-4b36-9509-67354a6fd537\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6rhpl" Oct 05 21:31:25 crc kubenswrapper[4754]: I1005 21:31:25.243049 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/61d4efe0-4a19-4b36-9509-67354a6fd537-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-6rhpl\" (UID: \"61d4efe0-4a19-4b36-9509-67354a6fd537\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6rhpl" Oct 05 21:31:25 crc kubenswrapper[4754]: I1005 21:31:25.243125 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/61d4efe0-4a19-4b36-9509-67354a6fd537-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-6rhpl\" (UID: \"61d4efe0-4a19-4b36-9509-67354a6fd537\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6rhpl" Oct 05 21:31:25 crc kubenswrapper[4754]: I1005 21:31:25.243188 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/61d4efe0-4a19-4b36-9509-67354a6fd537-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-6rhpl\" (UID: \"61d4efe0-4a19-4b36-9509-67354a6fd537\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6rhpl" Oct 05 21:31:25 crc kubenswrapper[4754]: I1005 21:31:25.248812 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/61d4efe0-4a19-4b36-9509-67354a6fd537-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-6rhpl\" (UID: \"61d4efe0-4a19-4b36-9509-67354a6fd537\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6rhpl" Oct 05 21:31:25 crc kubenswrapper[4754]: I1005 21:31:25.249304 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/61d4efe0-4a19-4b36-9509-67354a6fd537-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-6rhpl\" (UID: \"61d4efe0-4a19-4b36-9509-67354a6fd537\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6rhpl" Oct 05 21:31:25 crc kubenswrapper[4754]: I1005 21:31:25.249928 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/61d4efe0-4a19-4b36-9509-67354a6fd537-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-6rhpl\" (UID: \"61d4efe0-4a19-4b36-9509-67354a6fd537\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6rhpl" Oct 05 21:31:25 crc kubenswrapper[4754]: I1005 21:31:25.250015 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/61d4efe0-4a19-4b36-9509-67354a6fd537-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-6rhpl\" (UID: \"61d4efe0-4a19-4b36-9509-67354a6fd537\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6rhpl" Oct 05 21:31:25 crc kubenswrapper[4754]: I1005 21:31:25.251421 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61d4efe0-4a19-4b36-9509-67354a6fd537-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-6rhpl\" (UID: \"61d4efe0-4a19-4b36-9509-67354a6fd537\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6rhpl" Oct 05 21:31:25 crc kubenswrapper[4754]: I1005 21:31:25.269693 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dmcfw\" (UniqueName: \"kubernetes.io/projected/61d4efe0-4a19-4b36-9509-67354a6fd537-kube-api-access-dmcfw\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-6rhpl\" (UID: \"61d4efe0-4a19-4b36-9509-67354a6fd537\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6rhpl" Oct 05 21:31:25 crc kubenswrapper[4754]: I1005 21:31:25.328831 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6rhpl" Oct 05 21:31:25 crc kubenswrapper[4754]: I1005 21:31:25.956743 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6rhpl"] Oct 05 21:31:26 crc kubenswrapper[4754]: I1005 21:31:26.911912 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6rhpl" event={"ID":"61d4efe0-4a19-4b36-9509-67354a6fd537","Type":"ContainerStarted","Data":"8c732495912b694f4945de0a57b9c77a88596c6b63912704e23cf7c06a64968c"} Oct 05 21:31:26 crc kubenswrapper[4754]: I1005 21:31:26.912249 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6rhpl" event={"ID":"61d4efe0-4a19-4b36-9509-67354a6fd537","Type":"ContainerStarted","Data":"9dddae7244773096fc37130d479b78126a2b79b45af414675832306c305a8049"} Oct 05 21:31:26 crc kubenswrapper[4754]: I1005 21:31:26.944643 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6rhpl" podStartSLOduration=2.482023857 podStartE2EDuration="2.944619691s" podCreationTimestamp="2025-10-05 21:31:24 +0000 UTC" firstStartedPulling="2025-10-05 21:31:25.973033313 +0000 UTC m=+2209.877152063" lastFinishedPulling="2025-10-05 21:31:26.435629147 +0000 UTC m=+2210.339747897" observedRunningTime="2025-10-05 21:31:26.94003043 +0000 UTC m=+2210.844149160" watchObservedRunningTime="2025-10-05 21:31:26.944619691 +0000 UTC m=+2210.848738411" Oct 05 21:31:35 crc kubenswrapper[4754]: I1005 21:31:35.246329 4754 patch_prober.go:28] interesting pod/machine-config-daemon-b2h9k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 21:31:35 crc kubenswrapper[4754]: I1005 21:31:35.248906 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 21:31:46 crc kubenswrapper[4754]: I1005 21:31:46.196428 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-nkdmr"] Oct 05 21:31:46 crc kubenswrapper[4754]: I1005 21:31:46.199442 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nkdmr" Oct 05 21:31:46 crc kubenswrapper[4754]: I1005 21:31:46.217676 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-nkdmr"] Oct 05 21:31:46 crc kubenswrapper[4754]: I1005 21:31:46.265168 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6d5pv\" (UniqueName: \"kubernetes.io/projected/44fbb754-03cb-4f2d-8183-a90b4b750e5f-kube-api-access-6d5pv\") pod \"certified-operators-nkdmr\" (UID: \"44fbb754-03cb-4f2d-8183-a90b4b750e5f\") " pod="openshift-marketplace/certified-operators-nkdmr" Oct 05 21:31:46 crc kubenswrapper[4754]: I1005 21:31:46.265256 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44fbb754-03cb-4f2d-8183-a90b4b750e5f-utilities\") pod \"certified-operators-nkdmr\" (UID: \"44fbb754-03cb-4f2d-8183-a90b4b750e5f\") " pod="openshift-marketplace/certified-operators-nkdmr" Oct 05 21:31:46 crc kubenswrapper[4754]: I1005 21:31:46.265325 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44fbb754-03cb-4f2d-8183-a90b4b750e5f-catalog-content\") pod \"certified-operators-nkdmr\" (UID: \"44fbb754-03cb-4f2d-8183-a90b4b750e5f\") " pod="openshift-marketplace/certified-operators-nkdmr" Oct 05 21:31:46 crc kubenswrapper[4754]: I1005 21:31:46.367226 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6d5pv\" (UniqueName: \"kubernetes.io/projected/44fbb754-03cb-4f2d-8183-a90b4b750e5f-kube-api-access-6d5pv\") pod \"certified-operators-nkdmr\" (UID: \"44fbb754-03cb-4f2d-8183-a90b4b750e5f\") " pod="openshift-marketplace/certified-operators-nkdmr" Oct 05 21:31:46 crc kubenswrapper[4754]: I1005 21:31:46.367279 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44fbb754-03cb-4f2d-8183-a90b4b750e5f-utilities\") pod \"certified-operators-nkdmr\" (UID: \"44fbb754-03cb-4f2d-8183-a90b4b750e5f\") " pod="openshift-marketplace/certified-operators-nkdmr" Oct 05 21:31:46 crc kubenswrapper[4754]: I1005 21:31:46.367370 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44fbb754-03cb-4f2d-8183-a90b4b750e5f-catalog-content\") pod \"certified-operators-nkdmr\" (UID: \"44fbb754-03cb-4f2d-8183-a90b4b750e5f\") " pod="openshift-marketplace/certified-operators-nkdmr" Oct 05 21:31:46 crc kubenswrapper[4754]: I1005 21:31:46.367924 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44fbb754-03cb-4f2d-8183-a90b4b750e5f-utilities\") pod \"certified-operators-nkdmr\" (UID: \"44fbb754-03cb-4f2d-8183-a90b4b750e5f\") " pod="openshift-marketplace/certified-operators-nkdmr" Oct 05 21:31:46 crc kubenswrapper[4754]: I1005 21:31:46.367972 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44fbb754-03cb-4f2d-8183-a90b4b750e5f-catalog-content\") pod \"certified-operators-nkdmr\" (UID: \"44fbb754-03cb-4f2d-8183-a90b4b750e5f\") " pod="openshift-marketplace/certified-operators-nkdmr" Oct 05 21:31:46 crc kubenswrapper[4754]: I1005 21:31:46.392603 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6d5pv\" (UniqueName: \"kubernetes.io/projected/44fbb754-03cb-4f2d-8183-a90b4b750e5f-kube-api-access-6d5pv\") pod \"certified-operators-nkdmr\" (UID: \"44fbb754-03cb-4f2d-8183-a90b4b750e5f\") " pod="openshift-marketplace/certified-operators-nkdmr" Oct 05 21:31:46 crc kubenswrapper[4754]: I1005 21:31:46.520796 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nkdmr" Oct 05 21:31:47 crc kubenswrapper[4754]: I1005 21:31:47.043335 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-nkdmr"] Oct 05 21:31:47 crc kubenswrapper[4754]: I1005 21:31:47.183120 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nkdmr" event={"ID":"44fbb754-03cb-4f2d-8183-a90b4b750e5f","Type":"ContainerStarted","Data":"c898a7ad226827a4e45213a2ac93c582f1e459898b4703e1d5ab391ddfc9cc50"} Oct 05 21:31:48 crc kubenswrapper[4754]: I1005 21:31:48.199551 4754 generic.go:334] "Generic (PLEG): container finished" podID="44fbb754-03cb-4f2d-8183-a90b4b750e5f" containerID="82b67c03e18ff1398bf6fcc2d431c2877369ff205d2d91cac20ba362d9f19994" exitCode=0 Oct 05 21:31:48 crc kubenswrapper[4754]: I1005 21:31:48.199615 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nkdmr" event={"ID":"44fbb754-03cb-4f2d-8183-a90b4b750e5f","Type":"ContainerDied","Data":"82b67c03e18ff1398bf6fcc2d431c2877369ff205d2d91cac20ba362d9f19994"} Oct 05 21:31:49 crc kubenswrapper[4754]: I1005 21:31:49.213399 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nkdmr" event={"ID":"44fbb754-03cb-4f2d-8183-a90b4b750e5f","Type":"ContainerStarted","Data":"c3586619c53deeca588610949a8754dc1355d619012ce36d2941dfbe4d33d635"} Oct 05 21:31:50 crc kubenswrapper[4754]: I1005 21:31:50.231059 4754 generic.go:334] "Generic (PLEG): container finished" podID="44fbb754-03cb-4f2d-8183-a90b4b750e5f" containerID="c3586619c53deeca588610949a8754dc1355d619012ce36d2941dfbe4d33d635" exitCode=0 Oct 05 21:31:50 crc kubenswrapper[4754]: I1005 21:31:50.231134 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nkdmr" event={"ID":"44fbb754-03cb-4f2d-8183-a90b4b750e5f","Type":"ContainerDied","Data":"c3586619c53deeca588610949a8754dc1355d619012ce36d2941dfbe4d33d635"} Oct 05 21:31:51 crc kubenswrapper[4754]: I1005 21:31:51.248881 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nkdmr" event={"ID":"44fbb754-03cb-4f2d-8183-a90b4b750e5f","Type":"ContainerStarted","Data":"7224b6eef97cd87fba763d0168c58fa3cc4a1e3fa558222b0306ed8f3a8e78a8"} Oct 05 21:31:51 crc kubenswrapper[4754]: I1005 21:31:51.283736 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-nkdmr" podStartSLOduration=2.644648805 podStartE2EDuration="5.28370788s" podCreationTimestamp="2025-10-05 21:31:46 +0000 UTC" firstStartedPulling="2025-10-05 21:31:48.203798363 +0000 UTC m=+2232.107917073" lastFinishedPulling="2025-10-05 21:31:50.842857398 +0000 UTC m=+2234.746976148" observedRunningTime="2025-10-05 21:31:51.271359286 +0000 UTC m=+2235.175478006" watchObservedRunningTime="2025-10-05 21:31:51.28370788 +0000 UTC m=+2235.187826630" Oct 05 21:31:56 crc kubenswrapper[4754]: I1005 21:31:56.522122 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-nkdmr" Oct 05 21:31:56 crc kubenswrapper[4754]: I1005 21:31:56.522226 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-nkdmr" Oct 05 21:31:56 crc kubenswrapper[4754]: I1005 21:31:56.608195 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-nkdmr" Oct 05 21:31:57 crc kubenswrapper[4754]: I1005 21:31:57.396459 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-nkdmr" Oct 05 21:31:57 crc kubenswrapper[4754]: I1005 21:31:57.463543 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-nkdmr"] Oct 05 21:31:59 crc kubenswrapper[4754]: I1005 21:31:59.360482 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-nkdmr" podUID="44fbb754-03cb-4f2d-8183-a90b4b750e5f" containerName="registry-server" containerID="cri-o://7224b6eef97cd87fba763d0168c58fa3cc4a1e3fa558222b0306ed8f3a8e78a8" gracePeriod=2 Oct 05 21:31:59 crc kubenswrapper[4754]: I1005 21:31:59.896386 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nkdmr" Oct 05 21:31:59 crc kubenswrapper[4754]: I1005 21:31:59.969034 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44fbb754-03cb-4f2d-8183-a90b4b750e5f-utilities\") pod \"44fbb754-03cb-4f2d-8183-a90b4b750e5f\" (UID: \"44fbb754-03cb-4f2d-8183-a90b4b750e5f\") " Oct 05 21:31:59 crc kubenswrapper[4754]: I1005 21:31:59.969087 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44fbb754-03cb-4f2d-8183-a90b4b750e5f-catalog-content\") pod \"44fbb754-03cb-4f2d-8183-a90b4b750e5f\" (UID: \"44fbb754-03cb-4f2d-8183-a90b4b750e5f\") " Oct 05 21:31:59 crc kubenswrapper[4754]: I1005 21:31:59.969206 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6d5pv\" (UniqueName: \"kubernetes.io/projected/44fbb754-03cb-4f2d-8183-a90b4b750e5f-kube-api-access-6d5pv\") pod \"44fbb754-03cb-4f2d-8183-a90b4b750e5f\" (UID: \"44fbb754-03cb-4f2d-8183-a90b4b750e5f\") " Oct 05 21:31:59 crc kubenswrapper[4754]: I1005 21:31:59.973760 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/44fbb754-03cb-4f2d-8183-a90b4b750e5f-utilities" (OuterVolumeSpecName: "utilities") pod "44fbb754-03cb-4f2d-8183-a90b4b750e5f" (UID: "44fbb754-03cb-4f2d-8183-a90b4b750e5f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 21:31:59 crc kubenswrapper[4754]: I1005 21:31:59.983483 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44fbb754-03cb-4f2d-8183-a90b4b750e5f-kube-api-access-6d5pv" (OuterVolumeSpecName: "kube-api-access-6d5pv") pod "44fbb754-03cb-4f2d-8183-a90b4b750e5f" (UID: "44fbb754-03cb-4f2d-8183-a90b4b750e5f"). InnerVolumeSpecName "kube-api-access-6d5pv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:32:00 crc kubenswrapper[4754]: I1005 21:32:00.028438 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/44fbb754-03cb-4f2d-8183-a90b4b750e5f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "44fbb754-03cb-4f2d-8183-a90b4b750e5f" (UID: "44fbb754-03cb-4f2d-8183-a90b4b750e5f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 21:32:00 crc kubenswrapper[4754]: I1005 21:32:00.073240 4754 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44fbb754-03cb-4f2d-8183-a90b4b750e5f-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 21:32:00 crc kubenswrapper[4754]: I1005 21:32:00.073316 4754 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44fbb754-03cb-4f2d-8183-a90b4b750e5f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 21:32:00 crc kubenswrapper[4754]: I1005 21:32:00.073335 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6d5pv\" (UniqueName: \"kubernetes.io/projected/44fbb754-03cb-4f2d-8183-a90b4b750e5f-kube-api-access-6d5pv\") on node \"crc\" DevicePath \"\"" Oct 05 21:32:00 crc kubenswrapper[4754]: I1005 21:32:00.373332 4754 generic.go:334] "Generic (PLEG): container finished" podID="44fbb754-03cb-4f2d-8183-a90b4b750e5f" containerID="7224b6eef97cd87fba763d0168c58fa3cc4a1e3fa558222b0306ed8f3a8e78a8" exitCode=0 Oct 05 21:32:00 crc kubenswrapper[4754]: I1005 21:32:00.373644 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nkdmr" event={"ID":"44fbb754-03cb-4f2d-8183-a90b4b750e5f","Type":"ContainerDied","Data":"7224b6eef97cd87fba763d0168c58fa3cc4a1e3fa558222b0306ed8f3a8e78a8"} Oct 05 21:32:00 crc kubenswrapper[4754]: I1005 21:32:00.373815 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nkdmr" event={"ID":"44fbb754-03cb-4f2d-8183-a90b4b750e5f","Type":"ContainerDied","Data":"c898a7ad226827a4e45213a2ac93c582f1e459898b4703e1d5ab391ddfc9cc50"} Oct 05 21:32:00 crc kubenswrapper[4754]: I1005 21:32:00.374831 4754 scope.go:117] "RemoveContainer" containerID="7224b6eef97cd87fba763d0168c58fa3cc4a1e3fa558222b0306ed8f3a8e78a8" Oct 05 21:32:00 crc kubenswrapper[4754]: I1005 21:32:00.373740 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nkdmr" Oct 05 21:32:00 crc kubenswrapper[4754]: I1005 21:32:00.406664 4754 scope.go:117] "RemoveContainer" containerID="c3586619c53deeca588610949a8754dc1355d619012ce36d2941dfbe4d33d635" Oct 05 21:32:00 crc kubenswrapper[4754]: I1005 21:32:00.430832 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-nkdmr"] Oct 05 21:32:00 crc kubenswrapper[4754]: I1005 21:32:00.441958 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-nkdmr"] Oct 05 21:32:00 crc kubenswrapper[4754]: I1005 21:32:00.448467 4754 scope.go:117] "RemoveContainer" containerID="82b67c03e18ff1398bf6fcc2d431c2877369ff205d2d91cac20ba362d9f19994" Oct 05 21:32:00 crc kubenswrapper[4754]: I1005 21:32:00.501270 4754 scope.go:117] "RemoveContainer" containerID="7224b6eef97cd87fba763d0168c58fa3cc4a1e3fa558222b0306ed8f3a8e78a8" Oct 05 21:32:00 crc kubenswrapper[4754]: E1005 21:32:00.501890 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7224b6eef97cd87fba763d0168c58fa3cc4a1e3fa558222b0306ed8f3a8e78a8\": container with ID starting with 7224b6eef97cd87fba763d0168c58fa3cc4a1e3fa558222b0306ed8f3a8e78a8 not found: ID does not exist" containerID="7224b6eef97cd87fba763d0168c58fa3cc4a1e3fa558222b0306ed8f3a8e78a8" Oct 05 21:32:00 crc kubenswrapper[4754]: I1005 21:32:00.501937 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7224b6eef97cd87fba763d0168c58fa3cc4a1e3fa558222b0306ed8f3a8e78a8"} err="failed to get container status \"7224b6eef97cd87fba763d0168c58fa3cc4a1e3fa558222b0306ed8f3a8e78a8\": rpc error: code = NotFound desc = could not find container \"7224b6eef97cd87fba763d0168c58fa3cc4a1e3fa558222b0306ed8f3a8e78a8\": container with ID starting with 7224b6eef97cd87fba763d0168c58fa3cc4a1e3fa558222b0306ed8f3a8e78a8 not found: ID does not exist" Oct 05 21:32:00 crc kubenswrapper[4754]: I1005 21:32:00.501975 4754 scope.go:117] "RemoveContainer" containerID="c3586619c53deeca588610949a8754dc1355d619012ce36d2941dfbe4d33d635" Oct 05 21:32:00 crc kubenswrapper[4754]: E1005 21:32:00.502347 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c3586619c53deeca588610949a8754dc1355d619012ce36d2941dfbe4d33d635\": container with ID starting with c3586619c53deeca588610949a8754dc1355d619012ce36d2941dfbe4d33d635 not found: ID does not exist" containerID="c3586619c53deeca588610949a8754dc1355d619012ce36d2941dfbe4d33d635" Oct 05 21:32:00 crc kubenswrapper[4754]: I1005 21:32:00.502382 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c3586619c53deeca588610949a8754dc1355d619012ce36d2941dfbe4d33d635"} err="failed to get container status \"c3586619c53deeca588610949a8754dc1355d619012ce36d2941dfbe4d33d635\": rpc error: code = NotFound desc = could not find container \"c3586619c53deeca588610949a8754dc1355d619012ce36d2941dfbe4d33d635\": container with ID starting with c3586619c53deeca588610949a8754dc1355d619012ce36d2941dfbe4d33d635 not found: ID does not exist" Oct 05 21:32:00 crc kubenswrapper[4754]: I1005 21:32:00.502410 4754 scope.go:117] "RemoveContainer" containerID="82b67c03e18ff1398bf6fcc2d431c2877369ff205d2d91cac20ba362d9f19994" Oct 05 21:32:00 crc kubenswrapper[4754]: E1005 21:32:00.502800 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"82b67c03e18ff1398bf6fcc2d431c2877369ff205d2d91cac20ba362d9f19994\": container with ID starting with 82b67c03e18ff1398bf6fcc2d431c2877369ff205d2d91cac20ba362d9f19994 not found: ID does not exist" containerID="82b67c03e18ff1398bf6fcc2d431c2877369ff205d2d91cac20ba362d9f19994" Oct 05 21:32:00 crc kubenswrapper[4754]: I1005 21:32:00.502864 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"82b67c03e18ff1398bf6fcc2d431c2877369ff205d2d91cac20ba362d9f19994"} err="failed to get container status \"82b67c03e18ff1398bf6fcc2d431c2877369ff205d2d91cac20ba362d9f19994\": rpc error: code = NotFound desc = could not find container \"82b67c03e18ff1398bf6fcc2d431c2877369ff205d2d91cac20ba362d9f19994\": container with ID starting with 82b67c03e18ff1398bf6fcc2d431c2877369ff205d2d91cac20ba362d9f19994 not found: ID does not exist" Oct 05 21:32:00 crc kubenswrapper[4754]: I1005 21:32:00.852257 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44fbb754-03cb-4f2d-8183-a90b4b750e5f" path="/var/lib/kubelet/pods/44fbb754-03cb-4f2d-8183-a90b4b750e5f/volumes" Oct 05 21:32:05 crc kubenswrapper[4754]: I1005 21:32:05.246137 4754 patch_prober.go:28] interesting pod/machine-config-daemon-b2h9k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 21:32:05 crc kubenswrapper[4754]: I1005 21:32:05.246641 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 21:32:05 crc kubenswrapper[4754]: I1005 21:32:05.246710 4754 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" Oct 05 21:32:05 crc kubenswrapper[4754]: I1005 21:32:05.248043 4754 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0df70ace610d977d6937bca9a99a0d188993dfa6fcd8015008efc0870430fdb3"} pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 05 21:32:05 crc kubenswrapper[4754]: I1005 21:32:05.248145 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" containerName="machine-config-daemon" containerID="cri-o://0df70ace610d977d6937bca9a99a0d188993dfa6fcd8015008efc0870430fdb3" gracePeriod=600 Oct 05 21:32:05 crc kubenswrapper[4754]: E1005 21:32:05.413989 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 21:32:05 crc kubenswrapper[4754]: I1005 21:32:05.456811 4754 generic.go:334] "Generic (PLEG): container finished" podID="be95a413-6a59-45b4-84b7-b43cae694a26" containerID="0df70ace610d977d6937bca9a99a0d188993dfa6fcd8015008efc0870430fdb3" exitCode=0 Oct 05 21:32:05 crc kubenswrapper[4754]: I1005 21:32:05.456873 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" event={"ID":"be95a413-6a59-45b4-84b7-b43cae694a26","Type":"ContainerDied","Data":"0df70ace610d977d6937bca9a99a0d188993dfa6fcd8015008efc0870430fdb3"} Oct 05 21:32:05 crc kubenswrapper[4754]: I1005 21:32:05.456927 4754 scope.go:117] "RemoveContainer" containerID="fee409256f2bba9979dd69800f220c1257ab9bdb2e760315daae28edefc3f914" Oct 05 21:32:05 crc kubenswrapper[4754]: I1005 21:32:05.458063 4754 scope.go:117] "RemoveContainer" containerID="0df70ace610d977d6937bca9a99a0d188993dfa6fcd8015008efc0870430fdb3" Oct 05 21:32:05 crc kubenswrapper[4754]: E1005 21:32:05.458615 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 21:32:19 crc kubenswrapper[4754]: I1005 21:32:19.838709 4754 scope.go:117] "RemoveContainer" containerID="0df70ace610d977d6937bca9a99a0d188993dfa6fcd8015008efc0870430fdb3" Oct 05 21:32:19 crc kubenswrapper[4754]: E1005 21:32:19.840065 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 21:32:26 crc kubenswrapper[4754]: I1005 21:32:26.692462 4754 generic.go:334] "Generic (PLEG): container finished" podID="61d4efe0-4a19-4b36-9509-67354a6fd537" containerID="8c732495912b694f4945de0a57b9c77a88596c6b63912704e23cf7c06a64968c" exitCode=0 Oct 05 21:32:26 crc kubenswrapper[4754]: I1005 21:32:26.692621 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6rhpl" event={"ID":"61d4efe0-4a19-4b36-9509-67354a6fd537","Type":"ContainerDied","Data":"8c732495912b694f4945de0a57b9c77a88596c6b63912704e23cf7c06a64968c"} Oct 05 21:32:28 crc kubenswrapper[4754]: I1005 21:32:28.144647 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6rhpl" Oct 05 21:32:28 crc kubenswrapper[4754]: I1005 21:32:28.205989 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/61d4efe0-4a19-4b36-9509-67354a6fd537-nova-metadata-neutron-config-0\") pod \"61d4efe0-4a19-4b36-9509-67354a6fd537\" (UID: \"61d4efe0-4a19-4b36-9509-67354a6fd537\") " Oct 05 21:32:28 crc kubenswrapper[4754]: I1005 21:32:28.206188 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/61d4efe0-4a19-4b36-9509-67354a6fd537-inventory\") pod \"61d4efe0-4a19-4b36-9509-67354a6fd537\" (UID: \"61d4efe0-4a19-4b36-9509-67354a6fd537\") " Oct 05 21:32:28 crc kubenswrapper[4754]: I1005 21:32:28.206303 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61d4efe0-4a19-4b36-9509-67354a6fd537-neutron-metadata-combined-ca-bundle\") pod \"61d4efe0-4a19-4b36-9509-67354a6fd537\" (UID: \"61d4efe0-4a19-4b36-9509-67354a6fd537\") " Oct 05 21:32:28 crc kubenswrapper[4754]: I1005 21:32:28.206524 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/61d4efe0-4a19-4b36-9509-67354a6fd537-neutron-ovn-metadata-agent-neutron-config-0\") pod \"61d4efe0-4a19-4b36-9509-67354a6fd537\" (UID: \"61d4efe0-4a19-4b36-9509-67354a6fd537\") " Oct 05 21:32:28 crc kubenswrapper[4754]: I1005 21:32:28.206570 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/61d4efe0-4a19-4b36-9509-67354a6fd537-ssh-key\") pod \"61d4efe0-4a19-4b36-9509-67354a6fd537\" (UID: \"61d4efe0-4a19-4b36-9509-67354a6fd537\") " Oct 05 21:32:28 crc kubenswrapper[4754]: I1005 21:32:28.206730 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dmcfw\" (UniqueName: \"kubernetes.io/projected/61d4efe0-4a19-4b36-9509-67354a6fd537-kube-api-access-dmcfw\") pod \"61d4efe0-4a19-4b36-9509-67354a6fd537\" (UID: \"61d4efe0-4a19-4b36-9509-67354a6fd537\") " Oct 05 21:32:28 crc kubenswrapper[4754]: I1005 21:32:28.226154 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/61d4efe0-4a19-4b36-9509-67354a6fd537-kube-api-access-dmcfw" (OuterVolumeSpecName: "kube-api-access-dmcfw") pod "61d4efe0-4a19-4b36-9509-67354a6fd537" (UID: "61d4efe0-4a19-4b36-9509-67354a6fd537"). InnerVolumeSpecName "kube-api-access-dmcfw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:32:28 crc kubenswrapper[4754]: I1005 21:32:28.239914 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/61d4efe0-4a19-4b36-9509-67354a6fd537-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "61d4efe0-4a19-4b36-9509-67354a6fd537" (UID: "61d4efe0-4a19-4b36-9509-67354a6fd537"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:32:28 crc kubenswrapper[4754]: I1005 21:32:28.269778 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/61d4efe0-4a19-4b36-9509-67354a6fd537-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "61d4efe0-4a19-4b36-9509-67354a6fd537" (UID: "61d4efe0-4a19-4b36-9509-67354a6fd537"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:32:28 crc kubenswrapper[4754]: I1005 21:32:28.313346 4754 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61d4efe0-4a19-4b36-9509-67354a6fd537-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 21:32:28 crc kubenswrapper[4754]: I1005 21:32:28.313378 4754 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/61d4efe0-4a19-4b36-9509-67354a6fd537-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Oct 05 21:32:28 crc kubenswrapper[4754]: I1005 21:32:28.313390 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dmcfw\" (UniqueName: \"kubernetes.io/projected/61d4efe0-4a19-4b36-9509-67354a6fd537-kube-api-access-dmcfw\") on node \"crc\" DevicePath \"\"" Oct 05 21:32:28 crc kubenswrapper[4754]: I1005 21:32:28.357684 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/61d4efe0-4a19-4b36-9509-67354a6fd537-inventory" (OuterVolumeSpecName: "inventory") pod "61d4efe0-4a19-4b36-9509-67354a6fd537" (UID: "61d4efe0-4a19-4b36-9509-67354a6fd537"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:32:28 crc kubenswrapper[4754]: I1005 21:32:28.359619 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/61d4efe0-4a19-4b36-9509-67354a6fd537-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "61d4efe0-4a19-4b36-9509-67354a6fd537" (UID: "61d4efe0-4a19-4b36-9509-67354a6fd537"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:32:28 crc kubenswrapper[4754]: I1005 21:32:28.361715 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/61d4efe0-4a19-4b36-9509-67354a6fd537-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "61d4efe0-4a19-4b36-9509-67354a6fd537" (UID: "61d4efe0-4a19-4b36-9509-67354a6fd537"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:32:28 crc kubenswrapper[4754]: I1005 21:32:28.415942 4754 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/61d4efe0-4a19-4b36-9509-67354a6fd537-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 05 21:32:28 crc kubenswrapper[4754]: I1005 21:32:28.415984 4754 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/61d4efe0-4a19-4b36-9509-67354a6fd537-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Oct 05 21:32:28 crc kubenswrapper[4754]: I1005 21:32:28.415999 4754 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/61d4efe0-4a19-4b36-9509-67354a6fd537-inventory\") on node \"crc\" DevicePath \"\"" Oct 05 21:32:28 crc kubenswrapper[4754]: I1005 21:32:28.716621 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6rhpl" event={"ID":"61d4efe0-4a19-4b36-9509-67354a6fd537","Type":"ContainerDied","Data":"9dddae7244773096fc37130d479b78126a2b79b45af414675832306c305a8049"} Oct 05 21:32:28 crc kubenswrapper[4754]: I1005 21:32:28.716668 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9dddae7244773096fc37130d479b78126a2b79b45af414675832306c305a8049" Oct 05 21:32:28 crc kubenswrapper[4754]: I1005 21:32:28.716731 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6rhpl" Oct 05 21:32:28 crc kubenswrapper[4754]: I1005 21:32:28.832383 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-c7xwk"] Oct 05 21:32:28 crc kubenswrapper[4754]: E1005 21:32:28.832806 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44fbb754-03cb-4f2d-8183-a90b4b750e5f" containerName="registry-server" Oct 05 21:32:28 crc kubenswrapper[4754]: I1005 21:32:28.832824 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="44fbb754-03cb-4f2d-8183-a90b4b750e5f" containerName="registry-server" Oct 05 21:32:28 crc kubenswrapper[4754]: E1005 21:32:28.832854 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44fbb754-03cb-4f2d-8183-a90b4b750e5f" containerName="extract-utilities" Oct 05 21:32:28 crc kubenswrapper[4754]: I1005 21:32:28.832860 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="44fbb754-03cb-4f2d-8183-a90b4b750e5f" containerName="extract-utilities" Oct 05 21:32:28 crc kubenswrapper[4754]: E1005 21:32:28.832867 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61d4efe0-4a19-4b36-9509-67354a6fd537" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Oct 05 21:32:28 crc kubenswrapper[4754]: I1005 21:32:28.832875 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="61d4efe0-4a19-4b36-9509-67354a6fd537" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Oct 05 21:32:28 crc kubenswrapper[4754]: E1005 21:32:28.832890 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44fbb754-03cb-4f2d-8183-a90b4b750e5f" containerName="extract-content" Oct 05 21:32:28 crc kubenswrapper[4754]: I1005 21:32:28.832896 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="44fbb754-03cb-4f2d-8183-a90b4b750e5f" containerName="extract-content" Oct 05 21:32:28 crc kubenswrapper[4754]: I1005 21:32:28.833074 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="44fbb754-03cb-4f2d-8183-a90b4b750e5f" containerName="registry-server" Oct 05 21:32:28 crc kubenswrapper[4754]: I1005 21:32:28.833085 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="61d4efe0-4a19-4b36-9509-67354a6fd537" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Oct 05 21:32:28 crc kubenswrapper[4754]: I1005 21:32:28.833771 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-c7xwk" Oct 05 21:32:28 crc kubenswrapper[4754]: I1005 21:32:28.838604 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 05 21:32:28 crc kubenswrapper[4754]: I1005 21:32:28.838792 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 05 21:32:28 crc kubenswrapper[4754]: I1005 21:32:28.838937 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Oct 05 21:32:28 crc kubenswrapper[4754]: I1005 21:32:28.839070 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-v54jd" Oct 05 21:32:28 crc kubenswrapper[4754]: I1005 21:32:28.839194 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 05 21:32:28 crc kubenswrapper[4754]: I1005 21:32:28.853782 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-c7xwk"] Oct 05 21:32:28 crc kubenswrapper[4754]: I1005 21:32:28.926351 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/43f1b88f-48f4-4ee0-9fcd-8f53bee2f461-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-c7xwk\" (UID: \"43f1b88f-48f4-4ee0-9fcd-8f53bee2f461\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-c7xwk" Oct 05 21:32:28 crc kubenswrapper[4754]: I1005 21:32:28.926400 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7pkl8\" (UniqueName: \"kubernetes.io/projected/43f1b88f-48f4-4ee0-9fcd-8f53bee2f461-kube-api-access-7pkl8\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-c7xwk\" (UID: \"43f1b88f-48f4-4ee0-9fcd-8f53bee2f461\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-c7xwk" Oct 05 21:32:28 crc kubenswrapper[4754]: I1005 21:32:28.926710 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/43f1b88f-48f4-4ee0-9fcd-8f53bee2f461-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-c7xwk\" (UID: \"43f1b88f-48f4-4ee0-9fcd-8f53bee2f461\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-c7xwk" Oct 05 21:32:28 crc kubenswrapper[4754]: I1005 21:32:28.926837 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/43f1b88f-48f4-4ee0-9fcd-8f53bee2f461-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-c7xwk\" (UID: \"43f1b88f-48f4-4ee0-9fcd-8f53bee2f461\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-c7xwk" Oct 05 21:32:28 crc kubenswrapper[4754]: I1005 21:32:28.927203 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43f1b88f-48f4-4ee0-9fcd-8f53bee2f461-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-c7xwk\" (UID: \"43f1b88f-48f4-4ee0-9fcd-8f53bee2f461\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-c7xwk" Oct 05 21:32:29 crc kubenswrapper[4754]: I1005 21:32:29.029326 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/43f1b88f-48f4-4ee0-9fcd-8f53bee2f461-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-c7xwk\" (UID: \"43f1b88f-48f4-4ee0-9fcd-8f53bee2f461\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-c7xwk" Oct 05 21:32:29 crc kubenswrapper[4754]: I1005 21:32:29.029373 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7pkl8\" (UniqueName: \"kubernetes.io/projected/43f1b88f-48f4-4ee0-9fcd-8f53bee2f461-kube-api-access-7pkl8\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-c7xwk\" (UID: \"43f1b88f-48f4-4ee0-9fcd-8f53bee2f461\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-c7xwk" Oct 05 21:32:29 crc kubenswrapper[4754]: I1005 21:32:29.029452 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/43f1b88f-48f4-4ee0-9fcd-8f53bee2f461-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-c7xwk\" (UID: \"43f1b88f-48f4-4ee0-9fcd-8f53bee2f461\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-c7xwk" Oct 05 21:32:29 crc kubenswrapper[4754]: I1005 21:32:29.029515 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/43f1b88f-48f4-4ee0-9fcd-8f53bee2f461-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-c7xwk\" (UID: \"43f1b88f-48f4-4ee0-9fcd-8f53bee2f461\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-c7xwk" Oct 05 21:32:29 crc kubenswrapper[4754]: I1005 21:32:29.029594 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43f1b88f-48f4-4ee0-9fcd-8f53bee2f461-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-c7xwk\" (UID: \"43f1b88f-48f4-4ee0-9fcd-8f53bee2f461\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-c7xwk" Oct 05 21:32:29 crc kubenswrapper[4754]: I1005 21:32:29.034250 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/43f1b88f-48f4-4ee0-9fcd-8f53bee2f461-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-c7xwk\" (UID: \"43f1b88f-48f4-4ee0-9fcd-8f53bee2f461\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-c7xwk" Oct 05 21:32:29 crc kubenswrapper[4754]: I1005 21:32:29.040106 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/43f1b88f-48f4-4ee0-9fcd-8f53bee2f461-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-c7xwk\" (UID: \"43f1b88f-48f4-4ee0-9fcd-8f53bee2f461\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-c7xwk" Oct 05 21:32:29 crc kubenswrapper[4754]: I1005 21:32:29.041390 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43f1b88f-48f4-4ee0-9fcd-8f53bee2f461-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-c7xwk\" (UID: \"43f1b88f-48f4-4ee0-9fcd-8f53bee2f461\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-c7xwk" Oct 05 21:32:29 crc kubenswrapper[4754]: I1005 21:32:29.041475 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/43f1b88f-48f4-4ee0-9fcd-8f53bee2f461-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-c7xwk\" (UID: \"43f1b88f-48f4-4ee0-9fcd-8f53bee2f461\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-c7xwk" Oct 05 21:32:29 crc kubenswrapper[4754]: I1005 21:32:29.067095 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7pkl8\" (UniqueName: \"kubernetes.io/projected/43f1b88f-48f4-4ee0-9fcd-8f53bee2f461-kube-api-access-7pkl8\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-c7xwk\" (UID: \"43f1b88f-48f4-4ee0-9fcd-8f53bee2f461\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-c7xwk" Oct 05 21:32:29 crc kubenswrapper[4754]: I1005 21:32:29.161877 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-c7xwk" Oct 05 21:32:29 crc kubenswrapper[4754]: I1005 21:32:29.810065 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-c7xwk"] Oct 05 21:32:30 crc kubenswrapper[4754]: I1005 21:32:30.738524 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-c7xwk" event={"ID":"43f1b88f-48f4-4ee0-9fcd-8f53bee2f461","Type":"ContainerStarted","Data":"50ced9777fb51b87e2be86831b06d55b6f09a404f94d6c63e4134e18ee5fb0d7"} Oct 05 21:32:31 crc kubenswrapper[4754]: I1005 21:32:31.752489 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-c7xwk" event={"ID":"43f1b88f-48f4-4ee0-9fcd-8f53bee2f461","Type":"ContainerStarted","Data":"32e36d70d520718e3d13626c162110d89b2f80c7247126492987963f51721e5e"} Oct 05 21:32:31 crc kubenswrapper[4754]: I1005 21:32:31.779830 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-c7xwk" podStartSLOduration=3.049763522 podStartE2EDuration="3.779810454s" podCreationTimestamp="2025-10-05 21:32:28 +0000 UTC" firstStartedPulling="2025-10-05 21:32:29.803508926 +0000 UTC m=+2273.707627646" lastFinishedPulling="2025-10-05 21:32:30.533555868 +0000 UTC m=+2274.437674578" observedRunningTime="2025-10-05 21:32:31.772715368 +0000 UTC m=+2275.676834088" watchObservedRunningTime="2025-10-05 21:32:31.779810454 +0000 UTC m=+2275.683929174" Oct 05 21:32:32 crc kubenswrapper[4754]: I1005 21:32:32.837824 4754 scope.go:117] "RemoveContainer" containerID="0df70ace610d977d6937bca9a99a0d188993dfa6fcd8015008efc0870430fdb3" Oct 05 21:32:32 crc kubenswrapper[4754]: E1005 21:32:32.838271 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 21:32:44 crc kubenswrapper[4754]: I1005 21:32:44.838262 4754 scope.go:117] "RemoveContainer" containerID="0df70ace610d977d6937bca9a99a0d188993dfa6fcd8015008efc0870430fdb3" Oct 05 21:32:44 crc kubenswrapper[4754]: E1005 21:32:44.839399 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 21:32:57 crc kubenswrapper[4754]: I1005 21:32:57.838109 4754 scope.go:117] "RemoveContainer" containerID="0df70ace610d977d6937bca9a99a0d188993dfa6fcd8015008efc0870430fdb3" Oct 05 21:32:57 crc kubenswrapper[4754]: E1005 21:32:57.839374 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 21:33:08 crc kubenswrapper[4754]: I1005 21:33:08.841554 4754 scope.go:117] "RemoveContainer" containerID="0df70ace610d977d6937bca9a99a0d188993dfa6fcd8015008efc0870430fdb3" Oct 05 21:33:08 crc kubenswrapper[4754]: E1005 21:33:08.842553 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 21:33:22 crc kubenswrapper[4754]: I1005 21:33:22.838936 4754 scope.go:117] "RemoveContainer" containerID="0df70ace610d977d6937bca9a99a0d188993dfa6fcd8015008efc0870430fdb3" Oct 05 21:33:22 crc kubenswrapper[4754]: E1005 21:33:22.840537 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 21:33:34 crc kubenswrapper[4754]: I1005 21:33:34.837294 4754 scope.go:117] "RemoveContainer" containerID="0df70ace610d977d6937bca9a99a0d188993dfa6fcd8015008efc0870430fdb3" Oct 05 21:33:34 crc kubenswrapper[4754]: E1005 21:33:34.838100 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 21:33:45 crc kubenswrapper[4754]: I1005 21:33:45.838344 4754 scope.go:117] "RemoveContainer" containerID="0df70ace610d977d6937bca9a99a0d188993dfa6fcd8015008efc0870430fdb3" Oct 05 21:33:45 crc kubenswrapper[4754]: E1005 21:33:45.839816 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 21:33:58 crc kubenswrapper[4754]: I1005 21:33:58.837622 4754 scope.go:117] "RemoveContainer" containerID="0df70ace610d977d6937bca9a99a0d188993dfa6fcd8015008efc0870430fdb3" Oct 05 21:33:58 crc kubenswrapper[4754]: E1005 21:33:58.838425 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 21:34:11 crc kubenswrapper[4754]: I1005 21:34:11.837460 4754 scope.go:117] "RemoveContainer" containerID="0df70ace610d977d6937bca9a99a0d188993dfa6fcd8015008efc0870430fdb3" Oct 05 21:34:11 crc kubenswrapper[4754]: E1005 21:34:11.838668 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 21:34:25 crc kubenswrapper[4754]: I1005 21:34:25.838065 4754 scope.go:117] "RemoveContainer" containerID="0df70ace610d977d6937bca9a99a0d188993dfa6fcd8015008efc0870430fdb3" Oct 05 21:34:25 crc kubenswrapper[4754]: E1005 21:34:25.839095 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 21:34:38 crc kubenswrapper[4754]: I1005 21:34:38.847053 4754 scope.go:117] "RemoveContainer" containerID="0df70ace610d977d6937bca9a99a0d188993dfa6fcd8015008efc0870430fdb3" Oct 05 21:34:38 crc kubenswrapper[4754]: E1005 21:34:38.848837 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 21:34:51 crc kubenswrapper[4754]: I1005 21:34:51.837730 4754 scope.go:117] "RemoveContainer" containerID="0df70ace610d977d6937bca9a99a0d188993dfa6fcd8015008efc0870430fdb3" Oct 05 21:34:51 crc kubenswrapper[4754]: E1005 21:34:51.838437 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 21:35:03 crc kubenswrapper[4754]: I1005 21:35:03.837917 4754 scope.go:117] "RemoveContainer" containerID="0df70ace610d977d6937bca9a99a0d188993dfa6fcd8015008efc0870430fdb3" Oct 05 21:35:03 crc kubenswrapper[4754]: E1005 21:35:03.838726 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 21:35:14 crc kubenswrapper[4754]: I1005 21:35:14.838885 4754 scope.go:117] "RemoveContainer" containerID="0df70ace610d977d6937bca9a99a0d188993dfa6fcd8015008efc0870430fdb3" Oct 05 21:35:14 crc kubenswrapper[4754]: E1005 21:35:14.840016 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 21:35:28 crc kubenswrapper[4754]: I1005 21:35:28.838821 4754 scope.go:117] "RemoveContainer" containerID="0df70ace610d977d6937bca9a99a0d188993dfa6fcd8015008efc0870430fdb3" Oct 05 21:35:28 crc kubenswrapper[4754]: E1005 21:35:28.840305 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 21:35:39 crc kubenswrapper[4754]: I1005 21:35:39.838143 4754 scope.go:117] "RemoveContainer" containerID="0df70ace610d977d6937bca9a99a0d188993dfa6fcd8015008efc0870430fdb3" Oct 05 21:35:39 crc kubenswrapper[4754]: E1005 21:35:39.839339 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 21:35:51 crc kubenswrapper[4754]: I1005 21:35:51.837549 4754 scope.go:117] "RemoveContainer" containerID="0df70ace610d977d6937bca9a99a0d188993dfa6fcd8015008efc0870430fdb3" Oct 05 21:35:51 crc kubenswrapper[4754]: E1005 21:35:51.838779 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 21:36:03 crc kubenswrapper[4754]: I1005 21:36:03.838724 4754 scope.go:117] "RemoveContainer" containerID="0df70ace610d977d6937bca9a99a0d188993dfa6fcd8015008efc0870430fdb3" Oct 05 21:36:03 crc kubenswrapper[4754]: E1005 21:36:03.840561 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 21:36:15 crc kubenswrapper[4754]: I1005 21:36:15.838412 4754 scope.go:117] "RemoveContainer" containerID="0df70ace610d977d6937bca9a99a0d188993dfa6fcd8015008efc0870430fdb3" Oct 05 21:36:15 crc kubenswrapper[4754]: E1005 21:36:15.840580 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 21:36:27 crc kubenswrapper[4754]: I1005 21:36:27.857060 4754 scope.go:117] "RemoveContainer" containerID="0df70ace610d977d6937bca9a99a0d188993dfa6fcd8015008efc0870430fdb3" Oct 05 21:36:27 crc kubenswrapper[4754]: E1005 21:36:27.863036 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 21:36:39 crc kubenswrapper[4754]: I1005 21:36:39.837461 4754 scope.go:117] "RemoveContainer" containerID="0df70ace610d977d6937bca9a99a0d188993dfa6fcd8015008efc0870430fdb3" Oct 05 21:36:39 crc kubenswrapper[4754]: E1005 21:36:39.838303 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 21:36:52 crc kubenswrapper[4754]: I1005 21:36:52.837390 4754 scope.go:117] "RemoveContainer" containerID="0df70ace610d977d6937bca9a99a0d188993dfa6fcd8015008efc0870430fdb3" Oct 05 21:36:52 crc kubenswrapper[4754]: E1005 21:36:52.838103 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 21:37:06 crc kubenswrapper[4754]: I1005 21:37:06.844734 4754 scope.go:117] "RemoveContainer" containerID="0df70ace610d977d6937bca9a99a0d188993dfa6fcd8015008efc0870430fdb3" Oct 05 21:37:08 crc kubenswrapper[4754]: I1005 21:37:08.069030 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" event={"ID":"be95a413-6a59-45b4-84b7-b43cae694a26","Type":"ContainerStarted","Data":"5134bc33c72b286f6d3cee42c9ef42b9498b3009c16d4989384d741eded7eda2"} Oct 05 21:37:22 crc kubenswrapper[4754]: I1005 21:37:22.207544 4754 generic.go:334] "Generic (PLEG): container finished" podID="43f1b88f-48f4-4ee0-9fcd-8f53bee2f461" containerID="32e36d70d520718e3d13626c162110d89b2f80c7247126492987963f51721e5e" exitCode=0 Oct 05 21:37:22 crc kubenswrapper[4754]: I1005 21:37:22.207584 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-c7xwk" event={"ID":"43f1b88f-48f4-4ee0-9fcd-8f53bee2f461","Type":"ContainerDied","Data":"32e36d70d520718e3d13626c162110d89b2f80c7247126492987963f51721e5e"} Oct 05 21:37:23 crc kubenswrapper[4754]: I1005 21:37:23.709380 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-c7xwk" Oct 05 21:37:23 crc kubenswrapper[4754]: I1005 21:37:23.812600 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43f1b88f-48f4-4ee0-9fcd-8f53bee2f461-libvirt-combined-ca-bundle\") pod \"43f1b88f-48f4-4ee0-9fcd-8f53bee2f461\" (UID: \"43f1b88f-48f4-4ee0-9fcd-8f53bee2f461\") " Oct 05 21:37:23 crc kubenswrapper[4754]: I1005 21:37:23.812661 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/43f1b88f-48f4-4ee0-9fcd-8f53bee2f461-libvirt-secret-0\") pod \"43f1b88f-48f4-4ee0-9fcd-8f53bee2f461\" (UID: \"43f1b88f-48f4-4ee0-9fcd-8f53bee2f461\") " Oct 05 21:37:23 crc kubenswrapper[4754]: I1005 21:37:23.812723 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/43f1b88f-48f4-4ee0-9fcd-8f53bee2f461-ssh-key\") pod \"43f1b88f-48f4-4ee0-9fcd-8f53bee2f461\" (UID: \"43f1b88f-48f4-4ee0-9fcd-8f53bee2f461\") " Oct 05 21:37:23 crc kubenswrapper[4754]: I1005 21:37:23.812874 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/43f1b88f-48f4-4ee0-9fcd-8f53bee2f461-inventory\") pod \"43f1b88f-48f4-4ee0-9fcd-8f53bee2f461\" (UID: \"43f1b88f-48f4-4ee0-9fcd-8f53bee2f461\") " Oct 05 21:37:23 crc kubenswrapper[4754]: I1005 21:37:23.812945 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7pkl8\" (UniqueName: \"kubernetes.io/projected/43f1b88f-48f4-4ee0-9fcd-8f53bee2f461-kube-api-access-7pkl8\") pod \"43f1b88f-48f4-4ee0-9fcd-8f53bee2f461\" (UID: \"43f1b88f-48f4-4ee0-9fcd-8f53bee2f461\") " Oct 05 21:37:23 crc kubenswrapper[4754]: I1005 21:37:23.819656 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43f1b88f-48f4-4ee0-9fcd-8f53bee2f461-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "43f1b88f-48f4-4ee0-9fcd-8f53bee2f461" (UID: "43f1b88f-48f4-4ee0-9fcd-8f53bee2f461"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:37:23 crc kubenswrapper[4754]: I1005 21:37:23.825748 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43f1b88f-48f4-4ee0-9fcd-8f53bee2f461-kube-api-access-7pkl8" (OuterVolumeSpecName: "kube-api-access-7pkl8") pod "43f1b88f-48f4-4ee0-9fcd-8f53bee2f461" (UID: "43f1b88f-48f4-4ee0-9fcd-8f53bee2f461"). InnerVolumeSpecName "kube-api-access-7pkl8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:37:23 crc kubenswrapper[4754]: I1005 21:37:23.838914 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43f1b88f-48f4-4ee0-9fcd-8f53bee2f461-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "43f1b88f-48f4-4ee0-9fcd-8f53bee2f461" (UID: "43f1b88f-48f4-4ee0-9fcd-8f53bee2f461"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:37:23 crc kubenswrapper[4754]: I1005 21:37:23.847578 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43f1b88f-48f4-4ee0-9fcd-8f53bee2f461-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "43f1b88f-48f4-4ee0-9fcd-8f53bee2f461" (UID: "43f1b88f-48f4-4ee0-9fcd-8f53bee2f461"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:37:23 crc kubenswrapper[4754]: I1005 21:37:23.853754 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43f1b88f-48f4-4ee0-9fcd-8f53bee2f461-inventory" (OuterVolumeSpecName: "inventory") pod "43f1b88f-48f4-4ee0-9fcd-8f53bee2f461" (UID: "43f1b88f-48f4-4ee0-9fcd-8f53bee2f461"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:37:23 crc kubenswrapper[4754]: I1005 21:37:23.915308 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7pkl8\" (UniqueName: \"kubernetes.io/projected/43f1b88f-48f4-4ee0-9fcd-8f53bee2f461-kube-api-access-7pkl8\") on node \"crc\" DevicePath \"\"" Oct 05 21:37:23 crc kubenswrapper[4754]: I1005 21:37:23.915413 4754 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43f1b88f-48f4-4ee0-9fcd-8f53bee2f461-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 21:37:23 crc kubenswrapper[4754]: I1005 21:37:23.915486 4754 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/43f1b88f-48f4-4ee0-9fcd-8f53bee2f461-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Oct 05 21:37:23 crc kubenswrapper[4754]: I1005 21:37:23.915575 4754 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/43f1b88f-48f4-4ee0-9fcd-8f53bee2f461-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 05 21:37:23 crc kubenswrapper[4754]: I1005 21:37:23.915595 4754 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/43f1b88f-48f4-4ee0-9fcd-8f53bee2f461-inventory\") on node \"crc\" DevicePath \"\"" Oct 05 21:37:24 crc kubenswrapper[4754]: I1005 21:37:24.234251 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-c7xwk" event={"ID":"43f1b88f-48f4-4ee0-9fcd-8f53bee2f461","Type":"ContainerDied","Data":"50ced9777fb51b87e2be86831b06d55b6f09a404f94d6c63e4134e18ee5fb0d7"} Oct 05 21:37:24 crc kubenswrapper[4754]: I1005 21:37:24.234443 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="50ced9777fb51b87e2be86831b06d55b6f09a404f94d6c63e4134e18ee5fb0d7" Oct 05 21:37:24 crc kubenswrapper[4754]: I1005 21:37:24.234370 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-c7xwk" Oct 05 21:37:24 crc kubenswrapper[4754]: I1005 21:37:24.475294 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-fx4cg"] Oct 05 21:37:24 crc kubenswrapper[4754]: E1005 21:37:24.476084 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43f1b88f-48f4-4ee0-9fcd-8f53bee2f461" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Oct 05 21:37:24 crc kubenswrapper[4754]: I1005 21:37:24.476127 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="43f1b88f-48f4-4ee0-9fcd-8f53bee2f461" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Oct 05 21:37:24 crc kubenswrapper[4754]: I1005 21:37:24.476653 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="43f1b88f-48f4-4ee0-9fcd-8f53bee2f461" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Oct 05 21:37:24 crc kubenswrapper[4754]: I1005 21:37:24.477551 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-fx4cg" Oct 05 21:37:24 crc kubenswrapper[4754]: I1005 21:37:24.480289 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 05 21:37:24 crc kubenswrapper[4754]: I1005 21:37:24.481432 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 05 21:37:24 crc kubenswrapper[4754]: I1005 21:37:24.481916 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Oct 05 21:37:24 crc kubenswrapper[4754]: I1005 21:37:24.482385 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-v54jd" Oct 05 21:37:24 crc kubenswrapper[4754]: I1005 21:37:24.482871 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 05 21:37:24 crc kubenswrapper[4754]: I1005 21:37:24.489987 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Oct 05 21:37:24 crc kubenswrapper[4754]: I1005 21:37:24.490028 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Oct 05 21:37:24 crc kubenswrapper[4754]: I1005 21:37:24.528141 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-fx4cg"] Oct 05 21:37:24 crc kubenswrapper[4754]: I1005 21:37:24.528754 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/542410e6-102a-4e41-b3a5-528f37501180-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-fx4cg\" (UID: \"542410e6-102a-4e41-b3a5-528f37501180\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-fx4cg" Oct 05 21:37:24 crc kubenswrapper[4754]: I1005 21:37:24.528931 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/542410e6-102a-4e41-b3a5-528f37501180-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-fx4cg\" (UID: \"542410e6-102a-4e41-b3a5-528f37501180\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-fx4cg" Oct 05 21:37:24 crc kubenswrapper[4754]: I1005 21:37:24.529015 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/542410e6-102a-4e41-b3a5-528f37501180-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-fx4cg\" (UID: \"542410e6-102a-4e41-b3a5-528f37501180\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-fx4cg" Oct 05 21:37:24 crc kubenswrapper[4754]: I1005 21:37:24.529048 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cftpm\" (UniqueName: \"kubernetes.io/projected/542410e6-102a-4e41-b3a5-528f37501180-kube-api-access-cftpm\") pod \"nova-edpm-deployment-openstack-edpm-ipam-fx4cg\" (UID: \"542410e6-102a-4e41-b3a5-528f37501180\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-fx4cg" Oct 05 21:37:24 crc kubenswrapper[4754]: I1005 21:37:24.529085 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/542410e6-102a-4e41-b3a5-528f37501180-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-fx4cg\" (UID: \"542410e6-102a-4e41-b3a5-528f37501180\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-fx4cg" Oct 05 21:37:24 crc kubenswrapper[4754]: I1005 21:37:24.529251 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/542410e6-102a-4e41-b3a5-528f37501180-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-fx4cg\" (UID: \"542410e6-102a-4e41-b3a5-528f37501180\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-fx4cg" Oct 05 21:37:24 crc kubenswrapper[4754]: I1005 21:37:24.529319 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/542410e6-102a-4e41-b3a5-528f37501180-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-fx4cg\" (UID: \"542410e6-102a-4e41-b3a5-528f37501180\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-fx4cg" Oct 05 21:37:24 crc kubenswrapper[4754]: I1005 21:37:24.529345 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/542410e6-102a-4e41-b3a5-528f37501180-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-fx4cg\" (UID: \"542410e6-102a-4e41-b3a5-528f37501180\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-fx4cg" Oct 05 21:37:24 crc kubenswrapper[4754]: I1005 21:37:24.529528 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/542410e6-102a-4e41-b3a5-528f37501180-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-fx4cg\" (UID: \"542410e6-102a-4e41-b3a5-528f37501180\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-fx4cg" Oct 05 21:37:24 crc kubenswrapper[4754]: I1005 21:37:24.630878 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/542410e6-102a-4e41-b3a5-528f37501180-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-fx4cg\" (UID: \"542410e6-102a-4e41-b3a5-528f37501180\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-fx4cg" Oct 05 21:37:24 crc kubenswrapper[4754]: I1005 21:37:24.630935 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/542410e6-102a-4e41-b3a5-528f37501180-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-fx4cg\" (UID: \"542410e6-102a-4e41-b3a5-528f37501180\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-fx4cg" Oct 05 21:37:24 crc kubenswrapper[4754]: I1005 21:37:24.630960 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cftpm\" (UniqueName: \"kubernetes.io/projected/542410e6-102a-4e41-b3a5-528f37501180-kube-api-access-cftpm\") pod \"nova-edpm-deployment-openstack-edpm-ipam-fx4cg\" (UID: \"542410e6-102a-4e41-b3a5-528f37501180\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-fx4cg" Oct 05 21:37:24 crc kubenswrapper[4754]: I1005 21:37:24.630988 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/542410e6-102a-4e41-b3a5-528f37501180-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-fx4cg\" (UID: \"542410e6-102a-4e41-b3a5-528f37501180\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-fx4cg" Oct 05 21:37:24 crc kubenswrapper[4754]: I1005 21:37:24.631028 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/542410e6-102a-4e41-b3a5-528f37501180-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-fx4cg\" (UID: \"542410e6-102a-4e41-b3a5-528f37501180\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-fx4cg" Oct 05 21:37:24 crc kubenswrapper[4754]: I1005 21:37:24.631056 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/542410e6-102a-4e41-b3a5-528f37501180-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-fx4cg\" (UID: \"542410e6-102a-4e41-b3a5-528f37501180\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-fx4cg" Oct 05 21:37:24 crc kubenswrapper[4754]: I1005 21:37:24.631074 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/542410e6-102a-4e41-b3a5-528f37501180-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-fx4cg\" (UID: \"542410e6-102a-4e41-b3a5-528f37501180\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-fx4cg" Oct 05 21:37:24 crc kubenswrapper[4754]: I1005 21:37:24.631114 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/542410e6-102a-4e41-b3a5-528f37501180-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-fx4cg\" (UID: \"542410e6-102a-4e41-b3a5-528f37501180\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-fx4cg" Oct 05 21:37:24 crc kubenswrapper[4754]: I1005 21:37:24.631139 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/542410e6-102a-4e41-b3a5-528f37501180-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-fx4cg\" (UID: \"542410e6-102a-4e41-b3a5-528f37501180\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-fx4cg" Oct 05 21:37:24 crc kubenswrapper[4754]: I1005 21:37:24.633299 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/542410e6-102a-4e41-b3a5-528f37501180-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-fx4cg\" (UID: \"542410e6-102a-4e41-b3a5-528f37501180\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-fx4cg" Oct 05 21:37:24 crc kubenswrapper[4754]: I1005 21:37:24.634716 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/542410e6-102a-4e41-b3a5-528f37501180-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-fx4cg\" (UID: \"542410e6-102a-4e41-b3a5-528f37501180\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-fx4cg" Oct 05 21:37:24 crc kubenswrapper[4754]: I1005 21:37:24.635154 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/542410e6-102a-4e41-b3a5-528f37501180-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-fx4cg\" (UID: \"542410e6-102a-4e41-b3a5-528f37501180\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-fx4cg" Oct 05 21:37:24 crc kubenswrapper[4754]: I1005 21:37:24.637632 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/542410e6-102a-4e41-b3a5-528f37501180-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-fx4cg\" (UID: \"542410e6-102a-4e41-b3a5-528f37501180\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-fx4cg" Oct 05 21:37:24 crc kubenswrapper[4754]: I1005 21:37:24.638534 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/542410e6-102a-4e41-b3a5-528f37501180-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-fx4cg\" (UID: \"542410e6-102a-4e41-b3a5-528f37501180\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-fx4cg" Oct 05 21:37:24 crc kubenswrapper[4754]: I1005 21:37:24.639107 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/542410e6-102a-4e41-b3a5-528f37501180-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-fx4cg\" (UID: \"542410e6-102a-4e41-b3a5-528f37501180\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-fx4cg" Oct 05 21:37:24 crc kubenswrapper[4754]: I1005 21:37:24.645030 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/542410e6-102a-4e41-b3a5-528f37501180-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-fx4cg\" (UID: \"542410e6-102a-4e41-b3a5-528f37501180\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-fx4cg" Oct 05 21:37:24 crc kubenswrapper[4754]: I1005 21:37:24.650232 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/542410e6-102a-4e41-b3a5-528f37501180-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-fx4cg\" (UID: \"542410e6-102a-4e41-b3a5-528f37501180\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-fx4cg" Oct 05 21:37:24 crc kubenswrapper[4754]: I1005 21:37:24.650977 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cftpm\" (UniqueName: \"kubernetes.io/projected/542410e6-102a-4e41-b3a5-528f37501180-kube-api-access-cftpm\") pod \"nova-edpm-deployment-openstack-edpm-ipam-fx4cg\" (UID: \"542410e6-102a-4e41-b3a5-528f37501180\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-fx4cg" Oct 05 21:37:24 crc kubenswrapper[4754]: I1005 21:37:24.864061 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-fx4cg" Oct 05 21:37:25 crc kubenswrapper[4754]: I1005 21:37:25.391117 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-fx4cg"] Oct 05 21:37:25 crc kubenswrapper[4754]: W1005 21:37:25.402720 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod542410e6_102a_4e41_b3a5_528f37501180.slice/crio-5572f01993701d36b1a8e15fdfcc386bfb0db1eb477a9a8523be26608389a332 WatchSource:0}: Error finding container 5572f01993701d36b1a8e15fdfcc386bfb0db1eb477a9a8523be26608389a332: Status 404 returned error can't find the container with id 5572f01993701d36b1a8e15fdfcc386bfb0db1eb477a9a8523be26608389a332 Oct 05 21:37:25 crc kubenswrapper[4754]: I1005 21:37:25.410028 4754 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 05 21:37:26 crc kubenswrapper[4754]: I1005 21:37:26.262928 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-fx4cg" event={"ID":"542410e6-102a-4e41-b3a5-528f37501180","Type":"ContainerStarted","Data":"5572f01993701d36b1a8e15fdfcc386bfb0db1eb477a9a8523be26608389a332"} Oct 05 21:37:27 crc kubenswrapper[4754]: I1005 21:37:27.271796 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-fx4cg" event={"ID":"542410e6-102a-4e41-b3a5-528f37501180","Type":"ContainerStarted","Data":"85f2aa85759c257aa9638fcf726e10c26ef636131aaa91180cf0e2df67051256"} Oct 05 21:37:27 crc kubenswrapper[4754]: I1005 21:37:27.302259 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-fx4cg" podStartSLOduration=2.590468084 podStartE2EDuration="3.302233541s" podCreationTimestamp="2025-10-05 21:37:24 +0000 UTC" firstStartedPulling="2025-10-05 21:37:25.4097929 +0000 UTC m=+2569.313911600" lastFinishedPulling="2025-10-05 21:37:26.121558337 +0000 UTC m=+2570.025677057" observedRunningTime="2025-10-05 21:37:27.290950758 +0000 UTC m=+2571.195069498" watchObservedRunningTime="2025-10-05 21:37:27.302233541 +0000 UTC m=+2571.206352291" Oct 05 21:39:35 crc kubenswrapper[4754]: I1005 21:39:35.244876 4754 patch_prober.go:28] interesting pod/machine-config-daemon-b2h9k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 21:39:35 crc kubenswrapper[4754]: I1005 21:39:35.245603 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 21:40:05 crc kubenswrapper[4754]: I1005 21:40:05.245904 4754 patch_prober.go:28] interesting pod/machine-config-daemon-b2h9k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 21:40:05 crc kubenswrapper[4754]: I1005 21:40:05.247102 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 21:40:35 crc kubenswrapper[4754]: I1005 21:40:35.244805 4754 patch_prober.go:28] interesting pod/machine-config-daemon-b2h9k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 21:40:35 crc kubenswrapper[4754]: I1005 21:40:35.245591 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 21:40:35 crc kubenswrapper[4754]: I1005 21:40:35.245660 4754 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" Oct 05 21:40:35 crc kubenswrapper[4754]: I1005 21:40:35.246834 4754 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5134bc33c72b286f6d3cee42c9ef42b9498b3009c16d4989384d741eded7eda2"} pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 05 21:40:35 crc kubenswrapper[4754]: I1005 21:40:35.246935 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" containerName="machine-config-daemon" containerID="cri-o://5134bc33c72b286f6d3cee42c9ef42b9498b3009c16d4989384d741eded7eda2" gracePeriod=600 Oct 05 21:40:35 crc kubenswrapper[4754]: I1005 21:40:35.450544 4754 generic.go:334] "Generic (PLEG): container finished" podID="be95a413-6a59-45b4-84b7-b43cae694a26" containerID="5134bc33c72b286f6d3cee42c9ef42b9498b3009c16d4989384d741eded7eda2" exitCode=0 Oct 05 21:40:35 crc kubenswrapper[4754]: I1005 21:40:35.450654 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" event={"ID":"be95a413-6a59-45b4-84b7-b43cae694a26","Type":"ContainerDied","Data":"5134bc33c72b286f6d3cee42c9ef42b9498b3009c16d4989384d741eded7eda2"} Oct 05 21:40:35 crc kubenswrapper[4754]: I1005 21:40:35.450955 4754 scope.go:117] "RemoveContainer" containerID="0df70ace610d977d6937bca9a99a0d188993dfa6fcd8015008efc0870430fdb3" Oct 05 21:40:36 crc kubenswrapper[4754]: I1005 21:40:36.460396 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" event={"ID":"be95a413-6a59-45b4-84b7-b43cae694a26","Type":"ContainerStarted","Data":"1bb4b1f5def9e89ddfea5b5e0921b99d55f1ceb68868b957d9868c33a0211da3"} Oct 05 21:41:31 crc kubenswrapper[4754]: I1005 21:41:31.073880 4754 generic.go:334] "Generic (PLEG): container finished" podID="542410e6-102a-4e41-b3a5-528f37501180" containerID="85f2aa85759c257aa9638fcf726e10c26ef636131aaa91180cf0e2df67051256" exitCode=0 Oct 05 21:41:31 crc kubenswrapper[4754]: I1005 21:41:31.074156 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-fx4cg" event={"ID":"542410e6-102a-4e41-b3a5-528f37501180","Type":"ContainerDied","Data":"85f2aa85759c257aa9638fcf726e10c26ef636131aaa91180cf0e2df67051256"} Oct 05 21:41:32 crc kubenswrapper[4754]: I1005 21:41:32.563066 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-fx4cg" Oct 05 21:41:32 crc kubenswrapper[4754]: I1005 21:41:32.586449 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/542410e6-102a-4e41-b3a5-528f37501180-nova-cell1-compute-config-0\") pod \"542410e6-102a-4e41-b3a5-528f37501180\" (UID: \"542410e6-102a-4e41-b3a5-528f37501180\") " Oct 05 21:41:32 crc kubenswrapper[4754]: I1005 21:41:32.586537 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/542410e6-102a-4e41-b3a5-528f37501180-nova-migration-ssh-key-0\") pod \"542410e6-102a-4e41-b3a5-528f37501180\" (UID: \"542410e6-102a-4e41-b3a5-528f37501180\") " Oct 05 21:41:32 crc kubenswrapper[4754]: I1005 21:41:32.586590 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/542410e6-102a-4e41-b3a5-528f37501180-nova-extra-config-0\") pod \"542410e6-102a-4e41-b3a5-528f37501180\" (UID: \"542410e6-102a-4e41-b3a5-528f37501180\") " Oct 05 21:41:32 crc kubenswrapper[4754]: I1005 21:41:32.586669 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/542410e6-102a-4e41-b3a5-528f37501180-nova-combined-ca-bundle\") pod \"542410e6-102a-4e41-b3a5-528f37501180\" (UID: \"542410e6-102a-4e41-b3a5-528f37501180\") " Oct 05 21:41:32 crc kubenswrapper[4754]: I1005 21:41:32.586717 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/542410e6-102a-4e41-b3a5-528f37501180-nova-migration-ssh-key-1\") pod \"542410e6-102a-4e41-b3a5-528f37501180\" (UID: \"542410e6-102a-4e41-b3a5-528f37501180\") " Oct 05 21:41:32 crc kubenswrapper[4754]: I1005 21:41:32.586771 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/542410e6-102a-4e41-b3a5-528f37501180-inventory\") pod \"542410e6-102a-4e41-b3a5-528f37501180\" (UID: \"542410e6-102a-4e41-b3a5-528f37501180\") " Oct 05 21:41:32 crc kubenswrapper[4754]: I1005 21:41:32.586795 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cftpm\" (UniqueName: \"kubernetes.io/projected/542410e6-102a-4e41-b3a5-528f37501180-kube-api-access-cftpm\") pod \"542410e6-102a-4e41-b3a5-528f37501180\" (UID: \"542410e6-102a-4e41-b3a5-528f37501180\") " Oct 05 21:41:32 crc kubenswrapper[4754]: I1005 21:41:32.586838 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/542410e6-102a-4e41-b3a5-528f37501180-ssh-key\") pod \"542410e6-102a-4e41-b3a5-528f37501180\" (UID: \"542410e6-102a-4e41-b3a5-528f37501180\") " Oct 05 21:41:32 crc kubenswrapper[4754]: I1005 21:41:32.586858 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/542410e6-102a-4e41-b3a5-528f37501180-nova-cell1-compute-config-1\") pod \"542410e6-102a-4e41-b3a5-528f37501180\" (UID: \"542410e6-102a-4e41-b3a5-528f37501180\") " Oct 05 21:41:32 crc kubenswrapper[4754]: I1005 21:41:32.594744 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/542410e6-102a-4e41-b3a5-528f37501180-kube-api-access-cftpm" (OuterVolumeSpecName: "kube-api-access-cftpm") pod "542410e6-102a-4e41-b3a5-528f37501180" (UID: "542410e6-102a-4e41-b3a5-528f37501180"). InnerVolumeSpecName "kube-api-access-cftpm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:41:32 crc kubenswrapper[4754]: I1005 21:41:32.603625 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/542410e6-102a-4e41-b3a5-528f37501180-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "542410e6-102a-4e41-b3a5-528f37501180" (UID: "542410e6-102a-4e41-b3a5-528f37501180"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:41:32 crc kubenswrapper[4754]: I1005 21:41:32.620020 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/542410e6-102a-4e41-b3a5-528f37501180-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "542410e6-102a-4e41-b3a5-528f37501180" (UID: "542410e6-102a-4e41-b3a5-528f37501180"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:41:32 crc kubenswrapper[4754]: I1005 21:41:32.639545 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/542410e6-102a-4e41-b3a5-528f37501180-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "542410e6-102a-4e41-b3a5-528f37501180" (UID: "542410e6-102a-4e41-b3a5-528f37501180"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 21:41:32 crc kubenswrapper[4754]: I1005 21:41:32.646684 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/542410e6-102a-4e41-b3a5-528f37501180-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "542410e6-102a-4e41-b3a5-528f37501180" (UID: "542410e6-102a-4e41-b3a5-528f37501180"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:41:32 crc kubenswrapper[4754]: I1005 21:41:32.649637 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/542410e6-102a-4e41-b3a5-528f37501180-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "542410e6-102a-4e41-b3a5-528f37501180" (UID: "542410e6-102a-4e41-b3a5-528f37501180"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:41:32 crc kubenswrapper[4754]: I1005 21:41:32.650912 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/542410e6-102a-4e41-b3a5-528f37501180-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "542410e6-102a-4e41-b3a5-528f37501180" (UID: "542410e6-102a-4e41-b3a5-528f37501180"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:41:32 crc kubenswrapper[4754]: I1005 21:41:32.663391 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/542410e6-102a-4e41-b3a5-528f37501180-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "542410e6-102a-4e41-b3a5-528f37501180" (UID: "542410e6-102a-4e41-b3a5-528f37501180"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:41:32 crc kubenswrapper[4754]: I1005 21:41:32.675124 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/542410e6-102a-4e41-b3a5-528f37501180-inventory" (OuterVolumeSpecName: "inventory") pod "542410e6-102a-4e41-b3a5-528f37501180" (UID: "542410e6-102a-4e41-b3a5-528f37501180"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:41:32 crc kubenswrapper[4754]: I1005 21:41:32.688407 4754 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/542410e6-102a-4e41-b3a5-528f37501180-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Oct 05 21:41:32 crc kubenswrapper[4754]: I1005 21:41:32.688449 4754 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/542410e6-102a-4e41-b3a5-528f37501180-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 21:41:32 crc kubenswrapper[4754]: I1005 21:41:32.688460 4754 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/542410e6-102a-4e41-b3a5-528f37501180-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Oct 05 21:41:32 crc kubenswrapper[4754]: I1005 21:41:32.688468 4754 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/542410e6-102a-4e41-b3a5-528f37501180-inventory\") on node \"crc\" DevicePath \"\"" Oct 05 21:41:32 crc kubenswrapper[4754]: I1005 21:41:32.688476 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cftpm\" (UniqueName: \"kubernetes.io/projected/542410e6-102a-4e41-b3a5-528f37501180-kube-api-access-cftpm\") on node \"crc\" DevicePath \"\"" Oct 05 21:41:32 crc kubenswrapper[4754]: I1005 21:41:32.688484 4754 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/542410e6-102a-4e41-b3a5-528f37501180-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 05 21:41:32 crc kubenswrapper[4754]: I1005 21:41:32.688529 4754 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/542410e6-102a-4e41-b3a5-528f37501180-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Oct 05 21:41:32 crc kubenswrapper[4754]: I1005 21:41:32.688541 4754 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/542410e6-102a-4e41-b3a5-528f37501180-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Oct 05 21:41:32 crc kubenswrapper[4754]: I1005 21:41:32.688549 4754 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/542410e6-102a-4e41-b3a5-528f37501180-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Oct 05 21:41:33 crc kubenswrapper[4754]: I1005 21:41:33.107258 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-fx4cg" event={"ID":"542410e6-102a-4e41-b3a5-528f37501180","Type":"ContainerDied","Data":"5572f01993701d36b1a8e15fdfcc386bfb0db1eb477a9a8523be26608389a332"} Oct 05 21:41:33 crc kubenswrapper[4754]: I1005 21:41:33.107736 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5572f01993701d36b1a8e15fdfcc386bfb0db1eb477a9a8523be26608389a332" Oct 05 21:41:33 crc kubenswrapper[4754]: I1005 21:41:33.107355 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-fx4cg" Oct 05 21:41:33 crc kubenswrapper[4754]: I1005 21:41:33.221742 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s2xqb"] Oct 05 21:41:33 crc kubenswrapper[4754]: E1005 21:41:33.222314 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="542410e6-102a-4e41-b3a5-528f37501180" containerName="nova-edpm-deployment-openstack-edpm-ipam" Oct 05 21:41:33 crc kubenswrapper[4754]: I1005 21:41:33.222357 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="542410e6-102a-4e41-b3a5-528f37501180" containerName="nova-edpm-deployment-openstack-edpm-ipam" Oct 05 21:41:33 crc kubenswrapper[4754]: I1005 21:41:33.222741 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="542410e6-102a-4e41-b3a5-528f37501180" containerName="nova-edpm-deployment-openstack-edpm-ipam" Oct 05 21:41:33 crc kubenswrapper[4754]: I1005 21:41:33.223779 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s2xqb" Oct 05 21:41:33 crc kubenswrapper[4754]: I1005 21:41:33.228561 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 05 21:41:33 crc kubenswrapper[4754]: I1005 21:41:33.228672 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Oct 05 21:41:33 crc kubenswrapper[4754]: I1005 21:41:33.229858 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 05 21:41:33 crc kubenswrapper[4754]: I1005 21:41:33.229897 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 05 21:41:33 crc kubenswrapper[4754]: I1005 21:41:33.230550 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-v54jd" Oct 05 21:41:33 crc kubenswrapper[4754]: I1005 21:41:33.238394 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s2xqb"] Oct 05 21:41:33 crc kubenswrapper[4754]: I1005 21:41:33.299312 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7c094190-76ba-4310-ae5d-1e82d3caeac2-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-s2xqb\" (UID: \"7c094190-76ba-4310-ae5d-1e82d3caeac2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s2xqb" Oct 05 21:41:33 crc kubenswrapper[4754]: I1005 21:41:33.299625 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d8j48\" (UniqueName: \"kubernetes.io/projected/7c094190-76ba-4310-ae5d-1e82d3caeac2-kube-api-access-d8j48\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-s2xqb\" (UID: \"7c094190-76ba-4310-ae5d-1e82d3caeac2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s2xqb" Oct 05 21:41:33 crc kubenswrapper[4754]: I1005 21:41:33.299766 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c094190-76ba-4310-ae5d-1e82d3caeac2-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-s2xqb\" (UID: \"7c094190-76ba-4310-ae5d-1e82d3caeac2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s2xqb" Oct 05 21:41:33 crc kubenswrapper[4754]: I1005 21:41:33.299854 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7c094190-76ba-4310-ae5d-1e82d3caeac2-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-s2xqb\" (UID: \"7c094190-76ba-4310-ae5d-1e82d3caeac2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s2xqb" Oct 05 21:41:33 crc kubenswrapper[4754]: I1005 21:41:33.299957 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/7c094190-76ba-4310-ae5d-1e82d3caeac2-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-s2xqb\" (UID: \"7c094190-76ba-4310-ae5d-1e82d3caeac2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s2xqb" Oct 05 21:41:33 crc kubenswrapper[4754]: I1005 21:41:33.300042 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/7c094190-76ba-4310-ae5d-1e82d3caeac2-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-s2xqb\" (UID: \"7c094190-76ba-4310-ae5d-1e82d3caeac2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s2xqb" Oct 05 21:41:33 crc kubenswrapper[4754]: I1005 21:41:33.300260 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/7c094190-76ba-4310-ae5d-1e82d3caeac2-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-s2xqb\" (UID: \"7c094190-76ba-4310-ae5d-1e82d3caeac2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s2xqb" Oct 05 21:41:33 crc kubenswrapper[4754]: I1005 21:41:33.401241 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/7c094190-76ba-4310-ae5d-1e82d3caeac2-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-s2xqb\" (UID: \"7c094190-76ba-4310-ae5d-1e82d3caeac2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s2xqb" Oct 05 21:41:33 crc kubenswrapper[4754]: I1005 21:41:33.401521 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7c094190-76ba-4310-ae5d-1e82d3caeac2-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-s2xqb\" (UID: \"7c094190-76ba-4310-ae5d-1e82d3caeac2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s2xqb" Oct 05 21:41:33 crc kubenswrapper[4754]: I1005 21:41:33.401685 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d8j48\" (UniqueName: \"kubernetes.io/projected/7c094190-76ba-4310-ae5d-1e82d3caeac2-kube-api-access-d8j48\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-s2xqb\" (UID: \"7c094190-76ba-4310-ae5d-1e82d3caeac2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s2xqb" Oct 05 21:41:33 crc kubenswrapper[4754]: I1005 21:41:33.401775 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c094190-76ba-4310-ae5d-1e82d3caeac2-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-s2xqb\" (UID: \"7c094190-76ba-4310-ae5d-1e82d3caeac2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s2xqb" Oct 05 21:41:33 crc kubenswrapper[4754]: I1005 21:41:33.401858 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7c094190-76ba-4310-ae5d-1e82d3caeac2-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-s2xqb\" (UID: \"7c094190-76ba-4310-ae5d-1e82d3caeac2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s2xqb" Oct 05 21:41:33 crc kubenswrapper[4754]: I1005 21:41:33.401946 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/7c094190-76ba-4310-ae5d-1e82d3caeac2-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-s2xqb\" (UID: \"7c094190-76ba-4310-ae5d-1e82d3caeac2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s2xqb" Oct 05 21:41:33 crc kubenswrapper[4754]: I1005 21:41:33.402408 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/7c094190-76ba-4310-ae5d-1e82d3caeac2-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-s2xqb\" (UID: \"7c094190-76ba-4310-ae5d-1e82d3caeac2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s2xqb" Oct 05 21:41:33 crc kubenswrapper[4754]: I1005 21:41:33.408219 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c094190-76ba-4310-ae5d-1e82d3caeac2-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-s2xqb\" (UID: \"7c094190-76ba-4310-ae5d-1e82d3caeac2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s2xqb" Oct 05 21:41:33 crc kubenswrapper[4754]: I1005 21:41:33.408844 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/7c094190-76ba-4310-ae5d-1e82d3caeac2-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-s2xqb\" (UID: \"7c094190-76ba-4310-ae5d-1e82d3caeac2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s2xqb" Oct 05 21:41:33 crc kubenswrapper[4754]: I1005 21:41:33.409886 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7c094190-76ba-4310-ae5d-1e82d3caeac2-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-s2xqb\" (UID: \"7c094190-76ba-4310-ae5d-1e82d3caeac2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s2xqb" Oct 05 21:41:33 crc kubenswrapper[4754]: I1005 21:41:33.410993 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7c094190-76ba-4310-ae5d-1e82d3caeac2-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-s2xqb\" (UID: \"7c094190-76ba-4310-ae5d-1e82d3caeac2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s2xqb" Oct 05 21:41:33 crc kubenswrapper[4754]: I1005 21:41:33.411975 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/7c094190-76ba-4310-ae5d-1e82d3caeac2-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-s2xqb\" (UID: \"7c094190-76ba-4310-ae5d-1e82d3caeac2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s2xqb" Oct 05 21:41:33 crc kubenswrapper[4754]: I1005 21:41:33.421025 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/7c094190-76ba-4310-ae5d-1e82d3caeac2-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-s2xqb\" (UID: \"7c094190-76ba-4310-ae5d-1e82d3caeac2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s2xqb" Oct 05 21:41:33 crc kubenswrapper[4754]: I1005 21:41:33.423138 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d8j48\" (UniqueName: \"kubernetes.io/projected/7c094190-76ba-4310-ae5d-1e82d3caeac2-kube-api-access-d8j48\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-s2xqb\" (UID: \"7c094190-76ba-4310-ae5d-1e82d3caeac2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s2xqb" Oct 05 21:41:33 crc kubenswrapper[4754]: I1005 21:41:33.568842 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s2xqb" Oct 05 21:41:34 crc kubenswrapper[4754]: I1005 21:41:34.092365 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s2xqb"] Oct 05 21:41:34 crc kubenswrapper[4754]: W1005 21:41:34.105819 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7c094190_76ba_4310_ae5d_1e82d3caeac2.slice/crio-f6bf9abe0ec81763ca0e47f688310bb99684bf64aa0379e8dc9cae3ed457c336 WatchSource:0}: Error finding container f6bf9abe0ec81763ca0e47f688310bb99684bf64aa0379e8dc9cae3ed457c336: Status 404 returned error can't find the container with id f6bf9abe0ec81763ca0e47f688310bb99684bf64aa0379e8dc9cae3ed457c336 Oct 05 21:41:34 crc kubenswrapper[4754]: I1005 21:41:34.119823 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s2xqb" event={"ID":"7c094190-76ba-4310-ae5d-1e82d3caeac2","Type":"ContainerStarted","Data":"f6bf9abe0ec81763ca0e47f688310bb99684bf64aa0379e8dc9cae3ed457c336"} Oct 05 21:41:36 crc kubenswrapper[4754]: I1005 21:41:36.148057 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s2xqb" event={"ID":"7c094190-76ba-4310-ae5d-1e82d3caeac2","Type":"ContainerStarted","Data":"8335deacab79709ecaa34accbe0b8d24f3377c8107540e1746ab2269e7b31fcc"} Oct 05 21:41:36 crc kubenswrapper[4754]: I1005 21:41:36.168909 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s2xqb" podStartSLOduration=2.376690262 podStartE2EDuration="3.16888968s" podCreationTimestamp="2025-10-05 21:41:33 +0000 UTC" firstStartedPulling="2025-10-05 21:41:34.110090802 +0000 UTC m=+2818.014209512" lastFinishedPulling="2025-10-05 21:41:34.90229018 +0000 UTC m=+2818.806408930" observedRunningTime="2025-10-05 21:41:36.16732863 +0000 UTC m=+2820.071447370" watchObservedRunningTime="2025-10-05 21:41:36.16888968 +0000 UTC m=+2820.073008420" Oct 05 21:41:49 crc kubenswrapper[4754]: I1005 21:41:49.202438 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-9qzc5"] Oct 05 21:41:49 crc kubenswrapper[4754]: I1005 21:41:49.218393 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9qzc5"] Oct 05 21:41:49 crc kubenswrapper[4754]: I1005 21:41:49.218542 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9qzc5" Oct 05 21:41:49 crc kubenswrapper[4754]: I1005 21:41:49.363811 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1a58750e-ca20-4f84-af48-8ceef4967672-utilities\") pod \"redhat-operators-9qzc5\" (UID: \"1a58750e-ca20-4f84-af48-8ceef4967672\") " pod="openshift-marketplace/redhat-operators-9qzc5" Oct 05 21:41:49 crc kubenswrapper[4754]: I1005 21:41:49.363949 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1a58750e-ca20-4f84-af48-8ceef4967672-catalog-content\") pod \"redhat-operators-9qzc5\" (UID: \"1a58750e-ca20-4f84-af48-8ceef4967672\") " pod="openshift-marketplace/redhat-operators-9qzc5" Oct 05 21:41:49 crc kubenswrapper[4754]: I1005 21:41:49.364010 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ppfqz\" (UniqueName: \"kubernetes.io/projected/1a58750e-ca20-4f84-af48-8ceef4967672-kube-api-access-ppfqz\") pod \"redhat-operators-9qzc5\" (UID: \"1a58750e-ca20-4f84-af48-8ceef4967672\") " pod="openshift-marketplace/redhat-operators-9qzc5" Oct 05 21:41:49 crc kubenswrapper[4754]: I1005 21:41:49.465413 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1a58750e-ca20-4f84-af48-8ceef4967672-utilities\") pod \"redhat-operators-9qzc5\" (UID: \"1a58750e-ca20-4f84-af48-8ceef4967672\") " pod="openshift-marketplace/redhat-operators-9qzc5" Oct 05 21:41:49 crc kubenswrapper[4754]: I1005 21:41:49.465549 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1a58750e-ca20-4f84-af48-8ceef4967672-catalog-content\") pod \"redhat-operators-9qzc5\" (UID: \"1a58750e-ca20-4f84-af48-8ceef4967672\") " pod="openshift-marketplace/redhat-operators-9qzc5" Oct 05 21:41:49 crc kubenswrapper[4754]: I1005 21:41:49.465609 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ppfqz\" (UniqueName: \"kubernetes.io/projected/1a58750e-ca20-4f84-af48-8ceef4967672-kube-api-access-ppfqz\") pod \"redhat-operators-9qzc5\" (UID: \"1a58750e-ca20-4f84-af48-8ceef4967672\") " pod="openshift-marketplace/redhat-operators-9qzc5" Oct 05 21:41:49 crc kubenswrapper[4754]: I1005 21:41:49.465952 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1a58750e-ca20-4f84-af48-8ceef4967672-utilities\") pod \"redhat-operators-9qzc5\" (UID: \"1a58750e-ca20-4f84-af48-8ceef4967672\") " pod="openshift-marketplace/redhat-operators-9qzc5" Oct 05 21:41:49 crc kubenswrapper[4754]: I1005 21:41:49.466251 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1a58750e-ca20-4f84-af48-8ceef4967672-catalog-content\") pod \"redhat-operators-9qzc5\" (UID: \"1a58750e-ca20-4f84-af48-8ceef4967672\") " pod="openshift-marketplace/redhat-operators-9qzc5" Oct 05 21:41:49 crc kubenswrapper[4754]: I1005 21:41:49.494204 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ppfqz\" (UniqueName: \"kubernetes.io/projected/1a58750e-ca20-4f84-af48-8ceef4967672-kube-api-access-ppfqz\") pod \"redhat-operators-9qzc5\" (UID: \"1a58750e-ca20-4f84-af48-8ceef4967672\") " pod="openshift-marketplace/redhat-operators-9qzc5" Oct 05 21:41:49 crc kubenswrapper[4754]: I1005 21:41:49.551055 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9qzc5" Oct 05 21:41:50 crc kubenswrapper[4754]: I1005 21:41:50.051768 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9qzc5"] Oct 05 21:41:50 crc kubenswrapper[4754]: I1005 21:41:50.303026 4754 generic.go:334] "Generic (PLEG): container finished" podID="1a58750e-ca20-4f84-af48-8ceef4967672" containerID="feaf459efbd062460432cc4cc9757150734a82ba551f9d36f7069dee9d4b1ab3" exitCode=0 Oct 05 21:41:50 crc kubenswrapper[4754]: I1005 21:41:50.303243 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9qzc5" event={"ID":"1a58750e-ca20-4f84-af48-8ceef4967672","Type":"ContainerDied","Data":"feaf459efbd062460432cc4cc9757150734a82ba551f9d36f7069dee9d4b1ab3"} Oct 05 21:41:50 crc kubenswrapper[4754]: I1005 21:41:50.303295 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9qzc5" event={"ID":"1a58750e-ca20-4f84-af48-8ceef4967672","Type":"ContainerStarted","Data":"c3786ebe7a9b344fefd1071b0e8526e745304d39eb6d27b3999b3ef87cd4d70c"} Oct 05 21:41:52 crc kubenswrapper[4754]: I1005 21:41:52.322929 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9qzc5" event={"ID":"1a58750e-ca20-4f84-af48-8ceef4967672","Type":"ContainerStarted","Data":"3bfa4c311e2731c8178f32cdf75cfe626ecac11ce4dcf7f9bc9dc7a68fb12953"} Oct 05 21:41:59 crc kubenswrapper[4754]: I1005 21:41:59.398565 4754 generic.go:334] "Generic (PLEG): container finished" podID="1a58750e-ca20-4f84-af48-8ceef4967672" containerID="3bfa4c311e2731c8178f32cdf75cfe626ecac11ce4dcf7f9bc9dc7a68fb12953" exitCode=0 Oct 05 21:41:59 crc kubenswrapper[4754]: I1005 21:41:59.398672 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9qzc5" event={"ID":"1a58750e-ca20-4f84-af48-8ceef4967672","Type":"ContainerDied","Data":"3bfa4c311e2731c8178f32cdf75cfe626ecac11ce4dcf7f9bc9dc7a68fb12953"} Oct 05 21:42:02 crc kubenswrapper[4754]: I1005 21:42:02.436208 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9qzc5" event={"ID":"1a58750e-ca20-4f84-af48-8ceef4967672","Type":"ContainerStarted","Data":"8ab97c48108b985baeb9697e705c3fa9684705e0ca02a096480f4238f1357c91"} Oct 05 21:42:02 crc kubenswrapper[4754]: I1005 21:42:02.462604 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-9qzc5" podStartSLOduration=2.198059341 podStartE2EDuration="13.462577301s" podCreationTimestamp="2025-10-05 21:41:49 +0000 UTC" firstStartedPulling="2025-10-05 21:41:50.304868784 +0000 UTC m=+2834.208987494" lastFinishedPulling="2025-10-05 21:42:01.569386714 +0000 UTC m=+2845.473505454" observedRunningTime="2025-10-05 21:42:02.45447607 +0000 UTC m=+2846.358594780" watchObservedRunningTime="2025-10-05 21:42:02.462577301 +0000 UTC m=+2846.366696051" Oct 05 21:42:09 crc kubenswrapper[4754]: I1005 21:42:09.551934 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-9qzc5" Oct 05 21:42:09 crc kubenswrapper[4754]: I1005 21:42:09.552709 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-9qzc5" Oct 05 21:42:09 crc kubenswrapper[4754]: I1005 21:42:09.628407 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-9qzc5" Oct 05 21:42:10 crc kubenswrapper[4754]: I1005 21:42:10.655603 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-9qzc5" Oct 05 21:42:10 crc kubenswrapper[4754]: I1005 21:42:10.745045 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-9qzc5"] Oct 05 21:42:12 crc kubenswrapper[4754]: I1005 21:42:12.579165 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-9qzc5" podUID="1a58750e-ca20-4f84-af48-8ceef4967672" containerName="registry-server" containerID="cri-o://8ab97c48108b985baeb9697e705c3fa9684705e0ca02a096480f4238f1357c91" gracePeriod=2 Oct 05 21:42:13 crc kubenswrapper[4754]: I1005 21:42:13.103250 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9qzc5" Oct 05 21:42:13 crc kubenswrapper[4754]: I1005 21:42:13.170395 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1a58750e-ca20-4f84-af48-8ceef4967672-catalog-content\") pod \"1a58750e-ca20-4f84-af48-8ceef4967672\" (UID: \"1a58750e-ca20-4f84-af48-8ceef4967672\") " Oct 05 21:42:13 crc kubenswrapper[4754]: I1005 21:42:13.170440 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1a58750e-ca20-4f84-af48-8ceef4967672-utilities\") pod \"1a58750e-ca20-4f84-af48-8ceef4967672\" (UID: \"1a58750e-ca20-4f84-af48-8ceef4967672\") " Oct 05 21:42:13 crc kubenswrapper[4754]: I1005 21:42:13.170578 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ppfqz\" (UniqueName: \"kubernetes.io/projected/1a58750e-ca20-4f84-af48-8ceef4967672-kube-api-access-ppfqz\") pod \"1a58750e-ca20-4f84-af48-8ceef4967672\" (UID: \"1a58750e-ca20-4f84-af48-8ceef4967672\") " Oct 05 21:42:13 crc kubenswrapper[4754]: I1005 21:42:13.174298 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1a58750e-ca20-4f84-af48-8ceef4967672-utilities" (OuterVolumeSpecName: "utilities") pod "1a58750e-ca20-4f84-af48-8ceef4967672" (UID: "1a58750e-ca20-4f84-af48-8ceef4967672"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 21:42:13 crc kubenswrapper[4754]: I1005 21:42:13.178933 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1a58750e-ca20-4f84-af48-8ceef4967672-kube-api-access-ppfqz" (OuterVolumeSpecName: "kube-api-access-ppfqz") pod "1a58750e-ca20-4f84-af48-8ceef4967672" (UID: "1a58750e-ca20-4f84-af48-8ceef4967672"). InnerVolumeSpecName "kube-api-access-ppfqz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:42:13 crc kubenswrapper[4754]: I1005 21:42:13.261729 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1a58750e-ca20-4f84-af48-8ceef4967672-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1a58750e-ca20-4f84-af48-8ceef4967672" (UID: "1a58750e-ca20-4f84-af48-8ceef4967672"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 21:42:13 crc kubenswrapper[4754]: I1005 21:42:13.273914 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ppfqz\" (UniqueName: \"kubernetes.io/projected/1a58750e-ca20-4f84-af48-8ceef4967672-kube-api-access-ppfqz\") on node \"crc\" DevicePath \"\"" Oct 05 21:42:13 crc kubenswrapper[4754]: I1005 21:42:13.273939 4754 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1a58750e-ca20-4f84-af48-8ceef4967672-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 21:42:13 crc kubenswrapper[4754]: I1005 21:42:13.273950 4754 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1a58750e-ca20-4f84-af48-8ceef4967672-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 21:42:13 crc kubenswrapper[4754]: I1005 21:42:13.605545 4754 generic.go:334] "Generic (PLEG): container finished" podID="1a58750e-ca20-4f84-af48-8ceef4967672" containerID="8ab97c48108b985baeb9697e705c3fa9684705e0ca02a096480f4238f1357c91" exitCode=0 Oct 05 21:42:13 crc kubenswrapper[4754]: I1005 21:42:13.605599 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9qzc5" event={"ID":"1a58750e-ca20-4f84-af48-8ceef4967672","Type":"ContainerDied","Data":"8ab97c48108b985baeb9697e705c3fa9684705e0ca02a096480f4238f1357c91"} Oct 05 21:42:13 crc kubenswrapper[4754]: I1005 21:42:13.605634 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9qzc5" event={"ID":"1a58750e-ca20-4f84-af48-8ceef4967672","Type":"ContainerDied","Data":"c3786ebe7a9b344fefd1071b0e8526e745304d39eb6d27b3999b3ef87cd4d70c"} Oct 05 21:42:13 crc kubenswrapper[4754]: I1005 21:42:13.605654 4754 scope.go:117] "RemoveContainer" containerID="8ab97c48108b985baeb9697e705c3fa9684705e0ca02a096480f4238f1357c91" Oct 05 21:42:13 crc kubenswrapper[4754]: I1005 21:42:13.605705 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9qzc5" Oct 05 21:42:13 crc kubenswrapper[4754]: I1005 21:42:13.656816 4754 scope.go:117] "RemoveContainer" containerID="3bfa4c311e2731c8178f32cdf75cfe626ecac11ce4dcf7f9bc9dc7a68fb12953" Oct 05 21:42:13 crc kubenswrapper[4754]: I1005 21:42:13.668342 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-9qzc5"] Oct 05 21:42:13 crc kubenswrapper[4754]: I1005 21:42:13.694651 4754 scope.go:117] "RemoveContainer" containerID="feaf459efbd062460432cc4cc9757150734a82ba551f9d36f7069dee9d4b1ab3" Oct 05 21:42:13 crc kubenswrapper[4754]: I1005 21:42:13.696551 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-9qzc5"] Oct 05 21:42:13 crc kubenswrapper[4754]: I1005 21:42:13.728647 4754 scope.go:117] "RemoveContainer" containerID="8ab97c48108b985baeb9697e705c3fa9684705e0ca02a096480f4238f1357c91" Oct 05 21:42:13 crc kubenswrapper[4754]: E1005 21:42:13.729127 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8ab97c48108b985baeb9697e705c3fa9684705e0ca02a096480f4238f1357c91\": container with ID starting with 8ab97c48108b985baeb9697e705c3fa9684705e0ca02a096480f4238f1357c91 not found: ID does not exist" containerID="8ab97c48108b985baeb9697e705c3fa9684705e0ca02a096480f4238f1357c91" Oct 05 21:42:13 crc kubenswrapper[4754]: I1005 21:42:13.729202 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8ab97c48108b985baeb9697e705c3fa9684705e0ca02a096480f4238f1357c91"} err="failed to get container status \"8ab97c48108b985baeb9697e705c3fa9684705e0ca02a096480f4238f1357c91\": rpc error: code = NotFound desc = could not find container \"8ab97c48108b985baeb9697e705c3fa9684705e0ca02a096480f4238f1357c91\": container with ID starting with 8ab97c48108b985baeb9697e705c3fa9684705e0ca02a096480f4238f1357c91 not found: ID does not exist" Oct 05 21:42:13 crc kubenswrapper[4754]: I1005 21:42:13.729236 4754 scope.go:117] "RemoveContainer" containerID="3bfa4c311e2731c8178f32cdf75cfe626ecac11ce4dcf7f9bc9dc7a68fb12953" Oct 05 21:42:13 crc kubenswrapper[4754]: E1005 21:42:13.729622 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3bfa4c311e2731c8178f32cdf75cfe626ecac11ce4dcf7f9bc9dc7a68fb12953\": container with ID starting with 3bfa4c311e2731c8178f32cdf75cfe626ecac11ce4dcf7f9bc9dc7a68fb12953 not found: ID does not exist" containerID="3bfa4c311e2731c8178f32cdf75cfe626ecac11ce4dcf7f9bc9dc7a68fb12953" Oct 05 21:42:13 crc kubenswrapper[4754]: I1005 21:42:13.729660 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3bfa4c311e2731c8178f32cdf75cfe626ecac11ce4dcf7f9bc9dc7a68fb12953"} err="failed to get container status \"3bfa4c311e2731c8178f32cdf75cfe626ecac11ce4dcf7f9bc9dc7a68fb12953\": rpc error: code = NotFound desc = could not find container \"3bfa4c311e2731c8178f32cdf75cfe626ecac11ce4dcf7f9bc9dc7a68fb12953\": container with ID starting with 3bfa4c311e2731c8178f32cdf75cfe626ecac11ce4dcf7f9bc9dc7a68fb12953 not found: ID does not exist" Oct 05 21:42:13 crc kubenswrapper[4754]: I1005 21:42:13.729691 4754 scope.go:117] "RemoveContainer" containerID="feaf459efbd062460432cc4cc9757150734a82ba551f9d36f7069dee9d4b1ab3" Oct 05 21:42:13 crc kubenswrapper[4754]: E1005 21:42:13.730084 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"feaf459efbd062460432cc4cc9757150734a82ba551f9d36f7069dee9d4b1ab3\": container with ID starting with feaf459efbd062460432cc4cc9757150734a82ba551f9d36f7069dee9d4b1ab3 not found: ID does not exist" containerID="feaf459efbd062460432cc4cc9757150734a82ba551f9d36f7069dee9d4b1ab3" Oct 05 21:42:13 crc kubenswrapper[4754]: I1005 21:42:13.730131 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"feaf459efbd062460432cc4cc9757150734a82ba551f9d36f7069dee9d4b1ab3"} err="failed to get container status \"feaf459efbd062460432cc4cc9757150734a82ba551f9d36f7069dee9d4b1ab3\": rpc error: code = NotFound desc = could not find container \"feaf459efbd062460432cc4cc9757150734a82ba551f9d36f7069dee9d4b1ab3\": container with ID starting with feaf459efbd062460432cc4cc9757150734a82ba551f9d36f7069dee9d4b1ab3 not found: ID does not exist" Oct 05 21:42:14 crc kubenswrapper[4754]: I1005 21:42:14.861292 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1a58750e-ca20-4f84-af48-8ceef4967672" path="/var/lib/kubelet/pods/1a58750e-ca20-4f84-af48-8ceef4967672/volumes" Oct 05 21:42:35 crc kubenswrapper[4754]: I1005 21:42:35.245278 4754 patch_prober.go:28] interesting pod/machine-config-daemon-b2h9k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 21:42:35 crc kubenswrapper[4754]: I1005 21:42:35.246234 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 21:43:05 crc kubenswrapper[4754]: I1005 21:43:05.244891 4754 patch_prober.go:28] interesting pod/machine-config-daemon-b2h9k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 21:43:05 crc kubenswrapper[4754]: I1005 21:43:05.245376 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 21:43:15 crc kubenswrapper[4754]: I1005 21:43:15.123381 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-nqjkh"] Oct 05 21:43:15 crc kubenswrapper[4754]: E1005 21:43:15.124306 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a58750e-ca20-4f84-af48-8ceef4967672" containerName="extract-content" Oct 05 21:43:15 crc kubenswrapper[4754]: I1005 21:43:15.124322 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a58750e-ca20-4f84-af48-8ceef4967672" containerName="extract-content" Oct 05 21:43:15 crc kubenswrapper[4754]: E1005 21:43:15.124349 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a58750e-ca20-4f84-af48-8ceef4967672" containerName="extract-utilities" Oct 05 21:43:15 crc kubenswrapper[4754]: I1005 21:43:15.124357 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a58750e-ca20-4f84-af48-8ceef4967672" containerName="extract-utilities" Oct 05 21:43:15 crc kubenswrapper[4754]: E1005 21:43:15.124371 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a58750e-ca20-4f84-af48-8ceef4967672" containerName="registry-server" Oct 05 21:43:15 crc kubenswrapper[4754]: I1005 21:43:15.124379 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a58750e-ca20-4f84-af48-8ceef4967672" containerName="registry-server" Oct 05 21:43:15 crc kubenswrapper[4754]: I1005 21:43:15.124666 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a58750e-ca20-4f84-af48-8ceef4967672" containerName="registry-server" Oct 05 21:43:15 crc kubenswrapper[4754]: I1005 21:43:15.126262 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nqjkh" Oct 05 21:43:15 crc kubenswrapper[4754]: I1005 21:43:15.151286 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-nqjkh"] Oct 05 21:43:15 crc kubenswrapper[4754]: I1005 21:43:15.255004 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/188e4cf7-706c-4680-9a62-dd535162a954-utilities\") pod \"certified-operators-nqjkh\" (UID: \"188e4cf7-706c-4680-9a62-dd535162a954\") " pod="openshift-marketplace/certified-operators-nqjkh" Oct 05 21:43:15 crc kubenswrapper[4754]: I1005 21:43:15.255053 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xr9bj\" (UniqueName: \"kubernetes.io/projected/188e4cf7-706c-4680-9a62-dd535162a954-kube-api-access-xr9bj\") pod \"certified-operators-nqjkh\" (UID: \"188e4cf7-706c-4680-9a62-dd535162a954\") " pod="openshift-marketplace/certified-operators-nqjkh" Oct 05 21:43:15 crc kubenswrapper[4754]: I1005 21:43:15.255145 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/188e4cf7-706c-4680-9a62-dd535162a954-catalog-content\") pod \"certified-operators-nqjkh\" (UID: \"188e4cf7-706c-4680-9a62-dd535162a954\") " pod="openshift-marketplace/certified-operators-nqjkh" Oct 05 21:43:15 crc kubenswrapper[4754]: I1005 21:43:15.356437 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/188e4cf7-706c-4680-9a62-dd535162a954-utilities\") pod \"certified-operators-nqjkh\" (UID: \"188e4cf7-706c-4680-9a62-dd535162a954\") " pod="openshift-marketplace/certified-operators-nqjkh" Oct 05 21:43:15 crc kubenswrapper[4754]: I1005 21:43:15.356508 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xr9bj\" (UniqueName: \"kubernetes.io/projected/188e4cf7-706c-4680-9a62-dd535162a954-kube-api-access-xr9bj\") pod \"certified-operators-nqjkh\" (UID: \"188e4cf7-706c-4680-9a62-dd535162a954\") " pod="openshift-marketplace/certified-operators-nqjkh" Oct 05 21:43:15 crc kubenswrapper[4754]: I1005 21:43:15.356545 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/188e4cf7-706c-4680-9a62-dd535162a954-catalog-content\") pod \"certified-operators-nqjkh\" (UID: \"188e4cf7-706c-4680-9a62-dd535162a954\") " pod="openshift-marketplace/certified-operators-nqjkh" Oct 05 21:43:15 crc kubenswrapper[4754]: I1005 21:43:15.357081 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/188e4cf7-706c-4680-9a62-dd535162a954-catalog-content\") pod \"certified-operators-nqjkh\" (UID: \"188e4cf7-706c-4680-9a62-dd535162a954\") " pod="openshift-marketplace/certified-operators-nqjkh" Oct 05 21:43:15 crc kubenswrapper[4754]: I1005 21:43:15.357296 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/188e4cf7-706c-4680-9a62-dd535162a954-utilities\") pod \"certified-operators-nqjkh\" (UID: \"188e4cf7-706c-4680-9a62-dd535162a954\") " pod="openshift-marketplace/certified-operators-nqjkh" Oct 05 21:43:15 crc kubenswrapper[4754]: I1005 21:43:15.379993 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xr9bj\" (UniqueName: \"kubernetes.io/projected/188e4cf7-706c-4680-9a62-dd535162a954-kube-api-access-xr9bj\") pod \"certified-operators-nqjkh\" (UID: \"188e4cf7-706c-4680-9a62-dd535162a954\") " pod="openshift-marketplace/certified-operators-nqjkh" Oct 05 21:43:15 crc kubenswrapper[4754]: I1005 21:43:15.471086 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nqjkh" Oct 05 21:43:15 crc kubenswrapper[4754]: I1005 21:43:15.959634 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-nqjkh"] Oct 05 21:43:15 crc kubenswrapper[4754]: W1005 21:43:15.967927 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod188e4cf7_706c_4680_9a62_dd535162a954.slice/crio-da139ba63b68df6625fb156206232ff285de6120f05e3d74dd6d7143ca4ea6f4 WatchSource:0}: Error finding container da139ba63b68df6625fb156206232ff285de6120f05e3d74dd6d7143ca4ea6f4: Status 404 returned error can't find the container with id da139ba63b68df6625fb156206232ff285de6120f05e3d74dd6d7143ca4ea6f4 Oct 05 21:43:16 crc kubenswrapper[4754]: I1005 21:43:16.257525 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nqjkh" event={"ID":"188e4cf7-706c-4680-9a62-dd535162a954","Type":"ContainerDied","Data":"77c3915d6dd09418bc6a0e38069a14a9a09e78da33aa45b45d5f6cd355a6d974"} Oct 05 21:43:16 crc kubenswrapper[4754]: I1005 21:43:16.256648 4754 generic.go:334] "Generic (PLEG): container finished" podID="188e4cf7-706c-4680-9a62-dd535162a954" containerID="77c3915d6dd09418bc6a0e38069a14a9a09e78da33aa45b45d5f6cd355a6d974" exitCode=0 Oct 05 21:43:16 crc kubenswrapper[4754]: I1005 21:43:16.257870 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nqjkh" event={"ID":"188e4cf7-706c-4680-9a62-dd535162a954","Type":"ContainerStarted","Data":"da139ba63b68df6625fb156206232ff285de6120f05e3d74dd6d7143ca4ea6f4"} Oct 05 21:43:16 crc kubenswrapper[4754]: I1005 21:43:16.259352 4754 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 05 21:43:17 crc kubenswrapper[4754]: I1005 21:43:17.267035 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nqjkh" event={"ID":"188e4cf7-706c-4680-9a62-dd535162a954","Type":"ContainerStarted","Data":"6f6fe5b6702f5bedc0fa4c3c3493eb976f4562beda7c29d8bcf8e40ca54f16f4"} Oct 05 21:43:18 crc kubenswrapper[4754]: I1005 21:43:18.276729 4754 generic.go:334] "Generic (PLEG): container finished" podID="188e4cf7-706c-4680-9a62-dd535162a954" containerID="6f6fe5b6702f5bedc0fa4c3c3493eb976f4562beda7c29d8bcf8e40ca54f16f4" exitCode=0 Oct 05 21:43:18 crc kubenswrapper[4754]: I1005 21:43:18.276797 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nqjkh" event={"ID":"188e4cf7-706c-4680-9a62-dd535162a954","Type":"ContainerDied","Data":"6f6fe5b6702f5bedc0fa4c3c3493eb976f4562beda7c29d8bcf8e40ca54f16f4"} Oct 05 21:43:19 crc kubenswrapper[4754]: I1005 21:43:19.290148 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nqjkh" event={"ID":"188e4cf7-706c-4680-9a62-dd535162a954","Type":"ContainerStarted","Data":"1fc3d2ed6cdd385e2c42af4b82ffb935547057817a922fc2659f49045498c9e5"} Oct 05 21:43:19 crc kubenswrapper[4754]: I1005 21:43:19.320600 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-nqjkh" podStartSLOduration=1.855666901 podStartE2EDuration="4.320578714s" podCreationTimestamp="2025-10-05 21:43:15 +0000 UTC" firstStartedPulling="2025-10-05 21:43:16.258961418 +0000 UTC m=+2920.163080158" lastFinishedPulling="2025-10-05 21:43:18.723873231 +0000 UTC m=+2922.627991971" observedRunningTime="2025-10-05 21:43:19.314384683 +0000 UTC m=+2923.218503423" watchObservedRunningTime="2025-10-05 21:43:19.320578714 +0000 UTC m=+2923.224697434" Oct 05 21:43:25 crc kubenswrapper[4754]: I1005 21:43:25.471640 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-nqjkh" Oct 05 21:43:25 crc kubenswrapper[4754]: I1005 21:43:25.472164 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-nqjkh" Oct 05 21:43:25 crc kubenswrapper[4754]: I1005 21:43:25.569531 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-nqjkh" Oct 05 21:43:26 crc kubenswrapper[4754]: I1005 21:43:26.435918 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-nqjkh" Oct 05 21:43:26 crc kubenswrapper[4754]: I1005 21:43:26.508279 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-nqjkh"] Oct 05 21:43:28 crc kubenswrapper[4754]: I1005 21:43:28.392612 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-nqjkh" podUID="188e4cf7-706c-4680-9a62-dd535162a954" containerName="registry-server" containerID="cri-o://1fc3d2ed6cdd385e2c42af4b82ffb935547057817a922fc2659f49045498c9e5" gracePeriod=2 Oct 05 21:43:28 crc kubenswrapper[4754]: I1005 21:43:28.810281 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nqjkh" Oct 05 21:43:28 crc kubenswrapper[4754]: I1005 21:43:28.962891 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/188e4cf7-706c-4680-9a62-dd535162a954-catalog-content\") pod \"188e4cf7-706c-4680-9a62-dd535162a954\" (UID: \"188e4cf7-706c-4680-9a62-dd535162a954\") " Oct 05 21:43:28 crc kubenswrapper[4754]: I1005 21:43:28.962954 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xr9bj\" (UniqueName: \"kubernetes.io/projected/188e4cf7-706c-4680-9a62-dd535162a954-kube-api-access-xr9bj\") pod \"188e4cf7-706c-4680-9a62-dd535162a954\" (UID: \"188e4cf7-706c-4680-9a62-dd535162a954\") " Oct 05 21:43:28 crc kubenswrapper[4754]: I1005 21:43:28.963008 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/188e4cf7-706c-4680-9a62-dd535162a954-utilities\") pod \"188e4cf7-706c-4680-9a62-dd535162a954\" (UID: \"188e4cf7-706c-4680-9a62-dd535162a954\") " Oct 05 21:43:28 crc kubenswrapper[4754]: I1005 21:43:28.964827 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/188e4cf7-706c-4680-9a62-dd535162a954-utilities" (OuterVolumeSpecName: "utilities") pod "188e4cf7-706c-4680-9a62-dd535162a954" (UID: "188e4cf7-706c-4680-9a62-dd535162a954"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 21:43:28 crc kubenswrapper[4754]: I1005 21:43:28.970627 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/188e4cf7-706c-4680-9a62-dd535162a954-kube-api-access-xr9bj" (OuterVolumeSpecName: "kube-api-access-xr9bj") pod "188e4cf7-706c-4680-9a62-dd535162a954" (UID: "188e4cf7-706c-4680-9a62-dd535162a954"). InnerVolumeSpecName "kube-api-access-xr9bj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:43:29 crc kubenswrapper[4754]: I1005 21:43:29.012968 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/188e4cf7-706c-4680-9a62-dd535162a954-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "188e4cf7-706c-4680-9a62-dd535162a954" (UID: "188e4cf7-706c-4680-9a62-dd535162a954"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 21:43:29 crc kubenswrapper[4754]: I1005 21:43:29.064914 4754 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/188e4cf7-706c-4680-9a62-dd535162a954-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 21:43:29 crc kubenswrapper[4754]: I1005 21:43:29.064949 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xr9bj\" (UniqueName: \"kubernetes.io/projected/188e4cf7-706c-4680-9a62-dd535162a954-kube-api-access-xr9bj\") on node \"crc\" DevicePath \"\"" Oct 05 21:43:29 crc kubenswrapper[4754]: I1005 21:43:29.064962 4754 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/188e4cf7-706c-4680-9a62-dd535162a954-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 21:43:29 crc kubenswrapper[4754]: I1005 21:43:29.405216 4754 generic.go:334] "Generic (PLEG): container finished" podID="188e4cf7-706c-4680-9a62-dd535162a954" containerID="1fc3d2ed6cdd385e2c42af4b82ffb935547057817a922fc2659f49045498c9e5" exitCode=0 Oct 05 21:43:29 crc kubenswrapper[4754]: I1005 21:43:29.405291 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nqjkh" event={"ID":"188e4cf7-706c-4680-9a62-dd535162a954","Type":"ContainerDied","Data":"1fc3d2ed6cdd385e2c42af4b82ffb935547057817a922fc2659f49045498c9e5"} Oct 05 21:43:29 crc kubenswrapper[4754]: I1005 21:43:29.405323 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nqjkh" Oct 05 21:43:29 crc kubenswrapper[4754]: I1005 21:43:29.406072 4754 scope.go:117] "RemoveContainer" containerID="1fc3d2ed6cdd385e2c42af4b82ffb935547057817a922fc2659f49045498c9e5" Oct 05 21:43:29 crc kubenswrapper[4754]: I1005 21:43:29.405991 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nqjkh" event={"ID":"188e4cf7-706c-4680-9a62-dd535162a954","Type":"ContainerDied","Data":"da139ba63b68df6625fb156206232ff285de6120f05e3d74dd6d7143ca4ea6f4"} Oct 05 21:43:29 crc kubenswrapper[4754]: I1005 21:43:29.440936 4754 scope.go:117] "RemoveContainer" containerID="6f6fe5b6702f5bedc0fa4c3c3493eb976f4562beda7c29d8bcf8e40ca54f16f4" Oct 05 21:43:29 crc kubenswrapper[4754]: I1005 21:43:29.463346 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-nqjkh"] Oct 05 21:43:29 crc kubenswrapper[4754]: I1005 21:43:29.471619 4754 scope.go:117] "RemoveContainer" containerID="77c3915d6dd09418bc6a0e38069a14a9a09e78da33aa45b45d5f6cd355a6d974" Oct 05 21:43:29 crc kubenswrapper[4754]: I1005 21:43:29.477214 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-nqjkh"] Oct 05 21:43:29 crc kubenswrapper[4754]: I1005 21:43:29.522797 4754 scope.go:117] "RemoveContainer" containerID="1fc3d2ed6cdd385e2c42af4b82ffb935547057817a922fc2659f49045498c9e5" Oct 05 21:43:29 crc kubenswrapper[4754]: E1005 21:43:29.523767 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1fc3d2ed6cdd385e2c42af4b82ffb935547057817a922fc2659f49045498c9e5\": container with ID starting with 1fc3d2ed6cdd385e2c42af4b82ffb935547057817a922fc2659f49045498c9e5 not found: ID does not exist" containerID="1fc3d2ed6cdd385e2c42af4b82ffb935547057817a922fc2659f49045498c9e5" Oct 05 21:43:29 crc kubenswrapper[4754]: I1005 21:43:29.523845 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1fc3d2ed6cdd385e2c42af4b82ffb935547057817a922fc2659f49045498c9e5"} err="failed to get container status \"1fc3d2ed6cdd385e2c42af4b82ffb935547057817a922fc2659f49045498c9e5\": rpc error: code = NotFound desc = could not find container \"1fc3d2ed6cdd385e2c42af4b82ffb935547057817a922fc2659f49045498c9e5\": container with ID starting with 1fc3d2ed6cdd385e2c42af4b82ffb935547057817a922fc2659f49045498c9e5 not found: ID does not exist" Oct 05 21:43:29 crc kubenswrapper[4754]: I1005 21:43:29.523902 4754 scope.go:117] "RemoveContainer" containerID="6f6fe5b6702f5bedc0fa4c3c3493eb976f4562beda7c29d8bcf8e40ca54f16f4" Oct 05 21:43:29 crc kubenswrapper[4754]: E1005 21:43:29.524348 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6f6fe5b6702f5bedc0fa4c3c3493eb976f4562beda7c29d8bcf8e40ca54f16f4\": container with ID starting with 6f6fe5b6702f5bedc0fa4c3c3493eb976f4562beda7c29d8bcf8e40ca54f16f4 not found: ID does not exist" containerID="6f6fe5b6702f5bedc0fa4c3c3493eb976f4562beda7c29d8bcf8e40ca54f16f4" Oct 05 21:43:29 crc kubenswrapper[4754]: I1005 21:43:29.524419 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6f6fe5b6702f5bedc0fa4c3c3493eb976f4562beda7c29d8bcf8e40ca54f16f4"} err="failed to get container status \"6f6fe5b6702f5bedc0fa4c3c3493eb976f4562beda7c29d8bcf8e40ca54f16f4\": rpc error: code = NotFound desc = could not find container \"6f6fe5b6702f5bedc0fa4c3c3493eb976f4562beda7c29d8bcf8e40ca54f16f4\": container with ID starting with 6f6fe5b6702f5bedc0fa4c3c3493eb976f4562beda7c29d8bcf8e40ca54f16f4 not found: ID does not exist" Oct 05 21:43:29 crc kubenswrapper[4754]: I1005 21:43:29.524470 4754 scope.go:117] "RemoveContainer" containerID="77c3915d6dd09418bc6a0e38069a14a9a09e78da33aa45b45d5f6cd355a6d974" Oct 05 21:43:29 crc kubenswrapper[4754]: E1005 21:43:29.524927 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"77c3915d6dd09418bc6a0e38069a14a9a09e78da33aa45b45d5f6cd355a6d974\": container with ID starting with 77c3915d6dd09418bc6a0e38069a14a9a09e78da33aa45b45d5f6cd355a6d974 not found: ID does not exist" containerID="77c3915d6dd09418bc6a0e38069a14a9a09e78da33aa45b45d5f6cd355a6d974" Oct 05 21:43:29 crc kubenswrapper[4754]: I1005 21:43:29.525023 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"77c3915d6dd09418bc6a0e38069a14a9a09e78da33aa45b45d5f6cd355a6d974"} err="failed to get container status \"77c3915d6dd09418bc6a0e38069a14a9a09e78da33aa45b45d5f6cd355a6d974\": rpc error: code = NotFound desc = could not find container \"77c3915d6dd09418bc6a0e38069a14a9a09e78da33aa45b45d5f6cd355a6d974\": container with ID starting with 77c3915d6dd09418bc6a0e38069a14a9a09e78da33aa45b45d5f6cd355a6d974 not found: ID does not exist" Oct 05 21:43:30 crc kubenswrapper[4754]: I1005 21:43:30.853308 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="188e4cf7-706c-4680-9a62-dd535162a954" path="/var/lib/kubelet/pods/188e4cf7-706c-4680-9a62-dd535162a954/volumes" Oct 05 21:43:35 crc kubenswrapper[4754]: I1005 21:43:35.244951 4754 patch_prober.go:28] interesting pod/machine-config-daemon-b2h9k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 21:43:35 crc kubenswrapper[4754]: I1005 21:43:35.246146 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 21:43:35 crc kubenswrapper[4754]: I1005 21:43:35.246217 4754 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" Oct 05 21:43:35 crc kubenswrapper[4754]: I1005 21:43:35.247610 4754 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1bb4b1f5def9e89ddfea5b5e0921b99d55f1ceb68868b957d9868c33a0211da3"} pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 05 21:43:35 crc kubenswrapper[4754]: I1005 21:43:35.247721 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" containerName="machine-config-daemon" containerID="cri-o://1bb4b1f5def9e89ddfea5b5e0921b99d55f1ceb68868b957d9868c33a0211da3" gracePeriod=600 Oct 05 21:43:35 crc kubenswrapper[4754]: E1005 21:43:35.384279 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 21:43:35 crc kubenswrapper[4754]: I1005 21:43:35.469096 4754 generic.go:334] "Generic (PLEG): container finished" podID="be95a413-6a59-45b4-84b7-b43cae694a26" containerID="1bb4b1f5def9e89ddfea5b5e0921b99d55f1ceb68868b957d9868c33a0211da3" exitCode=0 Oct 05 21:43:35 crc kubenswrapper[4754]: I1005 21:43:35.469152 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" event={"ID":"be95a413-6a59-45b4-84b7-b43cae694a26","Type":"ContainerDied","Data":"1bb4b1f5def9e89ddfea5b5e0921b99d55f1ceb68868b957d9868c33a0211da3"} Oct 05 21:43:35 crc kubenswrapper[4754]: I1005 21:43:35.469194 4754 scope.go:117] "RemoveContainer" containerID="5134bc33c72b286f6d3cee42c9ef42b9498b3009c16d4989384d741eded7eda2" Oct 05 21:43:35 crc kubenswrapper[4754]: I1005 21:43:35.471791 4754 scope.go:117] "RemoveContainer" containerID="1bb4b1f5def9e89ddfea5b5e0921b99d55f1ceb68868b957d9868c33a0211da3" Oct 05 21:43:35 crc kubenswrapper[4754]: E1005 21:43:35.472330 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 21:43:47 crc kubenswrapper[4754]: I1005 21:43:47.838747 4754 scope.go:117] "RemoveContainer" containerID="1bb4b1f5def9e89ddfea5b5e0921b99d55f1ceb68868b957d9868c33a0211da3" Oct 05 21:43:47 crc kubenswrapper[4754]: E1005 21:43:47.840684 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 21:43:59 crc kubenswrapper[4754]: I1005 21:43:59.839106 4754 scope.go:117] "RemoveContainer" containerID="1bb4b1f5def9e89ddfea5b5e0921b99d55f1ceb68868b957d9868c33a0211da3" Oct 05 21:43:59 crc kubenswrapper[4754]: E1005 21:43:59.840280 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 21:44:12 crc kubenswrapper[4754]: I1005 21:44:12.838133 4754 scope.go:117] "RemoveContainer" containerID="1bb4b1f5def9e89ddfea5b5e0921b99d55f1ceb68868b957d9868c33a0211da3" Oct 05 21:44:12 crc kubenswrapper[4754]: E1005 21:44:12.839139 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 21:44:24 crc kubenswrapper[4754]: I1005 21:44:24.838119 4754 scope.go:117] "RemoveContainer" containerID="1bb4b1f5def9e89ddfea5b5e0921b99d55f1ceb68868b957d9868c33a0211da3" Oct 05 21:44:24 crc kubenswrapper[4754]: E1005 21:44:24.839206 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 21:44:35 crc kubenswrapper[4754]: I1005 21:44:35.837709 4754 scope.go:117] "RemoveContainer" containerID="1bb4b1f5def9e89ddfea5b5e0921b99d55f1ceb68868b957d9868c33a0211da3" Oct 05 21:44:35 crc kubenswrapper[4754]: E1005 21:44:35.838980 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 21:44:47 crc kubenswrapper[4754]: I1005 21:44:47.837206 4754 scope.go:117] "RemoveContainer" containerID="1bb4b1f5def9e89ddfea5b5e0921b99d55f1ceb68868b957d9868c33a0211da3" Oct 05 21:44:47 crc kubenswrapper[4754]: E1005 21:44:47.838895 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 21:44:58 crc kubenswrapper[4754]: I1005 21:44:58.837264 4754 scope.go:117] "RemoveContainer" containerID="1bb4b1f5def9e89ddfea5b5e0921b99d55f1ceb68868b957d9868c33a0211da3" Oct 05 21:44:58 crc kubenswrapper[4754]: E1005 21:44:58.838136 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 21:45:00 crc kubenswrapper[4754]: I1005 21:45:00.205033 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29328345-jz4bf"] Oct 05 21:45:00 crc kubenswrapper[4754]: E1005 21:45:00.206931 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="188e4cf7-706c-4680-9a62-dd535162a954" containerName="registry-server" Oct 05 21:45:00 crc kubenswrapper[4754]: I1005 21:45:00.206985 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="188e4cf7-706c-4680-9a62-dd535162a954" containerName="registry-server" Oct 05 21:45:00 crc kubenswrapper[4754]: E1005 21:45:00.207033 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="188e4cf7-706c-4680-9a62-dd535162a954" containerName="extract-content" Oct 05 21:45:00 crc kubenswrapper[4754]: I1005 21:45:00.207046 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="188e4cf7-706c-4680-9a62-dd535162a954" containerName="extract-content" Oct 05 21:45:00 crc kubenswrapper[4754]: E1005 21:45:00.207077 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="188e4cf7-706c-4680-9a62-dd535162a954" containerName="extract-utilities" Oct 05 21:45:00 crc kubenswrapper[4754]: I1005 21:45:00.207090 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="188e4cf7-706c-4680-9a62-dd535162a954" containerName="extract-utilities" Oct 05 21:45:00 crc kubenswrapper[4754]: I1005 21:45:00.207451 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="188e4cf7-706c-4680-9a62-dd535162a954" containerName="registry-server" Oct 05 21:45:00 crc kubenswrapper[4754]: I1005 21:45:00.208819 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29328345-jz4bf" Oct 05 21:45:00 crc kubenswrapper[4754]: I1005 21:45:00.218235 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 05 21:45:00 crc kubenswrapper[4754]: I1005 21:45:00.218696 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 05 21:45:00 crc kubenswrapper[4754]: I1005 21:45:00.222272 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29328345-jz4bf"] Oct 05 21:45:00 crc kubenswrapper[4754]: I1005 21:45:00.321964 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bpst9\" (UniqueName: \"kubernetes.io/projected/e18ed4e0-39b3-4ee1-9bb4-13fa1817f46b-kube-api-access-bpst9\") pod \"collect-profiles-29328345-jz4bf\" (UID: \"e18ed4e0-39b3-4ee1-9bb4-13fa1817f46b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29328345-jz4bf" Oct 05 21:45:00 crc kubenswrapper[4754]: I1005 21:45:00.322037 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e18ed4e0-39b3-4ee1-9bb4-13fa1817f46b-secret-volume\") pod \"collect-profiles-29328345-jz4bf\" (UID: \"e18ed4e0-39b3-4ee1-9bb4-13fa1817f46b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29328345-jz4bf" Oct 05 21:45:00 crc kubenswrapper[4754]: I1005 21:45:00.322671 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e18ed4e0-39b3-4ee1-9bb4-13fa1817f46b-config-volume\") pod \"collect-profiles-29328345-jz4bf\" (UID: \"e18ed4e0-39b3-4ee1-9bb4-13fa1817f46b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29328345-jz4bf" Oct 05 21:45:00 crc kubenswrapper[4754]: I1005 21:45:00.424866 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bpst9\" (UniqueName: \"kubernetes.io/projected/e18ed4e0-39b3-4ee1-9bb4-13fa1817f46b-kube-api-access-bpst9\") pod \"collect-profiles-29328345-jz4bf\" (UID: \"e18ed4e0-39b3-4ee1-9bb4-13fa1817f46b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29328345-jz4bf" Oct 05 21:45:00 crc kubenswrapper[4754]: I1005 21:45:00.424948 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e18ed4e0-39b3-4ee1-9bb4-13fa1817f46b-secret-volume\") pod \"collect-profiles-29328345-jz4bf\" (UID: \"e18ed4e0-39b3-4ee1-9bb4-13fa1817f46b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29328345-jz4bf" Oct 05 21:45:00 crc kubenswrapper[4754]: I1005 21:45:00.425181 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e18ed4e0-39b3-4ee1-9bb4-13fa1817f46b-config-volume\") pod \"collect-profiles-29328345-jz4bf\" (UID: \"e18ed4e0-39b3-4ee1-9bb4-13fa1817f46b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29328345-jz4bf" Oct 05 21:45:00 crc kubenswrapper[4754]: I1005 21:45:00.426762 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e18ed4e0-39b3-4ee1-9bb4-13fa1817f46b-config-volume\") pod \"collect-profiles-29328345-jz4bf\" (UID: \"e18ed4e0-39b3-4ee1-9bb4-13fa1817f46b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29328345-jz4bf" Oct 05 21:45:00 crc kubenswrapper[4754]: I1005 21:45:00.436628 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e18ed4e0-39b3-4ee1-9bb4-13fa1817f46b-secret-volume\") pod \"collect-profiles-29328345-jz4bf\" (UID: \"e18ed4e0-39b3-4ee1-9bb4-13fa1817f46b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29328345-jz4bf" Oct 05 21:45:00 crc kubenswrapper[4754]: I1005 21:45:00.446957 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bpst9\" (UniqueName: \"kubernetes.io/projected/e18ed4e0-39b3-4ee1-9bb4-13fa1817f46b-kube-api-access-bpst9\") pod \"collect-profiles-29328345-jz4bf\" (UID: \"e18ed4e0-39b3-4ee1-9bb4-13fa1817f46b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29328345-jz4bf" Oct 05 21:45:00 crc kubenswrapper[4754]: I1005 21:45:00.552262 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29328345-jz4bf" Oct 05 21:45:01 crc kubenswrapper[4754]: I1005 21:45:01.042792 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29328345-jz4bf"] Oct 05 21:45:01 crc kubenswrapper[4754]: W1005 21:45:01.057897 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode18ed4e0_39b3_4ee1_9bb4_13fa1817f46b.slice/crio-49a9edc89da3f83c82b91a69ec3e22a1f734f5323114147ce4cbf34f9a02970b WatchSource:0}: Error finding container 49a9edc89da3f83c82b91a69ec3e22a1f734f5323114147ce4cbf34f9a02970b: Status 404 returned error can't find the container with id 49a9edc89da3f83c82b91a69ec3e22a1f734f5323114147ce4cbf34f9a02970b Oct 05 21:45:01 crc kubenswrapper[4754]: I1005 21:45:01.464116 4754 generic.go:334] "Generic (PLEG): container finished" podID="e18ed4e0-39b3-4ee1-9bb4-13fa1817f46b" containerID="9492d779d7b02bfcd4a795f369c16a9291cd72ffa0b89d750fe4a56c67873f64" exitCode=0 Oct 05 21:45:01 crc kubenswrapper[4754]: I1005 21:45:01.464190 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29328345-jz4bf" event={"ID":"e18ed4e0-39b3-4ee1-9bb4-13fa1817f46b","Type":"ContainerDied","Data":"9492d779d7b02bfcd4a795f369c16a9291cd72ffa0b89d750fe4a56c67873f64"} Oct 05 21:45:01 crc kubenswrapper[4754]: I1005 21:45:01.464466 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29328345-jz4bf" event={"ID":"e18ed4e0-39b3-4ee1-9bb4-13fa1817f46b","Type":"ContainerStarted","Data":"49a9edc89da3f83c82b91a69ec3e22a1f734f5323114147ce4cbf34f9a02970b"} Oct 05 21:45:02 crc kubenswrapper[4754]: I1005 21:45:02.780058 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29328345-jz4bf" Oct 05 21:45:02 crc kubenswrapper[4754]: I1005 21:45:02.901988 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e18ed4e0-39b3-4ee1-9bb4-13fa1817f46b-config-volume\") pod \"e18ed4e0-39b3-4ee1-9bb4-13fa1817f46b\" (UID: \"e18ed4e0-39b3-4ee1-9bb4-13fa1817f46b\") " Oct 05 21:45:02 crc kubenswrapper[4754]: I1005 21:45:02.902160 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e18ed4e0-39b3-4ee1-9bb4-13fa1817f46b-secret-volume\") pod \"e18ed4e0-39b3-4ee1-9bb4-13fa1817f46b\" (UID: \"e18ed4e0-39b3-4ee1-9bb4-13fa1817f46b\") " Oct 05 21:45:02 crc kubenswrapper[4754]: I1005 21:45:02.902199 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bpst9\" (UniqueName: \"kubernetes.io/projected/e18ed4e0-39b3-4ee1-9bb4-13fa1817f46b-kube-api-access-bpst9\") pod \"e18ed4e0-39b3-4ee1-9bb4-13fa1817f46b\" (UID: \"e18ed4e0-39b3-4ee1-9bb4-13fa1817f46b\") " Oct 05 21:45:02 crc kubenswrapper[4754]: I1005 21:45:02.903016 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e18ed4e0-39b3-4ee1-9bb4-13fa1817f46b-config-volume" (OuterVolumeSpecName: "config-volume") pod "e18ed4e0-39b3-4ee1-9bb4-13fa1817f46b" (UID: "e18ed4e0-39b3-4ee1-9bb4-13fa1817f46b"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 21:45:02 crc kubenswrapper[4754]: I1005 21:45:02.909792 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e18ed4e0-39b3-4ee1-9bb4-13fa1817f46b-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "e18ed4e0-39b3-4ee1-9bb4-13fa1817f46b" (UID: "e18ed4e0-39b3-4ee1-9bb4-13fa1817f46b"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:45:02 crc kubenswrapper[4754]: I1005 21:45:02.909895 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e18ed4e0-39b3-4ee1-9bb4-13fa1817f46b-kube-api-access-bpst9" (OuterVolumeSpecName: "kube-api-access-bpst9") pod "e18ed4e0-39b3-4ee1-9bb4-13fa1817f46b" (UID: "e18ed4e0-39b3-4ee1-9bb4-13fa1817f46b"). InnerVolumeSpecName "kube-api-access-bpst9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:45:03 crc kubenswrapper[4754]: I1005 21:45:03.004566 4754 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e18ed4e0-39b3-4ee1-9bb4-13fa1817f46b-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 05 21:45:03 crc kubenswrapper[4754]: I1005 21:45:03.004842 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bpst9\" (UniqueName: \"kubernetes.io/projected/e18ed4e0-39b3-4ee1-9bb4-13fa1817f46b-kube-api-access-bpst9\") on node \"crc\" DevicePath \"\"" Oct 05 21:45:03 crc kubenswrapper[4754]: I1005 21:45:03.004853 4754 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e18ed4e0-39b3-4ee1-9bb4-13fa1817f46b-config-volume\") on node \"crc\" DevicePath \"\"" Oct 05 21:45:03 crc kubenswrapper[4754]: I1005 21:45:03.488230 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29328345-jz4bf" event={"ID":"e18ed4e0-39b3-4ee1-9bb4-13fa1817f46b","Type":"ContainerDied","Data":"49a9edc89da3f83c82b91a69ec3e22a1f734f5323114147ce4cbf34f9a02970b"} Oct 05 21:45:03 crc kubenswrapper[4754]: I1005 21:45:03.488290 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="49a9edc89da3f83c82b91a69ec3e22a1f734f5323114147ce4cbf34f9a02970b" Oct 05 21:45:03 crc kubenswrapper[4754]: I1005 21:45:03.488297 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29328345-jz4bf" Oct 05 21:45:03 crc kubenswrapper[4754]: I1005 21:45:03.867296 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29328300-gkv6p"] Oct 05 21:45:03 crc kubenswrapper[4754]: I1005 21:45:03.874810 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29328300-gkv6p"] Oct 05 21:45:04 crc kubenswrapper[4754]: I1005 21:45:04.880637 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e2c21c-e2af-4b3a-b990-76e7d4128b9f" path="/var/lib/kubelet/pods/25e2c21c-e2af-4b3a-b990-76e7d4128b9f/volumes" Oct 05 21:45:13 crc kubenswrapper[4754]: I1005 21:45:13.840419 4754 scope.go:117] "RemoveContainer" containerID="1bb4b1f5def9e89ddfea5b5e0921b99d55f1ceb68868b957d9868c33a0211da3" Oct 05 21:45:13 crc kubenswrapper[4754]: E1005 21:45:13.841214 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 21:45:28 crc kubenswrapper[4754]: I1005 21:45:28.841066 4754 scope.go:117] "RemoveContainer" containerID="1bb4b1f5def9e89ddfea5b5e0921b99d55f1ceb68868b957d9868c33a0211da3" Oct 05 21:45:28 crc kubenswrapper[4754]: E1005 21:45:28.842441 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 21:45:33 crc kubenswrapper[4754]: I1005 21:45:33.802392 4754 generic.go:334] "Generic (PLEG): container finished" podID="7c094190-76ba-4310-ae5d-1e82d3caeac2" containerID="8335deacab79709ecaa34accbe0b8d24f3377c8107540e1746ab2269e7b31fcc" exitCode=0 Oct 05 21:45:33 crc kubenswrapper[4754]: I1005 21:45:33.802468 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s2xqb" event={"ID":"7c094190-76ba-4310-ae5d-1e82d3caeac2","Type":"ContainerDied","Data":"8335deacab79709ecaa34accbe0b8d24f3377c8107540e1746ab2269e7b31fcc"} Oct 05 21:45:35 crc kubenswrapper[4754]: I1005 21:45:35.321400 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s2xqb" Oct 05 21:45:35 crc kubenswrapper[4754]: I1005 21:45:35.424255 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/7c094190-76ba-4310-ae5d-1e82d3caeac2-ceilometer-compute-config-data-2\") pod \"7c094190-76ba-4310-ae5d-1e82d3caeac2\" (UID: \"7c094190-76ba-4310-ae5d-1e82d3caeac2\") " Oct 05 21:45:35 crc kubenswrapper[4754]: I1005 21:45:35.424803 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/7c094190-76ba-4310-ae5d-1e82d3caeac2-ceilometer-compute-config-data-1\") pod \"7c094190-76ba-4310-ae5d-1e82d3caeac2\" (UID: \"7c094190-76ba-4310-ae5d-1e82d3caeac2\") " Oct 05 21:45:35 crc kubenswrapper[4754]: I1005 21:45:35.424849 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c094190-76ba-4310-ae5d-1e82d3caeac2-telemetry-combined-ca-bundle\") pod \"7c094190-76ba-4310-ae5d-1e82d3caeac2\" (UID: \"7c094190-76ba-4310-ae5d-1e82d3caeac2\") " Oct 05 21:45:35 crc kubenswrapper[4754]: I1005 21:45:35.424937 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7c094190-76ba-4310-ae5d-1e82d3caeac2-inventory\") pod \"7c094190-76ba-4310-ae5d-1e82d3caeac2\" (UID: \"7c094190-76ba-4310-ae5d-1e82d3caeac2\") " Oct 05 21:45:35 crc kubenswrapper[4754]: I1005 21:45:35.425030 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d8j48\" (UniqueName: \"kubernetes.io/projected/7c094190-76ba-4310-ae5d-1e82d3caeac2-kube-api-access-d8j48\") pod \"7c094190-76ba-4310-ae5d-1e82d3caeac2\" (UID: \"7c094190-76ba-4310-ae5d-1e82d3caeac2\") " Oct 05 21:45:35 crc kubenswrapper[4754]: I1005 21:45:35.425103 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7c094190-76ba-4310-ae5d-1e82d3caeac2-ssh-key\") pod \"7c094190-76ba-4310-ae5d-1e82d3caeac2\" (UID: \"7c094190-76ba-4310-ae5d-1e82d3caeac2\") " Oct 05 21:45:35 crc kubenswrapper[4754]: I1005 21:45:35.425196 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/7c094190-76ba-4310-ae5d-1e82d3caeac2-ceilometer-compute-config-data-0\") pod \"7c094190-76ba-4310-ae5d-1e82d3caeac2\" (UID: \"7c094190-76ba-4310-ae5d-1e82d3caeac2\") " Oct 05 21:45:35 crc kubenswrapper[4754]: I1005 21:45:35.435814 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7c094190-76ba-4310-ae5d-1e82d3caeac2-kube-api-access-d8j48" (OuterVolumeSpecName: "kube-api-access-d8j48") pod "7c094190-76ba-4310-ae5d-1e82d3caeac2" (UID: "7c094190-76ba-4310-ae5d-1e82d3caeac2"). InnerVolumeSpecName "kube-api-access-d8j48". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:45:35 crc kubenswrapper[4754]: I1005 21:45:35.440186 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7c094190-76ba-4310-ae5d-1e82d3caeac2-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "7c094190-76ba-4310-ae5d-1e82d3caeac2" (UID: "7c094190-76ba-4310-ae5d-1e82d3caeac2"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:45:35 crc kubenswrapper[4754]: I1005 21:45:35.457323 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7c094190-76ba-4310-ae5d-1e82d3caeac2-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "7c094190-76ba-4310-ae5d-1e82d3caeac2" (UID: "7c094190-76ba-4310-ae5d-1e82d3caeac2"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:45:35 crc kubenswrapper[4754]: I1005 21:45:35.462806 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7c094190-76ba-4310-ae5d-1e82d3caeac2-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "7c094190-76ba-4310-ae5d-1e82d3caeac2" (UID: "7c094190-76ba-4310-ae5d-1e82d3caeac2"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:45:35 crc kubenswrapper[4754]: I1005 21:45:35.466264 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7c094190-76ba-4310-ae5d-1e82d3caeac2-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "7c094190-76ba-4310-ae5d-1e82d3caeac2" (UID: "7c094190-76ba-4310-ae5d-1e82d3caeac2"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:45:35 crc kubenswrapper[4754]: I1005 21:45:35.472637 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7c094190-76ba-4310-ae5d-1e82d3caeac2-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "7c094190-76ba-4310-ae5d-1e82d3caeac2" (UID: "7c094190-76ba-4310-ae5d-1e82d3caeac2"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:45:35 crc kubenswrapper[4754]: I1005 21:45:35.476017 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7c094190-76ba-4310-ae5d-1e82d3caeac2-inventory" (OuterVolumeSpecName: "inventory") pod "7c094190-76ba-4310-ae5d-1e82d3caeac2" (UID: "7c094190-76ba-4310-ae5d-1e82d3caeac2"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 21:45:35 crc kubenswrapper[4754]: I1005 21:45:35.527587 4754 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/7c094190-76ba-4310-ae5d-1e82d3caeac2-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Oct 05 21:45:35 crc kubenswrapper[4754]: I1005 21:45:35.527622 4754 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c094190-76ba-4310-ae5d-1e82d3caeac2-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 21:45:35 crc kubenswrapper[4754]: I1005 21:45:35.527638 4754 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7c094190-76ba-4310-ae5d-1e82d3caeac2-inventory\") on node \"crc\" DevicePath \"\"" Oct 05 21:45:35 crc kubenswrapper[4754]: I1005 21:45:35.527649 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d8j48\" (UniqueName: \"kubernetes.io/projected/7c094190-76ba-4310-ae5d-1e82d3caeac2-kube-api-access-d8j48\") on node \"crc\" DevicePath \"\"" Oct 05 21:45:35 crc kubenswrapper[4754]: I1005 21:45:35.527660 4754 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7c094190-76ba-4310-ae5d-1e82d3caeac2-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 05 21:45:35 crc kubenswrapper[4754]: I1005 21:45:35.527670 4754 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/7c094190-76ba-4310-ae5d-1e82d3caeac2-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Oct 05 21:45:35 crc kubenswrapper[4754]: I1005 21:45:35.529289 4754 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/7c094190-76ba-4310-ae5d-1e82d3caeac2-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Oct 05 21:45:35 crc kubenswrapper[4754]: I1005 21:45:35.831430 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s2xqb" event={"ID":"7c094190-76ba-4310-ae5d-1e82d3caeac2","Type":"ContainerDied","Data":"f6bf9abe0ec81763ca0e47f688310bb99684bf64aa0379e8dc9cae3ed457c336"} Oct 05 21:45:35 crc kubenswrapper[4754]: I1005 21:45:35.831481 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f6bf9abe0ec81763ca0e47f688310bb99684bf64aa0379e8dc9cae3ed457c336" Oct 05 21:45:35 crc kubenswrapper[4754]: I1005 21:45:35.831569 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s2xqb" Oct 05 21:45:40 crc kubenswrapper[4754]: I1005 21:45:40.837702 4754 scope.go:117] "RemoveContainer" containerID="1bb4b1f5def9e89ddfea5b5e0921b99d55f1ceb68868b957d9868c33a0211da3" Oct 05 21:45:40 crc kubenswrapper[4754]: E1005 21:45:40.838657 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 21:45:51 crc kubenswrapper[4754]: I1005 21:45:51.838164 4754 scope.go:117] "RemoveContainer" containerID="1bb4b1f5def9e89ddfea5b5e0921b99d55f1ceb68868b957d9868c33a0211da3" Oct 05 21:45:51 crc kubenswrapper[4754]: E1005 21:45:51.839257 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 21:45:56 crc kubenswrapper[4754]: I1005 21:45:56.326191 4754 scope.go:117] "RemoveContainer" containerID="27e07597c80288def0893b4b8459ca7506f59fa2521265c1270b55514ae6c14e" Oct 05 21:46:05 crc kubenswrapper[4754]: I1005 21:46:05.847237 4754 scope.go:117] "RemoveContainer" containerID="1bb4b1f5def9e89ddfea5b5e0921b99d55f1ceb68868b957d9868c33a0211da3" Oct 05 21:46:05 crc kubenswrapper[4754]: E1005 21:46:05.848425 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 21:46:20 crc kubenswrapper[4754]: I1005 21:46:20.361225 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-g4jzc"] Oct 05 21:46:20 crc kubenswrapper[4754]: E1005 21:46:20.370217 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c094190-76ba-4310-ae5d-1e82d3caeac2" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Oct 05 21:46:20 crc kubenswrapper[4754]: I1005 21:46:20.370361 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c094190-76ba-4310-ae5d-1e82d3caeac2" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Oct 05 21:46:20 crc kubenswrapper[4754]: E1005 21:46:20.370527 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e18ed4e0-39b3-4ee1-9bb4-13fa1817f46b" containerName="collect-profiles" Oct 05 21:46:20 crc kubenswrapper[4754]: I1005 21:46:20.370615 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="e18ed4e0-39b3-4ee1-9bb4-13fa1817f46b" containerName="collect-profiles" Oct 05 21:46:20 crc kubenswrapper[4754]: I1005 21:46:20.370958 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="e18ed4e0-39b3-4ee1-9bb4-13fa1817f46b" containerName="collect-profiles" Oct 05 21:46:20 crc kubenswrapper[4754]: I1005 21:46:20.371360 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="7c094190-76ba-4310-ae5d-1e82d3caeac2" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Oct 05 21:46:20 crc kubenswrapper[4754]: I1005 21:46:20.373363 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-g4jzc" Oct 05 21:46:20 crc kubenswrapper[4754]: I1005 21:46:20.381079 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-g4jzc"] Oct 05 21:46:20 crc kubenswrapper[4754]: I1005 21:46:20.538892 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/be29e094-c72e-4ca0-9f8f-d9234767ed71-utilities\") pod \"community-operators-g4jzc\" (UID: \"be29e094-c72e-4ca0-9f8f-d9234767ed71\") " pod="openshift-marketplace/community-operators-g4jzc" Oct 05 21:46:20 crc kubenswrapper[4754]: I1005 21:46:20.539452 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rrd2j\" (UniqueName: \"kubernetes.io/projected/be29e094-c72e-4ca0-9f8f-d9234767ed71-kube-api-access-rrd2j\") pod \"community-operators-g4jzc\" (UID: \"be29e094-c72e-4ca0-9f8f-d9234767ed71\") " pod="openshift-marketplace/community-operators-g4jzc" Oct 05 21:46:20 crc kubenswrapper[4754]: I1005 21:46:20.539647 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/be29e094-c72e-4ca0-9f8f-d9234767ed71-catalog-content\") pod \"community-operators-g4jzc\" (UID: \"be29e094-c72e-4ca0-9f8f-d9234767ed71\") " pod="openshift-marketplace/community-operators-g4jzc" Oct 05 21:46:20 crc kubenswrapper[4754]: I1005 21:46:20.641968 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/be29e094-c72e-4ca0-9f8f-d9234767ed71-catalog-content\") pod \"community-operators-g4jzc\" (UID: \"be29e094-c72e-4ca0-9f8f-d9234767ed71\") " pod="openshift-marketplace/community-operators-g4jzc" Oct 05 21:46:20 crc kubenswrapper[4754]: I1005 21:46:20.642014 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/be29e094-c72e-4ca0-9f8f-d9234767ed71-utilities\") pod \"community-operators-g4jzc\" (UID: \"be29e094-c72e-4ca0-9f8f-d9234767ed71\") " pod="openshift-marketplace/community-operators-g4jzc" Oct 05 21:46:20 crc kubenswrapper[4754]: I1005 21:46:20.642155 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rrd2j\" (UniqueName: \"kubernetes.io/projected/be29e094-c72e-4ca0-9f8f-d9234767ed71-kube-api-access-rrd2j\") pod \"community-operators-g4jzc\" (UID: \"be29e094-c72e-4ca0-9f8f-d9234767ed71\") " pod="openshift-marketplace/community-operators-g4jzc" Oct 05 21:46:20 crc kubenswrapper[4754]: I1005 21:46:20.642457 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/be29e094-c72e-4ca0-9f8f-d9234767ed71-catalog-content\") pod \"community-operators-g4jzc\" (UID: \"be29e094-c72e-4ca0-9f8f-d9234767ed71\") " pod="openshift-marketplace/community-operators-g4jzc" Oct 05 21:46:20 crc kubenswrapper[4754]: I1005 21:46:20.642945 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/be29e094-c72e-4ca0-9f8f-d9234767ed71-utilities\") pod \"community-operators-g4jzc\" (UID: \"be29e094-c72e-4ca0-9f8f-d9234767ed71\") " pod="openshift-marketplace/community-operators-g4jzc" Oct 05 21:46:20 crc kubenswrapper[4754]: I1005 21:46:20.664929 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rrd2j\" (UniqueName: \"kubernetes.io/projected/be29e094-c72e-4ca0-9f8f-d9234767ed71-kube-api-access-rrd2j\") pod \"community-operators-g4jzc\" (UID: \"be29e094-c72e-4ca0-9f8f-d9234767ed71\") " pod="openshift-marketplace/community-operators-g4jzc" Oct 05 21:46:20 crc kubenswrapper[4754]: I1005 21:46:20.756559 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-g4jzc" Oct 05 21:46:20 crc kubenswrapper[4754]: I1005 21:46:20.842110 4754 scope.go:117] "RemoveContainer" containerID="1bb4b1f5def9e89ddfea5b5e0921b99d55f1ceb68868b957d9868c33a0211da3" Oct 05 21:46:20 crc kubenswrapper[4754]: E1005 21:46:20.842301 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 21:46:21 crc kubenswrapper[4754]: I1005 21:46:21.278115 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-g4jzc"] Oct 05 21:46:21 crc kubenswrapper[4754]: W1005 21:46:21.291871 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbe29e094_c72e_4ca0_9f8f_d9234767ed71.slice/crio-0a07fe21b2892d22061e0e9d460775b34adbd1715b2c676bc7edfc725f6240db WatchSource:0}: Error finding container 0a07fe21b2892d22061e0e9d460775b34adbd1715b2c676bc7edfc725f6240db: Status 404 returned error can't find the container with id 0a07fe21b2892d22061e0e9d460775b34adbd1715b2c676bc7edfc725f6240db Oct 05 21:46:21 crc kubenswrapper[4754]: I1005 21:46:21.349614 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g4jzc" event={"ID":"be29e094-c72e-4ca0-9f8f-d9234767ed71","Type":"ContainerStarted","Data":"0a07fe21b2892d22061e0e9d460775b34adbd1715b2c676bc7edfc725f6240db"} Oct 05 21:46:22 crc kubenswrapper[4754]: I1005 21:46:22.358989 4754 generic.go:334] "Generic (PLEG): container finished" podID="be29e094-c72e-4ca0-9f8f-d9234767ed71" containerID="8f645b8ac64d6c5e2f351ac4cf368c9e54ebfdfe12fb9cdc6b32f7b836ac474f" exitCode=0 Oct 05 21:46:22 crc kubenswrapper[4754]: I1005 21:46:22.359046 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g4jzc" event={"ID":"be29e094-c72e-4ca0-9f8f-d9234767ed71","Type":"ContainerDied","Data":"8f645b8ac64d6c5e2f351ac4cf368c9e54ebfdfe12fb9cdc6b32f7b836ac474f"} Oct 05 21:46:24 crc kubenswrapper[4754]: I1005 21:46:24.171134 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-ws6kv"] Oct 05 21:46:24 crc kubenswrapper[4754]: I1005 21:46:24.179544 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ws6kv" Oct 05 21:46:24 crc kubenswrapper[4754]: I1005 21:46:24.187727 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-ws6kv"] Oct 05 21:46:24 crc kubenswrapper[4754]: I1005 21:46:24.230903 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4209a211-0d5f-4a6c-95ef-5f1894bc0f11-utilities\") pod \"redhat-marketplace-ws6kv\" (UID: \"4209a211-0d5f-4a6c-95ef-5f1894bc0f11\") " pod="openshift-marketplace/redhat-marketplace-ws6kv" Oct 05 21:46:24 crc kubenswrapper[4754]: I1005 21:46:24.231212 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4209a211-0d5f-4a6c-95ef-5f1894bc0f11-catalog-content\") pod \"redhat-marketplace-ws6kv\" (UID: \"4209a211-0d5f-4a6c-95ef-5f1894bc0f11\") " pod="openshift-marketplace/redhat-marketplace-ws6kv" Oct 05 21:46:24 crc kubenswrapper[4754]: I1005 21:46:24.231382 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fh9pf\" (UniqueName: \"kubernetes.io/projected/4209a211-0d5f-4a6c-95ef-5f1894bc0f11-kube-api-access-fh9pf\") pod \"redhat-marketplace-ws6kv\" (UID: \"4209a211-0d5f-4a6c-95ef-5f1894bc0f11\") " pod="openshift-marketplace/redhat-marketplace-ws6kv" Oct 05 21:46:24 crc kubenswrapper[4754]: I1005 21:46:24.333617 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4209a211-0d5f-4a6c-95ef-5f1894bc0f11-catalog-content\") pod \"redhat-marketplace-ws6kv\" (UID: \"4209a211-0d5f-4a6c-95ef-5f1894bc0f11\") " pod="openshift-marketplace/redhat-marketplace-ws6kv" Oct 05 21:46:24 crc kubenswrapper[4754]: I1005 21:46:24.333699 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fh9pf\" (UniqueName: \"kubernetes.io/projected/4209a211-0d5f-4a6c-95ef-5f1894bc0f11-kube-api-access-fh9pf\") pod \"redhat-marketplace-ws6kv\" (UID: \"4209a211-0d5f-4a6c-95ef-5f1894bc0f11\") " pod="openshift-marketplace/redhat-marketplace-ws6kv" Oct 05 21:46:24 crc kubenswrapper[4754]: I1005 21:46:24.333750 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4209a211-0d5f-4a6c-95ef-5f1894bc0f11-utilities\") pod \"redhat-marketplace-ws6kv\" (UID: \"4209a211-0d5f-4a6c-95ef-5f1894bc0f11\") " pod="openshift-marketplace/redhat-marketplace-ws6kv" Oct 05 21:46:24 crc kubenswrapper[4754]: I1005 21:46:24.334298 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4209a211-0d5f-4a6c-95ef-5f1894bc0f11-catalog-content\") pod \"redhat-marketplace-ws6kv\" (UID: \"4209a211-0d5f-4a6c-95ef-5f1894bc0f11\") " pod="openshift-marketplace/redhat-marketplace-ws6kv" Oct 05 21:46:24 crc kubenswrapper[4754]: I1005 21:46:24.334413 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4209a211-0d5f-4a6c-95ef-5f1894bc0f11-utilities\") pod \"redhat-marketplace-ws6kv\" (UID: \"4209a211-0d5f-4a6c-95ef-5f1894bc0f11\") " pod="openshift-marketplace/redhat-marketplace-ws6kv" Oct 05 21:46:24 crc kubenswrapper[4754]: I1005 21:46:24.362247 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fh9pf\" (UniqueName: \"kubernetes.io/projected/4209a211-0d5f-4a6c-95ef-5f1894bc0f11-kube-api-access-fh9pf\") pod \"redhat-marketplace-ws6kv\" (UID: \"4209a211-0d5f-4a6c-95ef-5f1894bc0f11\") " pod="openshift-marketplace/redhat-marketplace-ws6kv" Oct 05 21:46:24 crc kubenswrapper[4754]: I1005 21:46:24.382397 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g4jzc" event={"ID":"be29e094-c72e-4ca0-9f8f-d9234767ed71","Type":"ContainerStarted","Data":"e1b641bc614adb094a2e1315ec153e31fa06d4e4e5c570bc89a4a1bb9a9fb38a"} Oct 05 21:46:24 crc kubenswrapper[4754]: I1005 21:46:24.554209 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ws6kv" Oct 05 21:46:24 crc kubenswrapper[4754]: I1005 21:46:24.961627 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-ws6kv"] Oct 05 21:46:25 crc kubenswrapper[4754]: I1005 21:46:25.397727 4754 generic.go:334] "Generic (PLEG): container finished" podID="4209a211-0d5f-4a6c-95ef-5f1894bc0f11" containerID="2ccb427c34d53b49856d812e5d85cde185241d3b516050ac4db330969480f2b1" exitCode=0 Oct 05 21:46:25 crc kubenswrapper[4754]: I1005 21:46:25.400810 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ws6kv" event={"ID":"4209a211-0d5f-4a6c-95ef-5f1894bc0f11","Type":"ContainerDied","Data":"2ccb427c34d53b49856d812e5d85cde185241d3b516050ac4db330969480f2b1"} Oct 05 21:46:25 crc kubenswrapper[4754]: I1005 21:46:25.401266 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ws6kv" event={"ID":"4209a211-0d5f-4a6c-95ef-5f1894bc0f11","Type":"ContainerStarted","Data":"d1254b95ad658668dc4861f4c32bafcda3f45127b9e8575c12589ddc6602693d"} Oct 05 21:46:27 crc kubenswrapper[4754]: I1005 21:46:27.414531 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ws6kv" event={"ID":"4209a211-0d5f-4a6c-95ef-5f1894bc0f11","Type":"ContainerStarted","Data":"91627fc10b845b6c745146fb3554fb90101bae859a29b27222dc271679c9d824"} Oct 05 21:46:27 crc kubenswrapper[4754]: I1005 21:46:27.418802 4754 generic.go:334] "Generic (PLEG): container finished" podID="be29e094-c72e-4ca0-9f8f-d9234767ed71" containerID="e1b641bc614adb094a2e1315ec153e31fa06d4e4e5c570bc89a4a1bb9a9fb38a" exitCode=0 Oct 05 21:46:27 crc kubenswrapper[4754]: I1005 21:46:27.418988 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g4jzc" event={"ID":"be29e094-c72e-4ca0-9f8f-d9234767ed71","Type":"ContainerDied","Data":"e1b641bc614adb094a2e1315ec153e31fa06d4e4e5c570bc89a4a1bb9a9fb38a"} Oct 05 21:46:28 crc kubenswrapper[4754]: I1005 21:46:28.444529 4754 generic.go:334] "Generic (PLEG): container finished" podID="4209a211-0d5f-4a6c-95ef-5f1894bc0f11" containerID="91627fc10b845b6c745146fb3554fb90101bae859a29b27222dc271679c9d824" exitCode=0 Oct 05 21:46:28 crc kubenswrapper[4754]: I1005 21:46:28.444627 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ws6kv" event={"ID":"4209a211-0d5f-4a6c-95ef-5f1894bc0f11","Type":"ContainerDied","Data":"91627fc10b845b6c745146fb3554fb90101bae859a29b27222dc271679c9d824"} Oct 05 21:46:28 crc kubenswrapper[4754]: I1005 21:46:28.463629 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g4jzc" event={"ID":"be29e094-c72e-4ca0-9f8f-d9234767ed71","Type":"ContainerStarted","Data":"31ed03b112336d7268f7fb1b95d954464c9cfcf25b94666b717d4347a7a2b8f2"} Oct 05 21:46:28 crc kubenswrapper[4754]: I1005 21:46:28.499344 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-g4jzc" podStartSLOduration=3.033720782 podStartE2EDuration="8.499321485s" podCreationTimestamp="2025-10-05 21:46:20 +0000 UTC" firstStartedPulling="2025-10-05 21:46:22.36086112 +0000 UTC m=+3106.264979830" lastFinishedPulling="2025-10-05 21:46:27.826461793 +0000 UTC m=+3111.730580533" observedRunningTime="2025-10-05 21:46:28.494114161 +0000 UTC m=+3112.398232871" watchObservedRunningTime="2025-10-05 21:46:28.499321485 +0000 UTC m=+3112.403440195" Oct 05 21:46:29 crc kubenswrapper[4754]: I1005 21:46:29.483217 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ws6kv" event={"ID":"4209a211-0d5f-4a6c-95ef-5f1894bc0f11","Type":"ContainerStarted","Data":"415488d808332c42d13bf27f9ea4bdc4d38404a4388fde6c9b30c226dc9b1161"} Oct 05 21:46:29 crc kubenswrapper[4754]: I1005 21:46:29.514312 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-ws6kv" podStartSLOduration=2.068620348 podStartE2EDuration="5.514290387s" podCreationTimestamp="2025-10-05 21:46:24 +0000 UTC" firstStartedPulling="2025-10-05 21:46:25.405199669 +0000 UTC m=+3109.309318379" lastFinishedPulling="2025-10-05 21:46:28.850869708 +0000 UTC m=+3112.754988418" observedRunningTime="2025-10-05 21:46:29.508312772 +0000 UTC m=+3113.412431512" watchObservedRunningTime="2025-10-05 21:46:29.514290387 +0000 UTC m=+3113.418409117" Oct 05 21:46:30 crc kubenswrapper[4754]: I1005 21:46:30.757079 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-g4jzc" Oct 05 21:46:30 crc kubenswrapper[4754]: I1005 21:46:30.757416 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-g4jzc" Oct 05 21:46:30 crc kubenswrapper[4754]: I1005 21:46:30.834158 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-g4jzc" Oct 05 21:46:31 crc kubenswrapper[4754]: I1005 21:46:31.838683 4754 scope.go:117] "RemoveContainer" containerID="1bb4b1f5def9e89ddfea5b5e0921b99d55f1ceb68868b957d9868c33a0211da3" Oct 05 21:46:31 crc kubenswrapper[4754]: E1005 21:46:31.839444 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 21:46:34 crc kubenswrapper[4754]: I1005 21:46:34.562616 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-ws6kv" Oct 05 21:46:34 crc kubenswrapper[4754]: I1005 21:46:34.564030 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-ws6kv" Oct 05 21:46:34 crc kubenswrapper[4754]: I1005 21:46:34.631955 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-ws6kv" Oct 05 21:46:35 crc kubenswrapper[4754]: I1005 21:46:35.656230 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-ws6kv" Oct 05 21:46:35 crc kubenswrapper[4754]: I1005 21:46:35.952828 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-ws6kv"] Oct 05 21:46:37 crc kubenswrapper[4754]: I1005 21:46:37.588416 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-ws6kv" podUID="4209a211-0d5f-4a6c-95ef-5f1894bc0f11" containerName="registry-server" containerID="cri-o://415488d808332c42d13bf27f9ea4bdc4d38404a4388fde6c9b30c226dc9b1161" gracePeriod=2 Oct 05 21:46:37 crc kubenswrapper[4754]: I1005 21:46:37.744118 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest"] Oct 05 21:46:37 crc kubenswrapper[4754]: I1005 21:46:37.749458 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Oct 05 21:46:37 crc kubenswrapper[4754]: I1005 21:46:37.755413 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-dx5w8" Oct 05 21:46:37 crc kubenswrapper[4754]: I1005 21:46:37.756388 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Oct 05 21:46:37 crc kubenswrapper[4754]: I1005 21:46:37.756532 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Oct 05 21:46:37 crc kubenswrapper[4754]: I1005 21:46:37.758599 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Oct 05 21:46:37 crc kubenswrapper[4754]: I1005 21:46:37.778774 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Oct 05 21:46:37 crc kubenswrapper[4754]: I1005 21:46:37.784397 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/5922ff4d-2ec1-4156-acc7-8e6427819b3e-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"5922ff4d-2ec1-4156-acc7-8e6427819b3e\") " pod="openstack/tempest-tests-tempest" Oct 05 21:46:37 crc kubenswrapper[4754]: I1005 21:46:37.784785 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5922ff4d-2ec1-4156-acc7-8e6427819b3e-config-data\") pod \"tempest-tests-tempest\" (UID: \"5922ff4d-2ec1-4156-acc7-8e6427819b3e\") " pod="openstack/tempest-tests-tempest" Oct 05 21:46:37 crc kubenswrapper[4754]: I1005 21:46:37.785212 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/5922ff4d-2ec1-4156-acc7-8e6427819b3e-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"5922ff4d-2ec1-4156-acc7-8e6427819b3e\") " pod="openstack/tempest-tests-tempest" Oct 05 21:46:37 crc kubenswrapper[4754]: I1005 21:46:37.888587 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/5922ff4d-2ec1-4156-acc7-8e6427819b3e-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"5922ff4d-2ec1-4156-acc7-8e6427819b3e\") " pod="openstack/tempest-tests-tempest" Oct 05 21:46:37 crc kubenswrapper[4754]: I1005 21:46:37.888661 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"tempest-tests-tempest\" (UID: \"5922ff4d-2ec1-4156-acc7-8e6427819b3e\") " pod="openstack/tempest-tests-tempest" Oct 05 21:46:37 crc kubenswrapper[4754]: I1005 21:46:37.888714 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/5922ff4d-2ec1-4156-acc7-8e6427819b3e-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"5922ff4d-2ec1-4156-acc7-8e6427819b3e\") " pod="openstack/tempest-tests-tempest" Oct 05 21:46:37 crc kubenswrapper[4754]: I1005 21:46:37.888820 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/5922ff4d-2ec1-4156-acc7-8e6427819b3e-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"5922ff4d-2ec1-4156-acc7-8e6427819b3e\") " pod="openstack/tempest-tests-tempest" Oct 05 21:46:37 crc kubenswrapper[4754]: I1005 21:46:37.888864 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/5922ff4d-2ec1-4156-acc7-8e6427819b3e-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"5922ff4d-2ec1-4156-acc7-8e6427819b3e\") " pod="openstack/tempest-tests-tempest" Oct 05 21:46:37 crc kubenswrapper[4754]: I1005 21:46:37.889105 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5922ff4d-2ec1-4156-acc7-8e6427819b3e-config-data\") pod \"tempest-tests-tempest\" (UID: \"5922ff4d-2ec1-4156-acc7-8e6427819b3e\") " pod="openstack/tempest-tests-tempest" Oct 05 21:46:37 crc kubenswrapper[4754]: I1005 21:46:37.889287 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x5j8w\" (UniqueName: \"kubernetes.io/projected/5922ff4d-2ec1-4156-acc7-8e6427819b3e-kube-api-access-x5j8w\") pod \"tempest-tests-tempest\" (UID: \"5922ff4d-2ec1-4156-acc7-8e6427819b3e\") " pod="openstack/tempest-tests-tempest" Oct 05 21:46:37 crc kubenswrapper[4754]: I1005 21:46:37.889635 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5922ff4d-2ec1-4156-acc7-8e6427819b3e-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"5922ff4d-2ec1-4156-acc7-8e6427819b3e\") " pod="openstack/tempest-tests-tempest" Oct 05 21:46:37 crc kubenswrapper[4754]: I1005 21:46:37.889691 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/5922ff4d-2ec1-4156-acc7-8e6427819b3e-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"5922ff4d-2ec1-4156-acc7-8e6427819b3e\") " pod="openstack/tempest-tests-tempest" Oct 05 21:46:37 crc kubenswrapper[4754]: I1005 21:46:37.890158 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/5922ff4d-2ec1-4156-acc7-8e6427819b3e-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"5922ff4d-2ec1-4156-acc7-8e6427819b3e\") " pod="openstack/tempest-tests-tempest" Oct 05 21:46:37 crc kubenswrapper[4754]: I1005 21:46:37.891673 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5922ff4d-2ec1-4156-acc7-8e6427819b3e-config-data\") pod \"tempest-tests-tempest\" (UID: \"5922ff4d-2ec1-4156-acc7-8e6427819b3e\") " pod="openstack/tempest-tests-tempest" Oct 05 21:46:37 crc kubenswrapper[4754]: I1005 21:46:37.900654 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/5922ff4d-2ec1-4156-acc7-8e6427819b3e-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"5922ff4d-2ec1-4156-acc7-8e6427819b3e\") " pod="openstack/tempest-tests-tempest" Oct 05 21:46:37 crc kubenswrapper[4754]: I1005 21:46:37.991871 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"tempest-tests-tempest\" (UID: \"5922ff4d-2ec1-4156-acc7-8e6427819b3e\") " pod="openstack/tempest-tests-tempest" Oct 05 21:46:37 crc kubenswrapper[4754]: I1005 21:46:37.991937 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/5922ff4d-2ec1-4156-acc7-8e6427819b3e-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"5922ff4d-2ec1-4156-acc7-8e6427819b3e\") " pod="openstack/tempest-tests-tempest" Oct 05 21:46:37 crc kubenswrapper[4754]: I1005 21:46:37.992024 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/5922ff4d-2ec1-4156-acc7-8e6427819b3e-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"5922ff4d-2ec1-4156-acc7-8e6427819b3e\") " pod="openstack/tempest-tests-tempest" Oct 05 21:46:37 crc kubenswrapper[4754]: I1005 21:46:37.992086 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x5j8w\" (UniqueName: \"kubernetes.io/projected/5922ff4d-2ec1-4156-acc7-8e6427819b3e-kube-api-access-x5j8w\") pod \"tempest-tests-tempest\" (UID: \"5922ff4d-2ec1-4156-acc7-8e6427819b3e\") " pod="openstack/tempest-tests-tempest" Oct 05 21:46:37 crc kubenswrapper[4754]: I1005 21:46:37.992149 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5922ff4d-2ec1-4156-acc7-8e6427819b3e-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"5922ff4d-2ec1-4156-acc7-8e6427819b3e\") " pod="openstack/tempest-tests-tempest" Oct 05 21:46:37 crc kubenswrapper[4754]: I1005 21:46:37.992189 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/5922ff4d-2ec1-4156-acc7-8e6427819b3e-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"5922ff4d-2ec1-4156-acc7-8e6427819b3e\") " pod="openstack/tempest-tests-tempest" Oct 05 21:46:37 crc kubenswrapper[4754]: I1005 21:46:37.992744 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/5922ff4d-2ec1-4156-acc7-8e6427819b3e-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"5922ff4d-2ec1-4156-acc7-8e6427819b3e\") " pod="openstack/tempest-tests-tempest" Oct 05 21:46:37 crc kubenswrapper[4754]: I1005 21:46:37.993200 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/5922ff4d-2ec1-4156-acc7-8e6427819b3e-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"5922ff4d-2ec1-4156-acc7-8e6427819b3e\") " pod="openstack/tempest-tests-tempest" Oct 05 21:46:38 crc kubenswrapper[4754]: I1005 21:46:38.000657 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/5922ff4d-2ec1-4156-acc7-8e6427819b3e-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"5922ff4d-2ec1-4156-acc7-8e6427819b3e\") " pod="openstack/tempest-tests-tempest" Oct 05 21:46:38 crc kubenswrapper[4754]: I1005 21:46:38.004097 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5922ff4d-2ec1-4156-acc7-8e6427819b3e-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"5922ff4d-2ec1-4156-acc7-8e6427819b3e\") " pod="openstack/tempest-tests-tempest" Oct 05 21:46:38 crc kubenswrapper[4754]: I1005 21:46:38.014459 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x5j8w\" (UniqueName: \"kubernetes.io/projected/5922ff4d-2ec1-4156-acc7-8e6427819b3e-kube-api-access-x5j8w\") pod \"tempest-tests-tempest\" (UID: \"5922ff4d-2ec1-4156-acc7-8e6427819b3e\") " pod="openstack/tempest-tests-tempest" Oct 05 21:46:38 crc kubenswrapper[4754]: I1005 21:46:38.022105 4754 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"tempest-tests-tempest\" (UID: \"5922ff4d-2ec1-4156-acc7-8e6427819b3e\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/tempest-tests-tempest" Oct 05 21:46:38 crc kubenswrapper[4754]: I1005 21:46:38.065661 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"tempest-tests-tempest\" (UID: \"5922ff4d-2ec1-4156-acc7-8e6427819b3e\") " pod="openstack/tempest-tests-tempest" Oct 05 21:46:38 crc kubenswrapper[4754]: I1005 21:46:38.179415 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Oct 05 21:46:38 crc kubenswrapper[4754]: I1005 21:46:38.500568 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Oct 05 21:46:38 crc kubenswrapper[4754]: I1005 21:46:38.602342 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"5922ff4d-2ec1-4156-acc7-8e6427819b3e","Type":"ContainerStarted","Data":"e2712f6081e90acaba2b35a7652cc92d1a43dea8c02a3267d9c8e5c3f5438972"} Oct 05 21:46:38 crc kubenswrapper[4754]: I1005 21:46:38.607443 4754 generic.go:334] "Generic (PLEG): container finished" podID="4209a211-0d5f-4a6c-95ef-5f1894bc0f11" containerID="415488d808332c42d13bf27f9ea4bdc4d38404a4388fde6c9b30c226dc9b1161" exitCode=0 Oct 05 21:46:38 crc kubenswrapper[4754]: I1005 21:46:38.607535 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ws6kv" event={"ID":"4209a211-0d5f-4a6c-95ef-5f1894bc0f11","Type":"ContainerDied","Data":"415488d808332c42d13bf27f9ea4bdc4d38404a4388fde6c9b30c226dc9b1161"} Oct 05 21:46:39 crc kubenswrapper[4754]: I1005 21:46:39.030161 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ws6kv" Oct 05 21:46:39 crc kubenswrapper[4754]: I1005 21:46:39.120626 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4209a211-0d5f-4a6c-95ef-5f1894bc0f11-catalog-content\") pod \"4209a211-0d5f-4a6c-95ef-5f1894bc0f11\" (UID: \"4209a211-0d5f-4a6c-95ef-5f1894bc0f11\") " Oct 05 21:46:39 crc kubenswrapper[4754]: I1005 21:46:39.121363 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fh9pf\" (UniqueName: \"kubernetes.io/projected/4209a211-0d5f-4a6c-95ef-5f1894bc0f11-kube-api-access-fh9pf\") pod \"4209a211-0d5f-4a6c-95ef-5f1894bc0f11\" (UID: \"4209a211-0d5f-4a6c-95ef-5f1894bc0f11\") " Oct 05 21:46:39 crc kubenswrapper[4754]: I1005 21:46:39.121897 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4209a211-0d5f-4a6c-95ef-5f1894bc0f11-utilities\") pod \"4209a211-0d5f-4a6c-95ef-5f1894bc0f11\" (UID: \"4209a211-0d5f-4a6c-95ef-5f1894bc0f11\") " Oct 05 21:46:39 crc kubenswrapper[4754]: I1005 21:46:39.122722 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4209a211-0d5f-4a6c-95ef-5f1894bc0f11-utilities" (OuterVolumeSpecName: "utilities") pod "4209a211-0d5f-4a6c-95ef-5f1894bc0f11" (UID: "4209a211-0d5f-4a6c-95ef-5f1894bc0f11"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 21:46:39 crc kubenswrapper[4754]: I1005 21:46:39.129887 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4209a211-0d5f-4a6c-95ef-5f1894bc0f11-kube-api-access-fh9pf" (OuterVolumeSpecName: "kube-api-access-fh9pf") pod "4209a211-0d5f-4a6c-95ef-5f1894bc0f11" (UID: "4209a211-0d5f-4a6c-95ef-5f1894bc0f11"). InnerVolumeSpecName "kube-api-access-fh9pf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:46:39 crc kubenswrapper[4754]: I1005 21:46:39.142908 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4209a211-0d5f-4a6c-95ef-5f1894bc0f11-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4209a211-0d5f-4a6c-95ef-5f1894bc0f11" (UID: "4209a211-0d5f-4a6c-95ef-5f1894bc0f11"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 21:46:39 crc kubenswrapper[4754]: I1005 21:46:39.225510 4754 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4209a211-0d5f-4a6c-95ef-5f1894bc0f11-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 21:46:39 crc kubenswrapper[4754]: I1005 21:46:39.225554 4754 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4209a211-0d5f-4a6c-95ef-5f1894bc0f11-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 21:46:39 crc kubenswrapper[4754]: I1005 21:46:39.225570 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fh9pf\" (UniqueName: \"kubernetes.io/projected/4209a211-0d5f-4a6c-95ef-5f1894bc0f11-kube-api-access-fh9pf\") on node \"crc\" DevicePath \"\"" Oct 05 21:46:39 crc kubenswrapper[4754]: I1005 21:46:39.619945 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ws6kv" event={"ID":"4209a211-0d5f-4a6c-95ef-5f1894bc0f11","Type":"ContainerDied","Data":"d1254b95ad658668dc4861f4c32bafcda3f45127b9e8575c12589ddc6602693d"} Oct 05 21:46:39 crc kubenswrapper[4754]: I1005 21:46:39.620010 4754 scope.go:117] "RemoveContainer" containerID="415488d808332c42d13bf27f9ea4bdc4d38404a4388fde6c9b30c226dc9b1161" Oct 05 21:46:39 crc kubenswrapper[4754]: I1005 21:46:39.620181 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ws6kv" Oct 05 21:46:39 crc kubenswrapper[4754]: I1005 21:46:39.658828 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-ws6kv"] Oct 05 21:46:39 crc kubenswrapper[4754]: I1005 21:46:39.666854 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-ws6kv"] Oct 05 21:46:39 crc kubenswrapper[4754]: I1005 21:46:39.672913 4754 scope.go:117] "RemoveContainer" containerID="91627fc10b845b6c745146fb3554fb90101bae859a29b27222dc271679c9d824" Oct 05 21:46:39 crc kubenswrapper[4754]: I1005 21:46:39.718247 4754 scope.go:117] "RemoveContainer" containerID="2ccb427c34d53b49856d812e5d85cde185241d3b516050ac4db330969480f2b1" Oct 05 21:46:40 crc kubenswrapper[4754]: I1005 21:46:40.816619 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-g4jzc" Oct 05 21:46:40 crc kubenswrapper[4754]: I1005 21:46:40.851552 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4209a211-0d5f-4a6c-95ef-5f1894bc0f11" path="/var/lib/kubelet/pods/4209a211-0d5f-4a6c-95ef-5f1894bc0f11/volumes" Oct 05 21:46:41 crc kubenswrapper[4754]: I1005 21:46:41.345616 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-g4jzc"] Oct 05 21:46:41 crc kubenswrapper[4754]: I1005 21:46:41.643734 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-g4jzc" podUID="be29e094-c72e-4ca0-9f8f-d9234767ed71" containerName="registry-server" containerID="cri-o://31ed03b112336d7268f7fb1b95d954464c9cfcf25b94666b717d4347a7a2b8f2" gracePeriod=2 Oct 05 21:46:42 crc kubenswrapper[4754]: I1005 21:46:42.837435 4754 scope.go:117] "RemoveContainer" containerID="1bb4b1f5def9e89ddfea5b5e0921b99d55f1ceb68868b957d9868c33a0211da3" Oct 05 21:46:42 crc kubenswrapper[4754]: E1005 21:46:42.838140 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 21:46:43 crc kubenswrapper[4754]: I1005 21:46:43.661672 4754 generic.go:334] "Generic (PLEG): container finished" podID="be29e094-c72e-4ca0-9f8f-d9234767ed71" containerID="31ed03b112336d7268f7fb1b95d954464c9cfcf25b94666b717d4347a7a2b8f2" exitCode=0 Oct 05 21:46:43 crc kubenswrapper[4754]: I1005 21:46:43.661728 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g4jzc" event={"ID":"be29e094-c72e-4ca0-9f8f-d9234767ed71","Type":"ContainerDied","Data":"31ed03b112336d7268f7fb1b95d954464c9cfcf25b94666b717d4347a7a2b8f2"} Oct 05 21:46:50 crc kubenswrapper[4754]: E1005 21:46:50.757844 4754 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 31ed03b112336d7268f7fb1b95d954464c9cfcf25b94666b717d4347a7a2b8f2 is running failed: container process not found" containerID="31ed03b112336d7268f7fb1b95d954464c9cfcf25b94666b717d4347a7a2b8f2" cmd=["grpc_health_probe","-addr=:50051"] Oct 05 21:46:50 crc kubenswrapper[4754]: E1005 21:46:50.760251 4754 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 31ed03b112336d7268f7fb1b95d954464c9cfcf25b94666b717d4347a7a2b8f2 is running failed: container process not found" containerID="31ed03b112336d7268f7fb1b95d954464c9cfcf25b94666b717d4347a7a2b8f2" cmd=["grpc_health_probe","-addr=:50051"] Oct 05 21:46:50 crc kubenswrapper[4754]: E1005 21:46:50.760571 4754 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 31ed03b112336d7268f7fb1b95d954464c9cfcf25b94666b717d4347a7a2b8f2 is running failed: container process not found" containerID="31ed03b112336d7268f7fb1b95d954464c9cfcf25b94666b717d4347a7a2b8f2" cmd=["grpc_health_probe","-addr=:50051"] Oct 05 21:46:50 crc kubenswrapper[4754]: E1005 21:46:50.760600 4754 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 31ed03b112336d7268f7fb1b95d954464c9cfcf25b94666b717d4347a7a2b8f2 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/community-operators-g4jzc" podUID="be29e094-c72e-4ca0-9f8f-d9234767ed71" containerName="registry-server" Oct 05 21:46:56 crc kubenswrapper[4754]: I1005 21:46:56.851087 4754 scope.go:117] "RemoveContainer" containerID="1bb4b1f5def9e89ddfea5b5e0921b99d55f1ceb68868b957d9868c33a0211da3" Oct 05 21:46:56 crc kubenswrapper[4754]: E1005 21:46:56.853255 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 21:47:00 crc kubenswrapper[4754]: E1005 21:47:00.757724 4754 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 31ed03b112336d7268f7fb1b95d954464c9cfcf25b94666b717d4347a7a2b8f2 is running failed: container process not found" containerID="31ed03b112336d7268f7fb1b95d954464c9cfcf25b94666b717d4347a7a2b8f2" cmd=["grpc_health_probe","-addr=:50051"] Oct 05 21:47:00 crc kubenswrapper[4754]: E1005 21:47:00.758731 4754 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 31ed03b112336d7268f7fb1b95d954464c9cfcf25b94666b717d4347a7a2b8f2 is running failed: container process not found" containerID="31ed03b112336d7268f7fb1b95d954464c9cfcf25b94666b717d4347a7a2b8f2" cmd=["grpc_health_probe","-addr=:50051"] Oct 05 21:47:00 crc kubenswrapper[4754]: E1005 21:47:00.759227 4754 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 31ed03b112336d7268f7fb1b95d954464c9cfcf25b94666b717d4347a7a2b8f2 is running failed: container process not found" containerID="31ed03b112336d7268f7fb1b95d954464c9cfcf25b94666b717d4347a7a2b8f2" cmd=["grpc_health_probe","-addr=:50051"] Oct 05 21:47:00 crc kubenswrapper[4754]: E1005 21:47:00.759284 4754 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 31ed03b112336d7268f7fb1b95d954464c9cfcf25b94666b717d4347a7a2b8f2 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/community-operators-g4jzc" podUID="be29e094-c72e-4ca0-9f8f-d9234767ed71" containerName="registry-server" Oct 05 21:47:10 crc kubenswrapper[4754]: E1005 21:47:10.758354 4754 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 31ed03b112336d7268f7fb1b95d954464c9cfcf25b94666b717d4347a7a2b8f2 is running failed: container process not found" containerID="31ed03b112336d7268f7fb1b95d954464c9cfcf25b94666b717d4347a7a2b8f2" cmd=["grpc_health_probe","-addr=:50051"] Oct 05 21:47:10 crc kubenswrapper[4754]: E1005 21:47:10.759720 4754 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 31ed03b112336d7268f7fb1b95d954464c9cfcf25b94666b717d4347a7a2b8f2 is running failed: container process not found" containerID="31ed03b112336d7268f7fb1b95d954464c9cfcf25b94666b717d4347a7a2b8f2" cmd=["grpc_health_probe","-addr=:50051"] Oct 05 21:47:10 crc kubenswrapper[4754]: E1005 21:47:10.760146 4754 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 31ed03b112336d7268f7fb1b95d954464c9cfcf25b94666b717d4347a7a2b8f2 is running failed: container process not found" containerID="31ed03b112336d7268f7fb1b95d954464c9cfcf25b94666b717d4347a7a2b8f2" cmd=["grpc_health_probe","-addr=:50051"] Oct 05 21:47:10 crc kubenswrapper[4754]: E1005 21:47:10.760199 4754 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 31ed03b112336d7268f7fb1b95d954464c9cfcf25b94666b717d4347a7a2b8f2 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/community-operators-g4jzc" podUID="be29e094-c72e-4ca0-9f8f-d9234767ed71" containerName="registry-server" Oct 05 21:47:11 crc kubenswrapper[4754]: I1005 21:47:11.838675 4754 scope.go:117] "RemoveContainer" containerID="1bb4b1f5def9e89ddfea5b5e0921b99d55f1ceb68868b957d9868c33a0211da3" Oct 05 21:47:11 crc kubenswrapper[4754]: E1005 21:47:11.839528 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 21:47:20 crc kubenswrapper[4754]: E1005 21:47:20.758014 4754 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 31ed03b112336d7268f7fb1b95d954464c9cfcf25b94666b717d4347a7a2b8f2 is running failed: container process not found" containerID="31ed03b112336d7268f7fb1b95d954464c9cfcf25b94666b717d4347a7a2b8f2" cmd=["grpc_health_probe","-addr=:50051"] Oct 05 21:47:20 crc kubenswrapper[4754]: E1005 21:47:20.759849 4754 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 31ed03b112336d7268f7fb1b95d954464c9cfcf25b94666b717d4347a7a2b8f2 is running failed: container process not found" containerID="31ed03b112336d7268f7fb1b95d954464c9cfcf25b94666b717d4347a7a2b8f2" cmd=["grpc_health_probe","-addr=:50051"] Oct 05 21:47:20 crc kubenswrapper[4754]: E1005 21:47:20.760381 4754 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 31ed03b112336d7268f7fb1b95d954464c9cfcf25b94666b717d4347a7a2b8f2 is running failed: container process not found" containerID="31ed03b112336d7268f7fb1b95d954464c9cfcf25b94666b717d4347a7a2b8f2" cmd=["grpc_health_probe","-addr=:50051"] Oct 05 21:47:20 crc kubenswrapper[4754]: E1005 21:47:20.760448 4754 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 31ed03b112336d7268f7fb1b95d954464c9cfcf25b94666b717d4347a7a2b8f2 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/community-operators-g4jzc" podUID="be29e094-c72e-4ca0-9f8f-d9234767ed71" containerName="registry-server" Oct 05 21:47:26 crc kubenswrapper[4754]: I1005 21:47:26.850543 4754 scope.go:117] "RemoveContainer" containerID="1bb4b1f5def9e89ddfea5b5e0921b99d55f1ceb68868b957d9868c33a0211da3" Oct 05 21:47:26 crc kubenswrapper[4754]: E1005 21:47:26.851765 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 21:47:29 crc kubenswrapper[4754]: I1005 21:47:29.277657 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g4jzc" event={"ID":"be29e094-c72e-4ca0-9f8f-d9234767ed71","Type":"ContainerDied","Data":"0a07fe21b2892d22061e0e9d460775b34adbd1715b2c676bc7edfc725f6240db"} Oct 05 21:47:29 crc kubenswrapper[4754]: I1005 21:47:29.278015 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0a07fe21b2892d22061e0e9d460775b34adbd1715b2c676bc7edfc725f6240db" Oct 05 21:47:29 crc kubenswrapper[4754]: E1005 21:47:29.334300 4754 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified" Oct 05 21:47:29 crc kubenswrapper[4754]: E1005 21:47:29.341841 4754 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tempest-tests-tempest-tests-runner,Image:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/tempest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/test_operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/tempest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/var/lib/tempest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/var/lib/tempest/id_ecdsa,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-x5j8w,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42480,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42480,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-custom-data-s0,},Optional:nil,},SecretRef:nil,},EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-env-vars-s0,},Optional:nil,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tempest-tests-tempest_openstack(5922ff4d-2ec1-4156-acc7-8e6427819b3e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 05 21:47:29 crc kubenswrapper[4754]: E1005 21:47:29.343024 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/tempest-tests-tempest" podUID="5922ff4d-2ec1-4156-acc7-8e6427819b3e" Oct 05 21:47:29 crc kubenswrapper[4754]: I1005 21:47:29.386051 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-g4jzc" Oct 05 21:47:29 crc kubenswrapper[4754]: I1005 21:47:29.491872 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rrd2j\" (UniqueName: \"kubernetes.io/projected/be29e094-c72e-4ca0-9f8f-d9234767ed71-kube-api-access-rrd2j\") pod \"be29e094-c72e-4ca0-9f8f-d9234767ed71\" (UID: \"be29e094-c72e-4ca0-9f8f-d9234767ed71\") " Oct 05 21:47:29 crc kubenswrapper[4754]: I1005 21:47:29.492004 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/be29e094-c72e-4ca0-9f8f-d9234767ed71-utilities\") pod \"be29e094-c72e-4ca0-9f8f-d9234767ed71\" (UID: \"be29e094-c72e-4ca0-9f8f-d9234767ed71\") " Oct 05 21:47:29 crc kubenswrapper[4754]: I1005 21:47:29.492126 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/be29e094-c72e-4ca0-9f8f-d9234767ed71-catalog-content\") pod \"be29e094-c72e-4ca0-9f8f-d9234767ed71\" (UID: \"be29e094-c72e-4ca0-9f8f-d9234767ed71\") " Oct 05 21:47:29 crc kubenswrapper[4754]: I1005 21:47:29.492791 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/be29e094-c72e-4ca0-9f8f-d9234767ed71-utilities" (OuterVolumeSpecName: "utilities") pod "be29e094-c72e-4ca0-9f8f-d9234767ed71" (UID: "be29e094-c72e-4ca0-9f8f-d9234767ed71"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 21:47:29 crc kubenswrapper[4754]: I1005 21:47:29.493186 4754 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/be29e094-c72e-4ca0-9f8f-d9234767ed71-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 21:47:29 crc kubenswrapper[4754]: I1005 21:47:29.514769 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/be29e094-c72e-4ca0-9f8f-d9234767ed71-kube-api-access-rrd2j" (OuterVolumeSpecName: "kube-api-access-rrd2j") pod "be29e094-c72e-4ca0-9f8f-d9234767ed71" (UID: "be29e094-c72e-4ca0-9f8f-d9234767ed71"). InnerVolumeSpecName "kube-api-access-rrd2j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:47:29 crc kubenswrapper[4754]: I1005 21:47:29.530379 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/be29e094-c72e-4ca0-9f8f-d9234767ed71-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "be29e094-c72e-4ca0-9f8f-d9234767ed71" (UID: "be29e094-c72e-4ca0-9f8f-d9234767ed71"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 21:47:29 crc kubenswrapper[4754]: I1005 21:47:29.596478 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rrd2j\" (UniqueName: \"kubernetes.io/projected/be29e094-c72e-4ca0-9f8f-d9234767ed71-kube-api-access-rrd2j\") on node \"crc\" DevicePath \"\"" Oct 05 21:47:29 crc kubenswrapper[4754]: I1005 21:47:29.596944 4754 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/be29e094-c72e-4ca0-9f8f-d9234767ed71-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 21:47:30 crc kubenswrapper[4754]: I1005 21:47:30.288728 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-g4jzc" Oct 05 21:47:30 crc kubenswrapper[4754]: E1005 21:47:30.293074 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified\\\"\"" pod="openstack/tempest-tests-tempest" podUID="5922ff4d-2ec1-4156-acc7-8e6427819b3e" Oct 05 21:47:30 crc kubenswrapper[4754]: I1005 21:47:30.366969 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-g4jzc"] Oct 05 21:47:30 crc kubenswrapper[4754]: I1005 21:47:30.379550 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-g4jzc"] Oct 05 21:47:30 crc kubenswrapper[4754]: I1005 21:47:30.861343 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="be29e094-c72e-4ca0-9f8f-d9234767ed71" path="/var/lib/kubelet/pods/be29e094-c72e-4ca0-9f8f-d9234767ed71/volumes" Oct 05 21:47:40 crc kubenswrapper[4754]: I1005 21:47:40.838781 4754 scope.go:117] "RemoveContainer" containerID="1bb4b1f5def9e89ddfea5b5e0921b99d55f1ceb68868b957d9868c33a0211da3" Oct 05 21:47:40 crc kubenswrapper[4754]: E1005 21:47:40.840048 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 21:47:46 crc kubenswrapper[4754]: I1005 21:47:46.036681 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Oct 05 21:47:49 crc kubenswrapper[4754]: I1005 21:47:49.521304 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"5922ff4d-2ec1-4156-acc7-8e6427819b3e","Type":"ContainerStarted","Data":"a7fc76a07a2380639510193b4c0f81120c5ee9beae1678cdbb5acac37cb15675"} Oct 05 21:47:49 crc kubenswrapper[4754]: I1005 21:47:49.552356 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest" podStartSLOduration=6.027068519 podStartE2EDuration="1m13.552325672s" podCreationTimestamp="2025-10-05 21:46:36 +0000 UTC" firstStartedPulling="2025-10-05 21:46:38.507024096 +0000 UTC m=+3122.411142806" lastFinishedPulling="2025-10-05 21:47:46.032281209 +0000 UTC m=+3189.936399959" observedRunningTime="2025-10-05 21:47:49.543964026 +0000 UTC m=+3193.448082776" watchObservedRunningTime="2025-10-05 21:47:49.552325672 +0000 UTC m=+3193.456444422" Oct 05 21:47:53 crc kubenswrapper[4754]: I1005 21:47:53.846029 4754 scope.go:117] "RemoveContainer" containerID="1bb4b1f5def9e89ddfea5b5e0921b99d55f1ceb68868b957d9868c33a0211da3" Oct 05 21:47:53 crc kubenswrapper[4754]: E1005 21:47:53.846886 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 21:48:06 crc kubenswrapper[4754]: I1005 21:48:06.843749 4754 scope.go:117] "RemoveContainer" containerID="1bb4b1f5def9e89ddfea5b5e0921b99d55f1ceb68868b957d9868c33a0211da3" Oct 05 21:48:06 crc kubenswrapper[4754]: E1005 21:48:06.844695 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 21:48:19 crc kubenswrapper[4754]: I1005 21:48:19.838320 4754 scope.go:117] "RemoveContainer" containerID="1bb4b1f5def9e89ddfea5b5e0921b99d55f1ceb68868b957d9868c33a0211da3" Oct 05 21:48:19 crc kubenswrapper[4754]: E1005 21:48:19.839291 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 21:48:34 crc kubenswrapper[4754]: I1005 21:48:34.838713 4754 scope.go:117] "RemoveContainer" containerID="1bb4b1f5def9e89ddfea5b5e0921b99d55f1ceb68868b957d9868c33a0211da3" Oct 05 21:48:34 crc kubenswrapper[4754]: E1005 21:48:34.841285 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 21:48:46 crc kubenswrapper[4754]: I1005 21:48:46.844345 4754 scope.go:117] "RemoveContainer" containerID="1bb4b1f5def9e89ddfea5b5e0921b99d55f1ceb68868b957d9868c33a0211da3" Oct 05 21:48:47 crc kubenswrapper[4754]: I1005 21:48:47.171722 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" event={"ID":"be95a413-6a59-45b4-84b7-b43cae694a26","Type":"ContainerStarted","Data":"8db5d0c4b86f257e1a334bb36f42cc5226cb59650fc43cdcfb1aaccb42c9a511"} Oct 05 21:51:05 crc kubenswrapper[4754]: I1005 21:51:05.244891 4754 patch_prober.go:28] interesting pod/machine-config-daemon-b2h9k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 21:51:05 crc kubenswrapper[4754]: I1005 21:51:05.247016 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 21:51:35 crc kubenswrapper[4754]: I1005 21:51:35.244689 4754 patch_prober.go:28] interesting pod/machine-config-daemon-b2h9k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 21:51:35 crc kubenswrapper[4754]: I1005 21:51:35.245147 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 21:52:05 crc kubenswrapper[4754]: I1005 21:52:05.244874 4754 patch_prober.go:28] interesting pod/machine-config-daemon-b2h9k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 21:52:05 crc kubenswrapper[4754]: I1005 21:52:05.245633 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 21:52:05 crc kubenswrapper[4754]: I1005 21:52:05.245717 4754 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" Oct 05 21:52:05 crc kubenswrapper[4754]: I1005 21:52:05.246974 4754 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8db5d0c4b86f257e1a334bb36f42cc5226cb59650fc43cdcfb1aaccb42c9a511"} pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 05 21:52:05 crc kubenswrapper[4754]: I1005 21:52:05.247087 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" containerName="machine-config-daemon" containerID="cri-o://8db5d0c4b86f257e1a334bb36f42cc5226cb59650fc43cdcfb1aaccb42c9a511" gracePeriod=600 Oct 05 21:52:06 crc kubenswrapper[4754]: I1005 21:52:06.093208 4754 generic.go:334] "Generic (PLEG): container finished" podID="be95a413-6a59-45b4-84b7-b43cae694a26" containerID="8db5d0c4b86f257e1a334bb36f42cc5226cb59650fc43cdcfb1aaccb42c9a511" exitCode=0 Oct 05 21:52:06 crc kubenswrapper[4754]: I1005 21:52:06.093268 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" event={"ID":"be95a413-6a59-45b4-84b7-b43cae694a26","Type":"ContainerDied","Data":"8db5d0c4b86f257e1a334bb36f42cc5226cb59650fc43cdcfb1aaccb42c9a511"} Oct 05 21:52:06 crc kubenswrapper[4754]: I1005 21:52:06.093674 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" event={"ID":"be95a413-6a59-45b4-84b7-b43cae694a26","Type":"ContainerStarted","Data":"f21169793903f3552e21e22d202f837dbee5ee4d48ced3d4666fe154c7e5b95c"} Oct 05 21:52:06 crc kubenswrapper[4754]: I1005 21:52:06.093700 4754 scope.go:117] "RemoveContainer" containerID="1bb4b1f5def9e89ddfea5b5e0921b99d55f1ceb68868b957d9868c33a0211da3" Oct 05 21:52:29 crc kubenswrapper[4754]: I1005 21:52:29.339010 4754 scope.go:117] "RemoveContainer" containerID="8f645b8ac64d6c5e2f351ac4cf368c9e54ebfdfe12fb9cdc6b32f7b836ac474f" Oct 05 21:52:29 crc kubenswrapper[4754]: I1005 21:52:29.376598 4754 scope.go:117] "RemoveContainer" containerID="31ed03b112336d7268f7fb1b95d954464c9cfcf25b94666b717d4347a7a2b8f2" Oct 05 21:52:29 crc kubenswrapper[4754]: I1005 21:52:29.451922 4754 scope.go:117] "RemoveContainer" containerID="e1b641bc614adb094a2e1315ec153e31fa06d4e4e5c570bc89a4a1bb9a9fb38a" Oct 05 21:53:12 crc kubenswrapper[4754]: I1005 21:53:12.276000 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-z45cg"] Oct 05 21:53:12 crc kubenswrapper[4754]: E1005 21:53:12.276979 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4209a211-0d5f-4a6c-95ef-5f1894bc0f11" containerName="extract-content" Oct 05 21:53:12 crc kubenswrapper[4754]: I1005 21:53:12.276995 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="4209a211-0d5f-4a6c-95ef-5f1894bc0f11" containerName="extract-content" Oct 05 21:53:12 crc kubenswrapper[4754]: E1005 21:53:12.277015 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be29e094-c72e-4ca0-9f8f-d9234767ed71" containerName="registry-server" Oct 05 21:53:12 crc kubenswrapper[4754]: I1005 21:53:12.277023 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="be29e094-c72e-4ca0-9f8f-d9234767ed71" containerName="registry-server" Oct 05 21:53:12 crc kubenswrapper[4754]: E1005 21:53:12.277040 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4209a211-0d5f-4a6c-95ef-5f1894bc0f11" containerName="registry-server" Oct 05 21:53:12 crc kubenswrapper[4754]: I1005 21:53:12.277050 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="4209a211-0d5f-4a6c-95ef-5f1894bc0f11" containerName="registry-server" Oct 05 21:53:12 crc kubenswrapper[4754]: E1005 21:53:12.277075 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4209a211-0d5f-4a6c-95ef-5f1894bc0f11" containerName="extract-utilities" Oct 05 21:53:12 crc kubenswrapper[4754]: I1005 21:53:12.277083 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="4209a211-0d5f-4a6c-95ef-5f1894bc0f11" containerName="extract-utilities" Oct 05 21:53:12 crc kubenswrapper[4754]: E1005 21:53:12.277099 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be29e094-c72e-4ca0-9f8f-d9234767ed71" containerName="extract-utilities" Oct 05 21:53:12 crc kubenswrapper[4754]: I1005 21:53:12.277106 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="be29e094-c72e-4ca0-9f8f-d9234767ed71" containerName="extract-utilities" Oct 05 21:53:12 crc kubenswrapper[4754]: E1005 21:53:12.277122 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be29e094-c72e-4ca0-9f8f-d9234767ed71" containerName="extract-content" Oct 05 21:53:12 crc kubenswrapper[4754]: I1005 21:53:12.277130 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="be29e094-c72e-4ca0-9f8f-d9234767ed71" containerName="extract-content" Oct 05 21:53:12 crc kubenswrapper[4754]: I1005 21:53:12.277358 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="be29e094-c72e-4ca0-9f8f-d9234767ed71" containerName="registry-server" Oct 05 21:53:12 crc kubenswrapper[4754]: I1005 21:53:12.277371 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="4209a211-0d5f-4a6c-95ef-5f1894bc0f11" containerName="registry-server" Oct 05 21:53:12 crc kubenswrapper[4754]: I1005 21:53:12.279027 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-z45cg" Oct 05 21:53:12 crc kubenswrapper[4754]: I1005 21:53:12.349681 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-z45cg"] Oct 05 21:53:12 crc kubenswrapper[4754]: I1005 21:53:12.397788 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xc6dk\" (UniqueName: \"kubernetes.io/projected/f6ff64a6-20db-4dbd-9bc2-bff129dfd9b7-kube-api-access-xc6dk\") pod \"redhat-operators-z45cg\" (UID: \"f6ff64a6-20db-4dbd-9bc2-bff129dfd9b7\") " pod="openshift-marketplace/redhat-operators-z45cg" Oct 05 21:53:12 crc kubenswrapper[4754]: I1005 21:53:12.397906 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f6ff64a6-20db-4dbd-9bc2-bff129dfd9b7-catalog-content\") pod \"redhat-operators-z45cg\" (UID: \"f6ff64a6-20db-4dbd-9bc2-bff129dfd9b7\") " pod="openshift-marketplace/redhat-operators-z45cg" Oct 05 21:53:12 crc kubenswrapper[4754]: I1005 21:53:12.397954 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f6ff64a6-20db-4dbd-9bc2-bff129dfd9b7-utilities\") pod \"redhat-operators-z45cg\" (UID: \"f6ff64a6-20db-4dbd-9bc2-bff129dfd9b7\") " pod="openshift-marketplace/redhat-operators-z45cg" Oct 05 21:53:12 crc kubenswrapper[4754]: I1005 21:53:12.499965 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f6ff64a6-20db-4dbd-9bc2-bff129dfd9b7-utilities\") pod \"redhat-operators-z45cg\" (UID: \"f6ff64a6-20db-4dbd-9bc2-bff129dfd9b7\") " pod="openshift-marketplace/redhat-operators-z45cg" Oct 05 21:53:12 crc kubenswrapper[4754]: I1005 21:53:12.500202 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xc6dk\" (UniqueName: \"kubernetes.io/projected/f6ff64a6-20db-4dbd-9bc2-bff129dfd9b7-kube-api-access-xc6dk\") pod \"redhat-operators-z45cg\" (UID: \"f6ff64a6-20db-4dbd-9bc2-bff129dfd9b7\") " pod="openshift-marketplace/redhat-operators-z45cg" Oct 05 21:53:12 crc kubenswrapper[4754]: I1005 21:53:12.500338 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f6ff64a6-20db-4dbd-9bc2-bff129dfd9b7-catalog-content\") pod \"redhat-operators-z45cg\" (UID: \"f6ff64a6-20db-4dbd-9bc2-bff129dfd9b7\") " pod="openshift-marketplace/redhat-operators-z45cg" Oct 05 21:53:12 crc kubenswrapper[4754]: I1005 21:53:12.500697 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f6ff64a6-20db-4dbd-9bc2-bff129dfd9b7-utilities\") pod \"redhat-operators-z45cg\" (UID: \"f6ff64a6-20db-4dbd-9bc2-bff129dfd9b7\") " pod="openshift-marketplace/redhat-operators-z45cg" Oct 05 21:53:12 crc kubenswrapper[4754]: I1005 21:53:12.500803 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f6ff64a6-20db-4dbd-9bc2-bff129dfd9b7-catalog-content\") pod \"redhat-operators-z45cg\" (UID: \"f6ff64a6-20db-4dbd-9bc2-bff129dfd9b7\") " pod="openshift-marketplace/redhat-operators-z45cg" Oct 05 21:53:12 crc kubenswrapper[4754]: I1005 21:53:12.533317 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xc6dk\" (UniqueName: \"kubernetes.io/projected/f6ff64a6-20db-4dbd-9bc2-bff129dfd9b7-kube-api-access-xc6dk\") pod \"redhat-operators-z45cg\" (UID: \"f6ff64a6-20db-4dbd-9bc2-bff129dfd9b7\") " pod="openshift-marketplace/redhat-operators-z45cg" Oct 05 21:53:12 crc kubenswrapper[4754]: I1005 21:53:12.608351 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-z45cg" Oct 05 21:53:13 crc kubenswrapper[4754]: I1005 21:53:13.748739 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-z45cg"] Oct 05 21:53:14 crc kubenswrapper[4754]: I1005 21:53:14.760992 4754 generic.go:334] "Generic (PLEG): container finished" podID="f6ff64a6-20db-4dbd-9bc2-bff129dfd9b7" containerID="053f581e7ada7f18324454db8ffc62bbb2fa4a7591088d2b08b7a7e56a11ebdb" exitCode=0 Oct 05 21:53:14 crc kubenswrapper[4754]: I1005 21:53:14.761093 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z45cg" event={"ID":"f6ff64a6-20db-4dbd-9bc2-bff129dfd9b7","Type":"ContainerDied","Data":"053f581e7ada7f18324454db8ffc62bbb2fa4a7591088d2b08b7a7e56a11ebdb"} Oct 05 21:53:14 crc kubenswrapper[4754]: I1005 21:53:14.761726 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z45cg" event={"ID":"f6ff64a6-20db-4dbd-9bc2-bff129dfd9b7","Type":"ContainerStarted","Data":"150a7bd4669e4743ff16a03bd130f9504ffb565e037c03d8b481a7a7f18befb7"} Oct 05 21:53:14 crc kubenswrapper[4754]: I1005 21:53:14.764104 4754 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 05 21:53:16 crc kubenswrapper[4754]: I1005 21:53:16.788836 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z45cg" event={"ID":"f6ff64a6-20db-4dbd-9bc2-bff129dfd9b7","Type":"ContainerStarted","Data":"6b228fa4aeccef93dfb63b8812c2ed773f3bcccea7d41affd0f1ddbb76144d32"} Oct 05 21:53:18 crc kubenswrapper[4754]: I1005 21:53:18.686684 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-rm2wg"] Oct 05 21:53:18 crc kubenswrapper[4754]: I1005 21:53:18.690860 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rm2wg" Oct 05 21:53:18 crc kubenswrapper[4754]: I1005 21:53:18.713653 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rm2wg"] Oct 05 21:53:18 crc kubenswrapper[4754]: I1005 21:53:18.720667 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d0a6306-03be-4e22-8e16-b654f2acc455-utilities\") pod \"certified-operators-rm2wg\" (UID: \"1d0a6306-03be-4e22-8e16-b654f2acc455\") " pod="openshift-marketplace/certified-operators-rm2wg" Oct 05 21:53:18 crc kubenswrapper[4754]: I1005 21:53:18.720803 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d0a6306-03be-4e22-8e16-b654f2acc455-catalog-content\") pod \"certified-operators-rm2wg\" (UID: \"1d0a6306-03be-4e22-8e16-b654f2acc455\") " pod="openshift-marketplace/certified-operators-rm2wg" Oct 05 21:53:18 crc kubenswrapper[4754]: I1005 21:53:18.720904 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d25m2\" (UniqueName: \"kubernetes.io/projected/1d0a6306-03be-4e22-8e16-b654f2acc455-kube-api-access-d25m2\") pod \"certified-operators-rm2wg\" (UID: \"1d0a6306-03be-4e22-8e16-b654f2acc455\") " pod="openshift-marketplace/certified-operators-rm2wg" Oct 05 21:53:18 crc kubenswrapper[4754]: I1005 21:53:18.843888 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d25m2\" (UniqueName: \"kubernetes.io/projected/1d0a6306-03be-4e22-8e16-b654f2acc455-kube-api-access-d25m2\") pod \"certified-operators-rm2wg\" (UID: \"1d0a6306-03be-4e22-8e16-b654f2acc455\") " pod="openshift-marketplace/certified-operators-rm2wg" Oct 05 21:53:18 crc kubenswrapper[4754]: I1005 21:53:18.843950 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d0a6306-03be-4e22-8e16-b654f2acc455-utilities\") pod \"certified-operators-rm2wg\" (UID: \"1d0a6306-03be-4e22-8e16-b654f2acc455\") " pod="openshift-marketplace/certified-operators-rm2wg" Oct 05 21:53:18 crc kubenswrapper[4754]: I1005 21:53:18.844037 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d0a6306-03be-4e22-8e16-b654f2acc455-catalog-content\") pod \"certified-operators-rm2wg\" (UID: \"1d0a6306-03be-4e22-8e16-b654f2acc455\") " pod="openshift-marketplace/certified-operators-rm2wg" Oct 05 21:53:18 crc kubenswrapper[4754]: I1005 21:53:18.844518 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d0a6306-03be-4e22-8e16-b654f2acc455-catalog-content\") pod \"certified-operators-rm2wg\" (UID: \"1d0a6306-03be-4e22-8e16-b654f2acc455\") " pod="openshift-marketplace/certified-operators-rm2wg" Oct 05 21:53:18 crc kubenswrapper[4754]: I1005 21:53:18.844735 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d0a6306-03be-4e22-8e16-b654f2acc455-utilities\") pod \"certified-operators-rm2wg\" (UID: \"1d0a6306-03be-4e22-8e16-b654f2acc455\") " pod="openshift-marketplace/certified-operators-rm2wg" Oct 05 21:53:18 crc kubenswrapper[4754]: I1005 21:53:18.869671 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d25m2\" (UniqueName: \"kubernetes.io/projected/1d0a6306-03be-4e22-8e16-b654f2acc455-kube-api-access-d25m2\") pod \"certified-operators-rm2wg\" (UID: \"1d0a6306-03be-4e22-8e16-b654f2acc455\") " pod="openshift-marketplace/certified-operators-rm2wg" Oct 05 21:53:19 crc kubenswrapper[4754]: I1005 21:53:19.008582 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rm2wg" Oct 05 21:53:19 crc kubenswrapper[4754]: I1005 21:53:19.648901 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rm2wg"] Oct 05 21:53:19 crc kubenswrapper[4754]: I1005 21:53:19.853463 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rm2wg" event={"ID":"1d0a6306-03be-4e22-8e16-b654f2acc455","Type":"ContainerStarted","Data":"88d1d640471fc50109fbc0104d2f4abdc8aa95362f418685988cf73ca2e16ade"} Oct 05 21:53:20 crc kubenswrapper[4754]: I1005 21:53:20.869702 4754 generic.go:334] "Generic (PLEG): container finished" podID="1d0a6306-03be-4e22-8e16-b654f2acc455" containerID="0452a285fc8a982fa64309b67fd1f121cb1384a817b100627f62210d04b7d487" exitCode=0 Oct 05 21:53:20 crc kubenswrapper[4754]: I1005 21:53:20.869973 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rm2wg" event={"ID":"1d0a6306-03be-4e22-8e16-b654f2acc455","Type":"ContainerDied","Data":"0452a285fc8a982fa64309b67fd1f121cb1384a817b100627f62210d04b7d487"} Oct 05 21:53:22 crc kubenswrapper[4754]: I1005 21:53:22.904424 4754 generic.go:334] "Generic (PLEG): container finished" podID="f6ff64a6-20db-4dbd-9bc2-bff129dfd9b7" containerID="6b228fa4aeccef93dfb63b8812c2ed773f3bcccea7d41affd0f1ddbb76144d32" exitCode=0 Oct 05 21:53:22 crc kubenswrapper[4754]: I1005 21:53:22.904883 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z45cg" event={"ID":"f6ff64a6-20db-4dbd-9bc2-bff129dfd9b7","Type":"ContainerDied","Data":"6b228fa4aeccef93dfb63b8812c2ed773f3bcccea7d41affd0f1ddbb76144d32"} Oct 05 21:53:22 crc kubenswrapper[4754]: I1005 21:53:22.915376 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rm2wg" event={"ID":"1d0a6306-03be-4e22-8e16-b654f2acc455","Type":"ContainerStarted","Data":"c72d10144c1de981b5d5c2d13a19d9f361bca6bfa5ea078e04b3b92fca86320c"} Oct 05 21:53:23 crc kubenswrapper[4754]: I1005 21:53:23.928453 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z45cg" event={"ID":"f6ff64a6-20db-4dbd-9bc2-bff129dfd9b7","Type":"ContainerStarted","Data":"1893ea59b21c44fd497a6e9ae48a3d8289800cc753ce9393eff03555654e7dea"} Oct 05 21:53:23 crc kubenswrapper[4754]: I1005 21:53:23.948726 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-z45cg" podStartSLOduration=3.297773565 podStartE2EDuration="11.948709256s" podCreationTimestamp="2025-10-05 21:53:12 +0000 UTC" firstStartedPulling="2025-10-05 21:53:14.763780457 +0000 UTC m=+3518.667899167" lastFinishedPulling="2025-10-05 21:53:23.414716148 +0000 UTC m=+3527.318834858" observedRunningTime="2025-10-05 21:53:23.944751173 +0000 UTC m=+3527.848869883" watchObservedRunningTime="2025-10-05 21:53:23.948709256 +0000 UTC m=+3527.852827966" Oct 05 21:53:24 crc kubenswrapper[4754]: I1005 21:53:24.937846 4754 generic.go:334] "Generic (PLEG): container finished" podID="1d0a6306-03be-4e22-8e16-b654f2acc455" containerID="c72d10144c1de981b5d5c2d13a19d9f361bca6bfa5ea078e04b3b92fca86320c" exitCode=0 Oct 05 21:53:24 crc kubenswrapper[4754]: I1005 21:53:24.937919 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rm2wg" event={"ID":"1d0a6306-03be-4e22-8e16-b654f2acc455","Type":"ContainerDied","Data":"c72d10144c1de981b5d5c2d13a19d9f361bca6bfa5ea078e04b3b92fca86320c"} Oct 05 21:53:25 crc kubenswrapper[4754]: I1005 21:53:25.947029 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rm2wg" event={"ID":"1d0a6306-03be-4e22-8e16-b654f2acc455","Type":"ContainerStarted","Data":"7bee969956a79dc362a3ed0a14f0d12fb20387ec57145b14fe26c049893efa8b"} Oct 05 21:53:25 crc kubenswrapper[4754]: I1005 21:53:25.966767 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-rm2wg" podStartSLOduration=3.495258214 podStartE2EDuration="7.966748269s" podCreationTimestamp="2025-10-05 21:53:18 +0000 UTC" firstStartedPulling="2025-10-05 21:53:20.871700951 +0000 UTC m=+3524.775819661" lastFinishedPulling="2025-10-05 21:53:25.343190996 +0000 UTC m=+3529.247309716" observedRunningTime="2025-10-05 21:53:25.961925224 +0000 UTC m=+3529.866043924" watchObservedRunningTime="2025-10-05 21:53:25.966748269 +0000 UTC m=+3529.870866969" Oct 05 21:53:29 crc kubenswrapper[4754]: I1005 21:53:29.008913 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-rm2wg" Oct 05 21:53:29 crc kubenswrapper[4754]: I1005 21:53:29.009273 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-rm2wg" Oct 05 21:53:30 crc kubenswrapper[4754]: I1005 21:53:30.060343 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-rm2wg" podUID="1d0a6306-03be-4e22-8e16-b654f2acc455" containerName="registry-server" probeResult="failure" output=< Oct 05 21:53:30 crc kubenswrapper[4754]: timeout: failed to connect service ":50051" within 1s Oct 05 21:53:30 crc kubenswrapper[4754]: > Oct 05 21:53:32 crc kubenswrapper[4754]: I1005 21:53:32.609977 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-z45cg" Oct 05 21:53:32 crc kubenswrapper[4754]: I1005 21:53:32.610030 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-z45cg" Oct 05 21:53:32 crc kubenswrapper[4754]: I1005 21:53:32.664350 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-z45cg" Oct 05 21:53:33 crc kubenswrapper[4754]: I1005 21:53:33.062912 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-z45cg" Oct 05 21:53:33 crc kubenswrapper[4754]: I1005 21:53:33.102604 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-z45cg"] Oct 05 21:53:35 crc kubenswrapper[4754]: I1005 21:53:35.016681 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-z45cg" podUID="f6ff64a6-20db-4dbd-9bc2-bff129dfd9b7" containerName="registry-server" containerID="cri-o://1893ea59b21c44fd497a6e9ae48a3d8289800cc753ce9393eff03555654e7dea" gracePeriod=2 Oct 05 21:53:35 crc kubenswrapper[4754]: I1005 21:53:35.571939 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-z45cg" Oct 05 21:53:35 crc kubenswrapper[4754]: I1005 21:53:35.699197 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xc6dk\" (UniqueName: \"kubernetes.io/projected/f6ff64a6-20db-4dbd-9bc2-bff129dfd9b7-kube-api-access-xc6dk\") pod \"f6ff64a6-20db-4dbd-9bc2-bff129dfd9b7\" (UID: \"f6ff64a6-20db-4dbd-9bc2-bff129dfd9b7\") " Oct 05 21:53:35 crc kubenswrapper[4754]: I1005 21:53:35.699329 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f6ff64a6-20db-4dbd-9bc2-bff129dfd9b7-utilities\") pod \"f6ff64a6-20db-4dbd-9bc2-bff129dfd9b7\" (UID: \"f6ff64a6-20db-4dbd-9bc2-bff129dfd9b7\") " Oct 05 21:53:35 crc kubenswrapper[4754]: I1005 21:53:35.699584 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f6ff64a6-20db-4dbd-9bc2-bff129dfd9b7-catalog-content\") pod \"f6ff64a6-20db-4dbd-9bc2-bff129dfd9b7\" (UID: \"f6ff64a6-20db-4dbd-9bc2-bff129dfd9b7\") " Oct 05 21:53:35 crc kubenswrapper[4754]: I1005 21:53:35.700159 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f6ff64a6-20db-4dbd-9bc2-bff129dfd9b7-utilities" (OuterVolumeSpecName: "utilities") pod "f6ff64a6-20db-4dbd-9bc2-bff129dfd9b7" (UID: "f6ff64a6-20db-4dbd-9bc2-bff129dfd9b7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 21:53:35 crc kubenswrapper[4754]: I1005 21:53:35.721606 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f6ff64a6-20db-4dbd-9bc2-bff129dfd9b7-kube-api-access-xc6dk" (OuterVolumeSpecName: "kube-api-access-xc6dk") pod "f6ff64a6-20db-4dbd-9bc2-bff129dfd9b7" (UID: "f6ff64a6-20db-4dbd-9bc2-bff129dfd9b7"). InnerVolumeSpecName "kube-api-access-xc6dk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:53:35 crc kubenswrapper[4754]: I1005 21:53:35.797915 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f6ff64a6-20db-4dbd-9bc2-bff129dfd9b7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f6ff64a6-20db-4dbd-9bc2-bff129dfd9b7" (UID: "f6ff64a6-20db-4dbd-9bc2-bff129dfd9b7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 21:53:35 crc kubenswrapper[4754]: I1005 21:53:35.802044 4754 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f6ff64a6-20db-4dbd-9bc2-bff129dfd9b7-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 21:53:35 crc kubenswrapper[4754]: I1005 21:53:35.802069 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xc6dk\" (UniqueName: \"kubernetes.io/projected/f6ff64a6-20db-4dbd-9bc2-bff129dfd9b7-kube-api-access-xc6dk\") on node \"crc\" DevicePath \"\"" Oct 05 21:53:35 crc kubenswrapper[4754]: I1005 21:53:35.802081 4754 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f6ff64a6-20db-4dbd-9bc2-bff129dfd9b7-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 21:53:36 crc kubenswrapper[4754]: I1005 21:53:36.027706 4754 generic.go:334] "Generic (PLEG): container finished" podID="f6ff64a6-20db-4dbd-9bc2-bff129dfd9b7" containerID="1893ea59b21c44fd497a6e9ae48a3d8289800cc753ce9393eff03555654e7dea" exitCode=0 Oct 05 21:53:36 crc kubenswrapper[4754]: I1005 21:53:36.027753 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z45cg" event={"ID":"f6ff64a6-20db-4dbd-9bc2-bff129dfd9b7","Type":"ContainerDied","Data":"1893ea59b21c44fd497a6e9ae48a3d8289800cc753ce9393eff03555654e7dea"} Oct 05 21:53:36 crc kubenswrapper[4754]: I1005 21:53:36.027780 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z45cg" event={"ID":"f6ff64a6-20db-4dbd-9bc2-bff129dfd9b7","Type":"ContainerDied","Data":"150a7bd4669e4743ff16a03bd130f9504ffb565e037c03d8b481a7a7f18befb7"} Oct 05 21:53:36 crc kubenswrapper[4754]: I1005 21:53:36.027797 4754 scope.go:117] "RemoveContainer" containerID="1893ea59b21c44fd497a6e9ae48a3d8289800cc753ce9393eff03555654e7dea" Oct 05 21:53:36 crc kubenswrapper[4754]: I1005 21:53:36.027832 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-z45cg" Oct 05 21:53:36 crc kubenswrapper[4754]: I1005 21:53:36.068089 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-z45cg"] Oct 05 21:53:36 crc kubenswrapper[4754]: I1005 21:53:36.072045 4754 scope.go:117] "RemoveContainer" containerID="6b228fa4aeccef93dfb63b8812c2ed773f3bcccea7d41affd0f1ddbb76144d32" Oct 05 21:53:36 crc kubenswrapper[4754]: I1005 21:53:36.076184 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-z45cg"] Oct 05 21:53:36 crc kubenswrapper[4754]: I1005 21:53:36.100142 4754 scope.go:117] "RemoveContainer" containerID="053f581e7ada7f18324454db8ffc62bbb2fa4a7591088d2b08b7a7e56a11ebdb" Oct 05 21:53:36 crc kubenswrapper[4754]: I1005 21:53:36.138531 4754 scope.go:117] "RemoveContainer" containerID="1893ea59b21c44fd497a6e9ae48a3d8289800cc753ce9393eff03555654e7dea" Oct 05 21:53:36 crc kubenswrapper[4754]: E1005 21:53:36.139646 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1893ea59b21c44fd497a6e9ae48a3d8289800cc753ce9393eff03555654e7dea\": container with ID starting with 1893ea59b21c44fd497a6e9ae48a3d8289800cc753ce9393eff03555654e7dea not found: ID does not exist" containerID="1893ea59b21c44fd497a6e9ae48a3d8289800cc753ce9393eff03555654e7dea" Oct 05 21:53:36 crc kubenswrapper[4754]: I1005 21:53:36.139847 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1893ea59b21c44fd497a6e9ae48a3d8289800cc753ce9393eff03555654e7dea"} err="failed to get container status \"1893ea59b21c44fd497a6e9ae48a3d8289800cc753ce9393eff03555654e7dea\": rpc error: code = NotFound desc = could not find container \"1893ea59b21c44fd497a6e9ae48a3d8289800cc753ce9393eff03555654e7dea\": container with ID starting with 1893ea59b21c44fd497a6e9ae48a3d8289800cc753ce9393eff03555654e7dea not found: ID does not exist" Oct 05 21:53:36 crc kubenswrapper[4754]: I1005 21:53:36.139869 4754 scope.go:117] "RemoveContainer" containerID="6b228fa4aeccef93dfb63b8812c2ed773f3bcccea7d41affd0f1ddbb76144d32" Oct 05 21:53:36 crc kubenswrapper[4754]: E1005 21:53:36.140139 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6b228fa4aeccef93dfb63b8812c2ed773f3bcccea7d41affd0f1ddbb76144d32\": container with ID starting with 6b228fa4aeccef93dfb63b8812c2ed773f3bcccea7d41affd0f1ddbb76144d32 not found: ID does not exist" containerID="6b228fa4aeccef93dfb63b8812c2ed773f3bcccea7d41affd0f1ddbb76144d32" Oct 05 21:53:36 crc kubenswrapper[4754]: I1005 21:53:36.140153 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6b228fa4aeccef93dfb63b8812c2ed773f3bcccea7d41affd0f1ddbb76144d32"} err="failed to get container status \"6b228fa4aeccef93dfb63b8812c2ed773f3bcccea7d41affd0f1ddbb76144d32\": rpc error: code = NotFound desc = could not find container \"6b228fa4aeccef93dfb63b8812c2ed773f3bcccea7d41affd0f1ddbb76144d32\": container with ID starting with 6b228fa4aeccef93dfb63b8812c2ed773f3bcccea7d41affd0f1ddbb76144d32 not found: ID does not exist" Oct 05 21:53:36 crc kubenswrapper[4754]: I1005 21:53:36.140179 4754 scope.go:117] "RemoveContainer" containerID="053f581e7ada7f18324454db8ffc62bbb2fa4a7591088d2b08b7a7e56a11ebdb" Oct 05 21:53:36 crc kubenswrapper[4754]: E1005 21:53:36.140742 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"053f581e7ada7f18324454db8ffc62bbb2fa4a7591088d2b08b7a7e56a11ebdb\": container with ID starting with 053f581e7ada7f18324454db8ffc62bbb2fa4a7591088d2b08b7a7e56a11ebdb not found: ID does not exist" containerID="053f581e7ada7f18324454db8ffc62bbb2fa4a7591088d2b08b7a7e56a11ebdb" Oct 05 21:53:36 crc kubenswrapper[4754]: I1005 21:53:36.140761 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"053f581e7ada7f18324454db8ffc62bbb2fa4a7591088d2b08b7a7e56a11ebdb"} err="failed to get container status \"053f581e7ada7f18324454db8ffc62bbb2fa4a7591088d2b08b7a7e56a11ebdb\": rpc error: code = NotFound desc = could not find container \"053f581e7ada7f18324454db8ffc62bbb2fa4a7591088d2b08b7a7e56a11ebdb\": container with ID starting with 053f581e7ada7f18324454db8ffc62bbb2fa4a7591088d2b08b7a7e56a11ebdb not found: ID does not exist" Oct 05 21:53:36 crc kubenswrapper[4754]: I1005 21:53:36.859227 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f6ff64a6-20db-4dbd-9bc2-bff129dfd9b7" path="/var/lib/kubelet/pods/f6ff64a6-20db-4dbd-9bc2-bff129dfd9b7/volumes" Oct 05 21:53:39 crc kubenswrapper[4754]: I1005 21:53:39.082870 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-rm2wg" Oct 05 21:53:39 crc kubenswrapper[4754]: I1005 21:53:39.146001 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-rm2wg" Oct 05 21:53:39 crc kubenswrapper[4754]: I1005 21:53:39.322517 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rm2wg"] Oct 05 21:53:41 crc kubenswrapper[4754]: I1005 21:53:41.090639 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-rm2wg" podUID="1d0a6306-03be-4e22-8e16-b654f2acc455" containerName="registry-server" containerID="cri-o://7bee969956a79dc362a3ed0a14f0d12fb20387ec57145b14fe26c049893efa8b" gracePeriod=2 Oct 05 21:53:41 crc kubenswrapper[4754]: I1005 21:53:41.709088 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rm2wg" Oct 05 21:53:41 crc kubenswrapper[4754]: I1005 21:53:41.845374 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d0a6306-03be-4e22-8e16-b654f2acc455-utilities\") pod \"1d0a6306-03be-4e22-8e16-b654f2acc455\" (UID: \"1d0a6306-03be-4e22-8e16-b654f2acc455\") " Oct 05 21:53:41 crc kubenswrapper[4754]: I1005 21:53:41.845543 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d0a6306-03be-4e22-8e16-b654f2acc455-catalog-content\") pod \"1d0a6306-03be-4e22-8e16-b654f2acc455\" (UID: \"1d0a6306-03be-4e22-8e16-b654f2acc455\") " Oct 05 21:53:41 crc kubenswrapper[4754]: I1005 21:53:41.845568 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d25m2\" (UniqueName: \"kubernetes.io/projected/1d0a6306-03be-4e22-8e16-b654f2acc455-kube-api-access-d25m2\") pod \"1d0a6306-03be-4e22-8e16-b654f2acc455\" (UID: \"1d0a6306-03be-4e22-8e16-b654f2acc455\") " Oct 05 21:53:41 crc kubenswrapper[4754]: I1005 21:53:41.846167 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d0a6306-03be-4e22-8e16-b654f2acc455-utilities" (OuterVolumeSpecName: "utilities") pod "1d0a6306-03be-4e22-8e16-b654f2acc455" (UID: "1d0a6306-03be-4e22-8e16-b654f2acc455"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 21:53:41 crc kubenswrapper[4754]: I1005 21:53:41.850819 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d0a6306-03be-4e22-8e16-b654f2acc455-kube-api-access-d25m2" (OuterVolumeSpecName: "kube-api-access-d25m2") pod "1d0a6306-03be-4e22-8e16-b654f2acc455" (UID: "1d0a6306-03be-4e22-8e16-b654f2acc455"). InnerVolumeSpecName "kube-api-access-d25m2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:53:41 crc kubenswrapper[4754]: I1005 21:53:41.903781 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d0a6306-03be-4e22-8e16-b654f2acc455-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d0a6306-03be-4e22-8e16-b654f2acc455" (UID: "1d0a6306-03be-4e22-8e16-b654f2acc455"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 21:53:41 crc kubenswrapper[4754]: I1005 21:53:41.947762 4754 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d0a6306-03be-4e22-8e16-b654f2acc455-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 21:53:41 crc kubenswrapper[4754]: I1005 21:53:41.948068 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d25m2\" (UniqueName: \"kubernetes.io/projected/1d0a6306-03be-4e22-8e16-b654f2acc455-kube-api-access-d25m2\") on node \"crc\" DevicePath \"\"" Oct 05 21:53:41 crc kubenswrapper[4754]: I1005 21:53:41.948081 4754 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d0a6306-03be-4e22-8e16-b654f2acc455-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 21:53:42 crc kubenswrapper[4754]: I1005 21:53:42.104706 4754 generic.go:334] "Generic (PLEG): container finished" podID="1d0a6306-03be-4e22-8e16-b654f2acc455" containerID="7bee969956a79dc362a3ed0a14f0d12fb20387ec57145b14fe26c049893efa8b" exitCode=0 Oct 05 21:53:42 crc kubenswrapper[4754]: I1005 21:53:42.104771 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rm2wg" event={"ID":"1d0a6306-03be-4e22-8e16-b654f2acc455","Type":"ContainerDied","Data":"7bee969956a79dc362a3ed0a14f0d12fb20387ec57145b14fe26c049893efa8b"} Oct 05 21:53:42 crc kubenswrapper[4754]: I1005 21:53:42.104822 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rm2wg" event={"ID":"1d0a6306-03be-4e22-8e16-b654f2acc455","Type":"ContainerDied","Data":"88d1d640471fc50109fbc0104d2f4abdc8aa95362f418685988cf73ca2e16ade"} Oct 05 21:53:42 crc kubenswrapper[4754]: I1005 21:53:42.104851 4754 scope.go:117] "RemoveContainer" containerID="7bee969956a79dc362a3ed0a14f0d12fb20387ec57145b14fe26c049893efa8b" Oct 05 21:53:42 crc kubenswrapper[4754]: I1005 21:53:42.107348 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rm2wg" Oct 05 21:53:42 crc kubenswrapper[4754]: I1005 21:53:42.151062 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rm2wg"] Oct 05 21:53:42 crc kubenswrapper[4754]: I1005 21:53:42.152865 4754 scope.go:117] "RemoveContainer" containerID="c72d10144c1de981b5d5c2d13a19d9f361bca6bfa5ea078e04b3b92fca86320c" Oct 05 21:53:42 crc kubenswrapper[4754]: I1005 21:53:42.156364 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-rm2wg"] Oct 05 21:53:42 crc kubenswrapper[4754]: I1005 21:53:42.177849 4754 scope.go:117] "RemoveContainer" containerID="0452a285fc8a982fa64309b67fd1f121cb1384a817b100627f62210d04b7d487" Oct 05 21:53:42 crc kubenswrapper[4754]: I1005 21:53:42.225261 4754 scope.go:117] "RemoveContainer" containerID="7bee969956a79dc362a3ed0a14f0d12fb20387ec57145b14fe26c049893efa8b" Oct 05 21:53:42 crc kubenswrapper[4754]: E1005 21:53:42.226039 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7bee969956a79dc362a3ed0a14f0d12fb20387ec57145b14fe26c049893efa8b\": container with ID starting with 7bee969956a79dc362a3ed0a14f0d12fb20387ec57145b14fe26c049893efa8b not found: ID does not exist" containerID="7bee969956a79dc362a3ed0a14f0d12fb20387ec57145b14fe26c049893efa8b" Oct 05 21:53:42 crc kubenswrapper[4754]: I1005 21:53:42.226073 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7bee969956a79dc362a3ed0a14f0d12fb20387ec57145b14fe26c049893efa8b"} err="failed to get container status \"7bee969956a79dc362a3ed0a14f0d12fb20387ec57145b14fe26c049893efa8b\": rpc error: code = NotFound desc = could not find container \"7bee969956a79dc362a3ed0a14f0d12fb20387ec57145b14fe26c049893efa8b\": container with ID starting with 7bee969956a79dc362a3ed0a14f0d12fb20387ec57145b14fe26c049893efa8b not found: ID does not exist" Oct 05 21:53:42 crc kubenswrapper[4754]: I1005 21:53:42.226098 4754 scope.go:117] "RemoveContainer" containerID="c72d10144c1de981b5d5c2d13a19d9f361bca6bfa5ea078e04b3b92fca86320c" Oct 05 21:53:42 crc kubenswrapper[4754]: E1005 21:53:42.226775 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c72d10144c1de981b5d5c2d13a19d9f361bca6bfa5ea078e04b3b92fca86320c\": container with ID starting with c72d10144c1de981b5d5c2d13a19d9f361bca6bfa5ea078e04b3b92fca86320c not found: ID does not exist" containerID="c72d10144c1de981b5d5c2d13a19d9f361bca6bfa5ea078e04b3b92fca86320c" Oct 05 21:53:42 crc kubenswrapper[4754]: I1005 21:53:42.227054 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c72d10144c1de981b5d5c2d13a19d9f361bca6bfa5ea078e04b3b92fca86320c"} err="failed to get container status \"c72d10144c1de981b5d5c2d13a19d9f361bca6bfa5ea078e04b3b92fca86320c\": rpc error: code = NotFound desc = could not find container \"c72d10144c1de981b5d5c2d13a19d9f361bca6bfa5ea078e04b3b92fca86320c\": container with ID starting with c72d10144c1de981b5d5c2d13a19d9f361bca6bfa5ea078e04b3b92fca86320c not found: ID does not exist" Oct 05 21:53:42 crc kubenswrapper[4754]: I1005 21:53:42.227217 4754 scope.go:117] "RemoveContainer" containerID="0452a285fc8a982fa64309b67fd1f121cb1384a817b100627f62210d04b7d487" Oct 05 21:53:42 crc kubenswrapper[4754]: E1005 21:53:42.227745 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0452a285fc8a982fa64309b67fd1f121cb1384a817b100627f62210d04b7d487\": container with ID starting with 0452a285fc8a982fa64309b67fd1f121cb1384a817b100627f62210d04b7d487 not found: ID does not exist" containerID="0452a285fc8a982fa64309b67fd1f121cb1384a817b100627f62210d04b7d487" Oct 05 21:53:42 crc kubenswrapper[4754]: I1005 21:53:42.227784 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0452a285fc8a982fa64309b67fd1f121cb1384a817b100627f62210d04b7d487"} err="failed to get container status \"0452a285fc8a982fa64309b67fd1f121cb1384a817b100627f62210d04b7d487\": rpc error: code = NotFound desc = could not find container \"0452a285fc8a982fa64309b67fd1f121cb1384a817b100627f62210d04b7d487\": container with ID starting with 0452a285fc8a982fa64309b67fd1f121cb1384a817b100627f62210d04b7d487 not found: ID does not exist" Oct 05 21:53:42 crc kubenswrapper[4754]: I1005 21:53:42.851914 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d0a6306-03be-4e22-8e16-b654f2acc455" path="/var/lib/kubelet/pods/1d0a6306-03be-4e22-8e16-b654f2acc455/volumes" Oct 05 21:54:05 crc kubenswrapper[4754]: I1005 21:54:05.245873 4754 patch_prober.go:28] interesting pod/machine-config-daemon-b2h9k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 21:54:05 crc kubenswrapper[4754]: I1005 21:54:05.246708 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 21:54:35 crc kubenswrapper[4754]: I1005 21:54:35.245361 4754 patch_prober.go:28] interesting pod/machine-config-daemon-b2h9k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 21:54:35 crc kubenswrapper[4754]: I1005 21:54:35.245993 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 21:55:05 crc kubenswrapper[4754]: I1005 21:55:05.245353 4754 patch_prober.go:28] interesting pod/machine-config-daemon-b2h9k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 21:55:05 crc kubenswrapper[4754]: I1005 21:55:05.246068 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 21:55:05 crc kubenswrapper[4754]: I1005 21:55:05.246132 4754 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" Oct 05 21:55:05 crc kubenswrapper[4754]: I1005 21:55:05.247210 4754 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f21169793903f3552e21e22d202f837dbee5ee4d48ced3d4666fe154c7e5b95c"} pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 05 21:55:05 crc kubenswrapper[4754]: I1005 21:55:05.247299 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" containerName="machine-config-daemon" containerID="cri-o://f21169793903f3552e21e22d202f837dbee5ee4d48ced3d4666fe154c7e5b95c" gracePeriod=600 Oct 05 21:55:05 crc kubenswrapper[4754]: E1005 21:55:05.383271 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 21:55:05 crc kubenswrapper[4754]: I1005 21:55:05.934312 4754 generic.go:334] "Generic (PLEG): container finished" podID="be95a413-6a59-45b4-84b7-b43cae694a26" containerID="f21169793903f3552e21e22d202f837dbee5ee4d48ced3d4666fe154c7e5b95c" exitCode=0 Oct 05 21:55:05 crc kubenswrapper[4754]: I1005 21:55:05.934363 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" event={"ID":"be95a413-6a59-45b4-84b7-b43cae694a26","Type":"ContainerDied","Data":"f21169793903f3552e21e22d202f837dbee5ee4d48ced3d4666fe154c7e5b95c"} Oct 05 21:55:05 crc kubenswrapper[4754]: I1005 21:55:05.934406 4754 scope.go:117] "RemoveContainer" containerID="8db5d0c4b86f257e1a334bb36f42cc5226cb59650fc43cdcfb1aaccb42c9a511" Oct 05 21:55:05 crc kubenswrapper[4754]: I1005 21:55:05.935144 4754 scope.go:117] "RemoveContainer" containerID="f21169793903f3552e21e22d202f837dbee5ee4d48ced3d4666fe154c7e5b95c" Oct 05 21:55:05 crc kubenswrapper[4754]: E1005 21:55:05.935478 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 21:55:18 crc kubenswrapper[4754]: I1005 21:55:18.837364 4754 scope.go:117] "RemoveContainer" containerID="f21169793903f3552e21e22d202f837dbee5ee4d48ced3d4666fe154c7e5b95c" Oct 05 21:55:18 crc kubenswrapper[4754]: E1005 21:55:18.838278 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 21:55:33 crc kubenswrapper[4754]: I1005 21:55:33.837369 4754 scope.go:117] "RemoveContainer" containerID="f21169793903f3552e21e22d202f837dbee5ee4d48ced3d4666fe154c7e5b95c" Oct 05 21:55:33 crc kubenswrapper[4754]: E1005 21:55:33.839897 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 21:55:44 crc kubenswrapper[4754]: I1005 21:55:44.838267 4754 scope.go:117] "RemoveContainer" containerID="f21169793903f3552e21e22d202f837dbee5ee4d48ced3d4666fe154c7e5b95c" Oct 05 21:55:44 crc kubenswrapper[4754]: E1005 21:55:44.839154 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 21:55:58 crc kubenswrapper[4754]: I1005 21:55:58.842971 4754 scope.go:117] "RemoveContainer" containerID="f21169793903f3552e21e22d202f837dbee5ee4d48ced3d4666fe154c7e5b95c" Oct 05 21:55:58 crc kubenswrapper[4754]: E1005 21:55:58.844300 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 21:56:10 crc kubenswrapper[4754]: I1005 21:56:10.838577 4754 scope.go:117] "RemoveContainer" containerID="f21169793903f3552e21e22d202f837dbee5ee4d48ced3d4666fe154c7e5b95c" Oct 05 21:56:10 crc kubenswrapper[4754]: E1005 21:56:10.839568 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 21:56:24 crc kubenswrapper[4754]: I1005 21:56:24.845968 4754 scope.go:117] "RemoveContainer" containerID="f21169793903f3552e21e22d202f837dbee5ee4d48ced3d4666fe154c7e5b95c" Oct 05 21:56:24 crc kubenswrapper[4754]: E1005 21:56:24.848538 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 21:56:28 crc kubenswrapper[4754]: I1005 21:56:28.685744 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-7r6lj"] Oct 05 21:56:28 crc kubenswrapper[4754]: E1005 21:56:28.687268 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d0a6306-03be-4e22-8e16-b654f2acc455" containerName="registry-server" Oct 05 21:56:28 crc kubenswrapper[4754]: I1005 21:56:28.687347 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d0a6306-03be-4e22-8e16-b654f2acc455" containerName="registry-server" Oct 05 21:56:28 crc kubenswrapper[4754]: E1005 21:56:28.687416 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6ff64a6-20db-4dbd-9bc2-bff129dfd9b7" containerName="extract-utilities" Oct 05 21:56:28 crc kubenswrapper[4754]: I1005 21:56:28.687472 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6ff64a6-20db-4dbd-9bc2-bff129dfd9b7" containerName="extract-utilities" Oct 05 21:56:28 crc kubenswrapper[4754]: E1005 21:56:28.687552 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d0a6306-03be-4e22-8e16-b654f2acc455" containerName="extract-content" Oct 05 21:56:28 crc kubenswrapper[4754]: I1005 21:56:28.687612 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d0a6306-03be-4e22-8e16-b654f2acc455" containerName="extract-content" Oct 05 21:56:28 crc kubenswrapper[4754]: E1005 21:56:28.687691 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d0a6306-03be-4e22-8e16-b654f2acc455" containerName="extract-utilities" Oct 05 21:56:28 crc kubenswrapper[4754]: I1005 21:56:28.687756 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d0a6306-03be-4e22-8e16-b654f2acc455" containerName="extract-utilities" Oct 05 21:56:28 crc kubenswrapper[4754]: E1005 21:56:28.687827 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6ff64a6-20db-4dbd-9bc2-bff129dfd9b7" containerName="extract-content" Oct 05 21:56:28 crc kubenswrapper[4754]: I1005 21:56:28.687885 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6ff64a6-20db-4dbd-9bc2-bff129dfd9b7" containerName="extract-content" Oct 05 21:56:28 crc kubenswrapper[4754]: E1005 21:56:28.687964 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6ff64a6-20db-4dbd-9bc2-bff129dfd9b7" containerName="registry-server" Oct 05 21:56:28 crc kubenswrapper[4754]: I1005 21:56:28.688020 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6ff64a6-20db-4dbd-9bc2-bff129dfd9b7" containerName="registry-server" Oct 05 21:56:28 crc kubenswrapper[4754]: I1005 21:56:28.688263 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="f6ff64a6-20db-4dbd-9bc2-bff129dfd9b7" containerName="registry-server" Oct 05 21:56:28 crc kubenswrapper[4754]: I1005 21:56:28.688348 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d0a6306-03be-4e22-8e16-b654f2acc455" containerName="registry-server" Oct 05 21:56:28 crc kubenswrapper[4754]: I1005 21:56:28.691012 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7r6lj" Oct 05 21:56:28 crc kubenswrapper[4754]: I1005 21:56:28.701228 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7r6lj"] Oct 05 21:56:28 crc kubenswrapper[4754]: I1005 21:56:28.844403 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7509312a-ff7a-4a56-9a81-ba655b8e4465-catalog-content\") pod \"redhat-marketplace-7r6lj\" (UID: \"7509312a-ff7a-4a56-9a81-ba655b8e4465\") " pod="openshift-marketplace/redhat-marketplace-7r6lj" Oct 05 21:56:28 crc kubenswrapper[4754]: I1005 21:56:28.844727 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7509312a-ff7a-4a56-9a81-ba655b8e4465-utilities\") pod \"redhat-marketplace-7r6lj\" (UID: \"7509312a-ff7a-4a56-9a81-ba655b8e4465\") " pod="openshift-marketplace/redhat-marketplace-7r6lj" Oct 05 21:56:28 crc kubenswrapper[4754]: I1005 21:56:28.844776 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fbwhw\" (UniqueName: \"kubernetes.io/projected/7509312a-ff7a-4a56-9a81-ba655b8e4465-kube-api-access-fbwhw\") pod \"redhat-marketplace-7r6lj\" (UID: \"7509312a-ff7a-4a56-9a81-ba655b8e4465\") " pod="openshift-marketplace/redhat-marketplace-7r6lj" Oct 05 21:56:28 crc kubenswrapper[4754]: I1005 21:56:28.946588 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7509312a-ff7a-4a56-9a81-ba655b8e4465-catalog-content\") pod \"redhat-marketplace-7r6lj\" (UID: \"7509312a-ff7a-4a56-9a81-ba655b8e4465\") " pod="openshift-marketplace/redhat-marketplace-7r6lj" Oct 05 21:56:28 crc kubenswrapper[4754]: I1005 21:56:28.946642 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7509312a-ff7a-4a56-9a81-ba655b8e4465-utilities\") pod \"redhat-marketplace-7r6lj\" (UID: \"7509312a-ff7a-4a56-9a81-ba655b8e4465\") " pod="openshift-marketplace/redhat-marketplace-7r6lj" Oct 05 21:56:28 crc kubenswrapper[4754]: I1005 21:56:28.946684 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fbwhw\" (UniqueName: \"kubernetes.io/projected/7509312a-ff7a-4a56-9a81-ba655b8e4465-kube-api-access-fbwhw\") pod \"redhat-marketplace-7r6lj\" (UID: \"7509312a-ff7a-4a56-9a81-ba655b8e4465\") " pod="openshift-marketplace/redhat-marketplace-7r6lj" Oct 05 21:56:28 crc kubenswrapper[4754]: I1005 21:56:28.947617 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7509312a-ff7a-4a56-9a81-ba655b8e4465-catalog-content\") pod \"redhat-marketplace-7r6lj\" (UID: \"7509312a-ff7a-4a56-9a81-ba655b8e4465\") " pod="openshift-marketplace/redhat-marketplace-7r6lj" Oct 05 21:56:28 crc kubenswrapper[4754]: I1005 21:56:28.947700 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7509312a-ff7a-4a56-9a81-ba655b8e4465-utilities\") pod \"redhat-marketplace-7r6lj\" (UID: \"7509312a-ff7a-4a56-9a81-ba655b8e4465\") " pod="openshift-marketplace/redhat-marketplace-7r6lj" Oct 05 21:56:28 crc kubenswrapper[4754]: I1005 21:56:28.968454 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fbwhw\" (UniqueName: \"kubernetes.io/projected/7509312a-ff7a-4a56-9a81-ba655b8e4465-kube-api-access-fbwhw\") pod \"redhat-marketplace-7r6lj\" (UID: \"7509312a-ff7a-4a56-9a81-ba655b8e4465\") " pod="openshift-marketplace/redhat-marketplace-7r6lj" Oct 05 21:56:29 crc kubenswrapper[4754]: I1005 21:56:29.082693 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7r6lj" Oct 05 21:56:29 crc kubenswrapper[4754]: I1005 21:56:29.610180 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7r6lj"] Oct 05 21:56:29 crc kubenswrapper[4754]: I1005 21:56:29.865142 4754 generic.go:334] "Generic (PLEG): container finished" podID="7509312a-ff7a-4a56-9a81-ba655b8e4465" containerID="80614787497b399da214a781b37390251841bd29c41958cdf71993788f8bf082" exitCode=0 Oct 05 21:56:29 crc kubenswrapper[4754]: I1005 21:56:29.865249 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7r6lj" event={"ID":"7509312a-ff7a-4a56-9a81-ba655b8e4465","Type":"ContainerDied","Data":"80614787497b399da214a781b37390251841bd29c41958cdf71993788f8bf082"} Oct 05 21:56:29 crc kubenswrapper[4754]: I1005 21:56:29.865453 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7r6lj" event={"ID":"7509312a-ff7a-4a56-9a81-ba655b8e4465","Type":"ContainerStarted","Data":"d050712b8d72a7ed6bf63d181b3362a7780637926081d66f5bb802c934f09dd9"} Oct 05 21:56:30 crc kubenswrapper[4754]: I1005 21:56:30.874263 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7r6lj" event={"ID":"7509312a-ff7a-4a56-9a81-ba655b8e4465","Type":"ContainerStarted","Data":"f5f88c1309c691e8f0f2d02ef6da0ecb77025d8413ae1004eba33fa85d82f3ca"} Oct 05 21:56:31 crc kubenswrapper[4754]: I1005 21:56:31.882632 4754 generic.go:334] "Generic (PLEG): container finished" podID="7509312a-ff7a-4a56-9a81-ba655b8e4465" containerID="f5f88c1309c691e8f0f2d02ef6da0ecb77025d8413ae1004eba33fa85d82f3ca" exitCode=0 Oct 05 21:56:31 crc kubenswrapper[4754]: I1005 21:56:31.882697 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7r6lj" event={"ID":"7509312a-ff7a-4a56-9a81-ba655b8e4465","Type":"ContainerDied","Data":"f5f88c1309c691e8f0f2d02ef6da0ecb77025d8413ae1004eba33fa85d82f3ca"} Oct 05 21:56:32 crc kubenswrapper[4754]: I1005 21:56:32.892454 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7r6lj" event={"ID":"7509312a-ff7a-4a56-9a81-ba655b8e4465","Type":"ContainerStarted","Data":"01f700e6eef8dd6d1b948c7b6495d47d7e8ba828439988777b0e99db75198888"} Oct 05 21:56:39 crc kubenswrapper[4754]: I1005 21:56:39.083236 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-7r6lj" Oct 05 21:56:39 crc kubenswrapper[4754]: I1005 21:56:39.083894 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-7r6lj" Oct 05 21:56:39 crc kubenswrapper[4754]: I1005 21:56:39.152695 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-7r6lj" Oct 05 21:56:39 crc kubenswrapper[4754]: I1005 21:56:39.179896 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-7r6lj" podStartSLOduration=8.778390642 podStartE2EDuration="11.179878986s" podCreationTimestamp="2025-10-05 21:56:28 +0000 UTC" firstStartedPulling="2025-10-05 21:56:29.866288717 +0000 UTC m=+3713.770407427" lastFinishedPulling="2025-10-05 21:56:32.267777061 +0000 UTC m=+3716.171895771" observedRunningTime="2025-10-05 21:56:32.913470018 +0000 UTC m=+3716.817588738" watchObservedRunningTime="2025-10-05 21:56:39.179878986 +0000 UTC m=+3723.083997696" Oct 05 21:56:39 crc kubenswrapper[4754]: I1005 21:56:39.837951 4754 scope.go:117] "RemoveContainer" containerID="f21169793903f3552e21e22d202f837dbee5ee4d48ced3d4666fe154c7e5b95c" Oct 05 21:56:39 crc kubenswrapper[4754]: E1005 21:56:39.838760 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 21:56:40 crc kubenswrapper[4754]: I1005 21:56:40.009200 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-7r6lj" Oct 05 21:56:40 crc kubenswrapper[4754]: I1005 21:56:40.059749 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-7r6lj"] Oct 05 21:56:41 crc kubenswrapper[4754]: I1005 21:56:41.975037 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-7r6lj" podUID="7509312a-ff7a-4a56-9a81-ba655b8e4465" containerName="registry-server" containerID="cri-o://01f700e6eef8dd6d1b948c7b6495d47d7e8ba828439988777b0e99db75198888" gracePeriod=2 Oct 05 21:56:42 crc kubenswrapper[4754]: I1005 21:56:42.636126 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7r6lj" Oct 05 21:56:42 crc kubenswrapper[4754]: I1005 21:56:42.736861 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7509312a-ff7a-4a56-9a81-ba655b8e4465-catalog-content\") pod \"7509312a-ff7a-4a56-9a81-ba655b8e4465\" (UID: \"7509312a-ff7a-4a56-9a81-ba655b8e4465\") " Oct 05 21:56:42 crc kubenswrapper[4754]: I1005 21:56:42.736988 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fbwhw\" (UniqueName: \"kubernetes.io/projected/7509312a-ff7a-4a56-9a81-ba655b8e4465-kube-api-access-fbwhw\") pod \"7509312a-ff7a-4a56-9a81-ba655b8e4465\" (UID: \"7509312a-ff7a-4a56-9a81-ba655b8e4465\") " Oct 05 21:56:42 crc kubenswrapper[4754]: I1005 21:56:42.737097 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7509312a-ff7a-4a56-9a81-ba655b8e4465-utilities\") pod \"7509312a-ff7a-4a56-9a81-ba655b8e4465\" (UID: \"7509312a-ff7a-4a56-9a81-ba655b8e4465\") " Oct 05 21:56:42 crc kubenswrapper[4754]: I1005 21:56:42.738239 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7509312a-ff7a-4a56-9a81-ba655b8e4465-utilities" (OuterVolumeSpecName: "utilities") pod "7509312a-ff7a-4a56-9a81-ba655b8e4465" (UID: "7509312a-ff7a-4a56-9a81-ba655b8e4465"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 21:56:42 crc kubenswrapper[4754]: I1005 21:56:42.743128 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7509312a-ff7a-4a56-9a81-ba655b8e4465-kube-api-access-fbwhw" (OuterVolumeSpecName: "kube-api-access-fbwhw") pod "7509312a-ff7a-4a56-9a81-ba655b8e4465" (UID: "7509312a-ff7a-4a56-9a81-ba655b8e4465"). InnerVolumeSpecName "kube-api-access-fbwhw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:56:42 crc kubenswrapper[4754]: I1005 21:56:42.756215 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7509312a-ff7a-4a56-9a81-ba655b8e4465-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7509312a-ff7a-4a56-9a81-ba655b8e4465" (UID: "7509312a-ff7a-4a56-9a81-ba655b8e4465"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 21:56:42 crc kubenswrapper[4754]: I1005 21:56:42.839329 4754 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7509312a-ff7a-4a56-9a81-ba655b8e4465-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 21:56:42 crc kubenswrapper[4754]: I1005 21:56:42.839365 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fbwhw\" (UniqueName: \"kubernetes.io/projected/7509312a-ff7a-4a56-9a81-ba655b8e4465-kube-api-access-fbwhw\") on node \"crc\" DevicePath \"\"" Oct 05 21:56:42 crc kubenswrapper[4754]: I1005 21:56:42.839380 4754 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7509312a-ff7a-4a56-9a81-ba655b8e4465-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 21:56:42 crc kubenswrapper[4754]: I1005 21:56:42.997657 4754 generic.go:334] "Generic (PLEG): container finished" podID="7509312a-ff7a-4a56-9a81-ba655b8e4465" containerID="01f700e6eef8dd6d1b948c7b6495d47d7e8ba828439988777b0e99db75198888" exitCode=0 Oct 05 21:56:42 crc kubenswrapper[4754]: I1005 21:56:42.997793 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7r6lj" Oct 05 21:56:42 crc kubenswrapper[4754]: I1005 21:56:42.997803 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7r6lj" event={"ID":"7509312a-ff7a-4a56-9a81-ba655b8e4465","Type":"ContainerDied","Data":"01f700e6eef8dd6d1b948c7b6495d47d7e8ba828439988777b0e99db75198888"} Oct 05 21:56:42 crc kubenswrapper[4754]: I1005 21:56:42.998254 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7r6lj" event={"ID":"7509312a-ff7a-4a56-9a81-ba655b8e4465","Type":"ContainerDied","Data":"d050712b8d72a7ed6bf63d181b3362a7780637926081d66f5bb802c934f09dd9"} Oct 05 21:56:42 crc kubenswrapper[4754]: I1005 21:56:42.998288 4754 scope.go:117] "RemoveContainer" containerID="01f700e6eef8dd6d1b948c7b6495d47d7e8ba828439988777b0e99db75198888" Oct 05 21:56:43 crc kubenswrapper[4754]: I1005 21:56:43.028820 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-7r6lj"] Oct 05 21:56:43 crc kubenswrapper[4754]: I1005 21:56:43.039999 4754 scope.go:117] "RemoveContainer" containerID="f5f88c1309c691e8f0f2d02ef6da0ecb77025d8413ae1004eba33fa85d82f3ca" Oct 05 21:56:43 crc kubenswrapper[4754]: I1005 21:56:43.040083 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-7r6lj"] Oct 05 21:56:43 crc kubenswrapper[4754]: I1005 21:56:43.063178 4754 scope.go:117] "RemoveContainer" containerID="80614787497b399da214a781b37390251841bd29c41958cdf71993788f8bf082" Oct 05 21:56:43 crc kubenswrapper[4754]: I1005 21:56:43.105386 4754 scope.go:117] "RemoveContainer" containerID="01f700e6eef8dd6d1b948c7b6495d47d7e8ba828439988777b0e99db75198888" Oct 05 21:56:43 crc kubenswrapper[4754]: E1005 21:56:43.106149 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"01f700e6eef8dd6d1b948c7b6495d47d7e8ba828439988777b0e99db75198888\": container with ID starting with 01f700e6eef8dd6d1b948c7b6495d47d7e8ba828439988777b0e99db75198888 not found: ID does not exist" containerID="01f700e6eef8dd6d1b948c7b6495d47d7e8ba828439988777b0e99db75198888" Oct 05 21:56:43 crc kubenswrapper[4754]: I1005 21:56:43.106211 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"01f700e6eef8dd6d1b948c7b6495d47d7e8ba828439988777b0e99db75198888"} err="failed to get container status \"01f700e6eef8dd6d1b948c7b6495d47d7e8ba828439988777b0e99db75198888\": rpc error: code = NotFound desc = could not find container \"01f700e6eef8dd6d1b948c7b6495d47d7e8ba828439988777b0e99db75198888\": container with ID starting with 01f700e6eef8dd6d1b948c7b6495d47d7e8ba828439988777b0e99db75198888 not found: ID does not exist" Oct 05 21:56:43 crc kubenswrapper[4754]: I1005 21:56:43.106239 4754 scope.go:117] "RemoveContainer" containerID="f5f88c1309c691e8f0f2d02ef6da0ecb77025d8413ae1004eba33fa85d82f3ca" Oct 05 21:56:43 crc kubenswrapper[4754]: E1005 21:56:43.106835 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f5f88c1309c691e8f0f2d02ef6da0ecb77025d8413ae1004eba33fa85d82f3ca\": container with ID starting with f5f88c1309c691e8f0f2d02ef6da0ecb77025d8413ae1004eba33fa85d82f3ca not found: ID does not exist" containerID="f5f88c1309c691e8f0f2d02ef6da0ecb77025d8413ae1004eba33fa85d82f3ca" Oct 05 21:56:43 crc kubenswrapper[4754]: I1005 21:56:43.106867 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f5f88c1309c691e8f0f2d02ef6da0ecb77025d8413ae1004eba33fa85d82f3ca"} err="failed to get container status \"f5f88c1309c691e8f0f2d02ef6da0ecb77025d8413ae1004eba33fa85d82f3ca\": rpc error: code = NotFound desc = could not find container \"f5f88c1309c691e8f0f2d02ef6da0ecb77025d8413ae1004eba33fa85d82f3ca\": container with ID starting with f5f88c1309c691e8f0f2d02ef6da0ecb77025d8413ae1004eba33fa85d82f3ca not found: ID does not exist" Oct 05 21:56:43 crc kubenswrapper[4754]: I1005 21:56:43.106906 4754 scope.go:117] "RemoveContainer" containerID="80614787497b399da214a781b37390251841bd29c41958cdf71993788f8bf082" Oct 05 21:56:43 crc kubenswrapper[4754]: E1005 21:56:43.107337 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"80614787497b399da214a781b37390251841bd29c41958cdf71993788f8bf082\": container with ID starting with 80614787497b399da214a781b37390251841bd29c41958cdf71993788f8bf082 not found: ID does not exist" containerID="80614787497b399da214a781b37390251841bd29c41958cdf71993788f8bf082" Oct 05 21:56:43 crc kubenswrapper[4754]: I1005 21:56:43.107386 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"80614787497b399da214a781b37390251841bd29c41958cdf71993788f8bf082"} err="failed to get container status \"80614787497b399da214a781b37390251841bd29c41958cdf71993788f8bf082\": rpc error: code = NotFound desc = could not find container \"80614787497b399da214a781b37390251841bd29c41958cdf71993788f8bf082\": container with ID starting with 80614787497b399da214a781b37390251841bd29c41958cdf71993788f8bf082 not found: ID does not exist" Oct 05 21:56:44 crc kubenswrapper[4754]: I1005 21:56:44.856121 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7509312a-ff7a-4a56-9a81-ba655b8e4465" path="/var/lib/kubelet/pods/7509312a-ff7a-4a56-9a81-ba655b8e4465/volumes" Oct 05 21:56:53 crc kubenswrapper[4754]: I1005 21:56:53.838128 4754 scope.go:117] "RemoveContainer" containerID="f21169793903f3552e21e22d202f837dbee5ee4d48ced3d4666fe154c7e5b95c" Oct 05 21:56:53 crc kubenswrapper[4754]: E1005 21:56:53.839221 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 21:57:03 crc kubenswrapper[4754]: I1005 21:57:03.061303 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-dlznx"] Oct 05 21:57:03 crc kubenswrapper[4754]: E1005 21:57:03.063406 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7509312a-ff7a-4a56-9a81-ba655b8e4465" containerName="registry-server" Oct 05 21:57:03 crc kubenswrapper[4754]: I1005 21:57:03.063647 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="7509312a-ff7a-4a56-9a81-ba655b8e4465" containerName="registry-server" Oct 05 21:57:03 crc kubenswrapper[4754]: E1005 21:57:03.063753 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7509312a-ff7a-4a56-9a81-ba655b8e4465" containerName="extract-utilities" Oct 05 21:57:03 crc kubenswrapper[4754]: I1005 21:57:03.063845 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="7509312a-ff7a-4a56-9a81-ba655b8e4465" containerName="extract-utilities" Oct 05 21:57:03 crc kubenswrapper[4754]: E1005 21:57:03.063942 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7509312a-ff7a-4a56-9a81-ba655b8e4465" containerName="extract-content" Oct 05 21:57:03 crc kubenswrapper[4754]: I1005 21:57:03.064835 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="7509312a-ff7a-4a56-9a81-ba655b8e4465" containerName="extract-content" Oct 05 21:57:03 crc kubenswrapper[4754]: I1005 21:57:03.065275 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="7509312a-ff7a-4a56-9a81-ba655b8e4465" containerName="registry-server" Oct 05 21:57:03 crc kubenswrapper[4754]: I1005 21:57:03.068314 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dlznx" Oct 05 21:57:03 crc kubenswrapper[4754]: I1005 21:57:03.080885 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-dlznx"] Oct 05 21:57:03 crc kubenswrapper[4754]: I1005 21:57:03.254741 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ac7fab0-0e2a-47db-a7e6-5ab36fb9cf58-utilities\") pod \"community-operators-dlznx\" (UID: \"0ac7fab0-0e2a-47db-a7e6-5ab36fb9cf58\") " pod="openshift-marketplace/community-operators-dlznx" Oct 05 21:57:03 crc kubenswrapper[4754]: I1005 21:57:03.254956 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ac7fab0-0e2a-47db-a7e6-5ab36fb9cf58-catalog-content\") pod \"community-operators-dlznx\" (UID: \"0ac7fab0-0e2a-47db-a7e6-5ab36fb9cf58\") " pod="openshift-marketplace/community-operators-dlznx" Oct 05 21:57:03 crc kubenswrapper[4754]: I1005 21:57:03.255001 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sr69p\" (UniqueName: \"kubernetes.io/projected/0ac7fab0-0e2a-47db-a7e6-5ab36fb9cf58-kube-api-access-sr69p\") pod \"community-operators-dlznx\" (UID: \"0ac7fab0-0e2a-47db-a7e6-5ab36fb9cf58\") " pod="openshift-marketplace/community-operators-dlznx" Oct 05 21:57:03 crc kubenswrapper[4754]: I1005 21:57:03.356820 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ac7fab0-0e2a-47db-a7e6-5ab36fb9cf58-catalog-content\") pod \"community-operators-dlznx\" (UID: \"0ac7fab0-0e2a-47db-a7e6-5ab36fb9cf58\") " pod="openshift-marketplace/community-operators-dlznx" Oct 05 21:57:03 crc kubenswrapper[4754]: I1005 21:57:03.356878 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sr69p\" (UniqueName: \"kubernetes.io/projected/0ac7fab0-0e2a-47db-a7e6-5ab36fb9cf58-kube-api-access-sr69p\") pod \"community-operators-dlznx\" (UID: \"0ac7fab0-0e2a-47db-a7e6-5ab36fb9cf58\") " pod="openshift-marketplace/community-operators-dlznx" Oct 05 21:57:03 crc kubenswrapper[4754]: I1005 21:57:03.356920 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ac7fab0-0e2a-47db-a7e6-5ab36fb9cf58-utilities\") pod \"community-operators-dlznx\" (UID: \"0ac7fab0-0e2a-47db-a7e6-5ab36fb9cf58\") " pod="openshift-marketplace/community-operators-dlznx" Oct 05 21:57:03 crc kubenswrapper[4754]: I1005 21:57:03.357524 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ac7fab0-0e2a-47db-a7e6-5ab36fb9cf58-utilities\") pod \"community-operators-dlznx\" (UID: \"0ac7fab0-0e2a-47db-a7e6-5ab36fb9cf58\") " pod="openshift-marketplace/community-operators-dlznx" Oct 05 21:57:03 crc kubenswrapper[4754]: I1005 21:57:03.357632 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ac7fab0-0e2a-47db-a7e6-5ab36fb9cf58-catalog-content\") pod \"community-operators-dlznx\" (UID: \"0ac7fab0-0e2a-47db-a7e6-5ab36fb9cf58\") " pod="openshift-marketplace/community-operators-dlznx" Oct 05 21:57:03 crc kubenswrapper[4754]: I1005 21:57:03.386979 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sr69p\" (UniqueName: \"kubernetes.io/projected/0ac7fab0-0e2a-47db-a7e6-5ab36fb9cf58-kube-api-access-sr69p\") pod \"community-operators-dlznx\" (UID: \"0ac7fab0-0e2a-47db-a7e6-5ab36fb9cf58\") " pod="openshift-marketplace/community-operators-dlznx" Oct 05 21:57:03 crc kubenswrapper[4754]: I1005 21:57:03.403920 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dlznx" Oct 05 21:57:03 crc kubenswrapper[4754]: I1005 21:57:03.836659 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-dlznx"] Oct 05 21:57:04 crc kubenswrapper[4754]: I1005 21:57:04.217443 4754 generic.go:334] "Generic (PLEG): container finished" podID="0ac7fab0-0e2a-47db-a7e6-5ab36fb9cf58" containerID="38328d5301cf94ffa5ac5581e2773880da5b773a479df6f2e2a025c3eed9cc4d" exitCode=0 Oct 05 21:57:04 crc kubenswrapper[4754]: I1005 21:57:04.217544 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dlznx" event={"ID":"0ac7fab0-0e2a-47db-a7e6-5ab36fb9cf58","Type":"ContainerDied","Data":"38328d5301cf94ffa5ac5581e2773880da5b773a479df6f2e2a025c3eed9cc4d"} Oct 05 21:57:04 crc kubenswrapper[4754]: I1005 21:57:04.217817 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dlznx" event={"ID":"0ac7fab0-0e2a-47db-a7e6-5ab36fb9cf58","Type":"ContainerStarted","Data":"17e12586db201b838241c3dc2bc4640a3600dcf2c28e4a0b4d9c3e4367842b60"} Oct 05 21:57:05 crc kubenswrapper[4754]: I1005 21:57:05.228053 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dlznx" event={"ID":"0ac7fab0-0e2a-47db-a7e6-5ab36fb9cf58","Type":"ContainerStarted","Data":"72afe7a2a495320bdd860846c948bcbe1bf39b6e2b96c57c7d8758e1387def23"} Oct 05 21:57:06 crc kubenswrapper[4754]: I1005 21:57:06.838076 4754 scope.go:117] "RemoveContainer" containerID="f21169793903f3552e21e22d202f837dbee5ee4d48ced3d4666fe154c7e5b95c" Oct 05 21:57:06 crc kubenswrapper[4754]: E1005 21:57:06.838784 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 21:57:07 crc kubenswrapper[4754]: I1005 21:57:07.247058 4754 generic.go:334] "Generic (PLEG): container finished" podID="0ac7fab0-0e2a-47db-a7e6-5ab36fb9cf58" containerID="72afe7a2a495320bdd860846c948bcbe1bf39b6e2b96c57c7d8758e1387def23" exitCode=0 Oct 05 21:57:07 crc kubenswrapper[4754]: I1005 21:57:07.247109 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dlznx" event={"ID":"0ac7fab0-0e2a-47db-a7e6-5ab36fb9cf58","Type":"ContainerDied","Data":"72afe7a2a495320bdd860846c948bcbe1bf39b6e2b96c57c7d8758e1387def23"} Oct 05 21:57:08 crc kubenswrapper[4754]: I1005 21:57:08.266605 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dlznx" event={"ID":"0ac7fab0-0e2a-47db-a7e6-5ab36fb9cf58","Type":"ContainerStarted","Data":"c8b23b428984f279a44ebff8fb61fc69a8d1927e4f0ab7e6451e57b057d30b6c"} Oct 05 21:57:08 crc kubenswrapper[4754]: I1005 21:57:08.296832 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-dlznx" podStartSLOduration=1.793020185 podStartE2EDuration="5.296815016s" podCreationTimestamp="2025-10-05 21:57:03 +0000 UTC" firstStartedPulling="2025-10-05 21:57:04.219825349 +0000 UTC m=+3748.123944059" lastFinishedPulling="2025-10-05 21:57:07.72362018 +0000 UTC m=+3751.627738890" observedRunningTime="2025-10-05 21:57:08.289012183 +0000 UTC m=+3752.193130893" watchObservedRunningTime="2025-10-05 21:57:08.296815016 +0000 UTC m=+3752.200933726" Oct 05 21:57:13 crc kubenswrapper[4754]: I1005 21:57:13.404627 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-dlznx" Oct 05 21:57:13 crc kubenswrapper[4754]: I1005 21:57:13.406068 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-dlznx" Oct 05 21:57:14 crc kubenswrapper[4754]: I1005 21:57:14.474773 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-dlznx" podUID="0ac7fab0-0e2a-47db-a7e6-5ab36fb9cf58" containerName="registry-server" probeResult="failure" output=< Oct 05 21:57:14 crc kubenswrapper[4754]: timeout: failed to connect service ":50051" within 1s Oct 05 21:57:14 crc kubenswrapper[4754]: > Oct 05 21:57:19 crc kubenswrapper[4754]: I1005 21:57:19.837122 4754 scope.go:117] "RemoveContainer" containerID="f21169793903f3552e21e22d202f837dbee5ee4d48ced3d4666fe154c7e5b95c" Oct 05 21:57:19 crc kubenswrapper[4754]: E1005 21:57:19.838174 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 21:57:23 crc kubenswrapper[4754]: I1005 21:57:23.461345 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-dlznx" Oct 05 21:57:23 crc kubenswrapper[4754]: I1005 21:57:23.510076 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-dlznx" Oct 05 21:57:23 crc kubenswrapper[4754]: I1005 21:57:23.694272 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-dlznx"] Oct 05 21:57:25 crc kubenswrapper[4754]: I1005 21:57:25.433254 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-dlznx" podUID="0ac7fab0-0e2a-47db-a7e6-5ab36fb9cf58" containerName="registry-server" containerID="cri-o://c8b23b428984f279a44ebff8fb61fc69a8d1927e4f0ab7e6451e57b057d30b6c" gracePeriod=2 Oct 05 21:57:26 crc kubenswrapper[4754]: I1005 21:57:26.119081 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dlznx" Oct 05 21:57:26 crc kubenswrapper[4754]: I1005 21:57:26.148613 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ac7fab0-0e2a-47db-a7e6-5ab36fb9cf58-utilities\") pod \"0ac7fab0-0e2a-47db-a7e6-5ab36fb9cf58\" (UID: \"0ac7fab0-0e2a-47db-a7e6-5ab36fb9cf58\") " Oct 05 21:57:26 crc kubenswrapper[4754]: I1005 21:57:26.148713 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ac7fab0-0e2a-47db-a7e6-5ab36fb9cf58-catalog-content\") pod \"0ac7fab0-0e2a-47db-a7e6-5ab36fb9cf58\" (UID: \"0ac7fab0-0e2a-47db-a7e6-5ab36fb9cf58\") " Oct 05 21:57:26 crc kubenswrapper[4754]: I1005 21:57:26.148782 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sr69p\" (UniqueName: \"kubernetes.io/projected/0ac7fab0-0e2a-47db-a7e6-5ab36fb9cf58-kube-api-access-sr69p\") pod \"0ac7fab0-0e2a-47db-a7e6-5ab36fb9cf58\" (UID: \"0ac7fab0-0e2a-47db-a7e6-5ab36fb9cf58\") " Oct 05 21:57:26 crc kubenswrapper[4754]: I1005 21:57:26.149129 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0ac7fab0-0e2a-47db-a7e6-5ab36fb9cf58-utilities" (OuterVolumeSpecName: "utilities") pod "0ac7fab0-0e2a-47db-a7e6-5ab36fb9cf58" (UID: "0ac7fab0-0e2a-47db-a7e6-5ab36fb9cf58"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 21:57:26 crc kubenswrapper[4754]: I1005 21:57:26.149473 4754 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ac7fab0-0e2a-47db-a7e6-5ab36fb9cf58-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 21:57:26 crc kubenswrapper[4754]: I1005 21:57:26.167747 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0ac7fab0-0e2a-47db-a7e6-5ab36fb9cf58-kube-api-access-sr69p" (OuterVolumeSpecName: "kube-api-access-sr69p") pod "0ac7fab0-0e2a-47db-a7e6-5ab36fb9cf58" (UID: "0ac7fab0-0e2a-47db-a7e6-5ab36fb9cf58"). InnerVolumeSpecName "kube-api-access-sr69p". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 21:57:26 crc kubenswrapper[4754]: I1005 21:57:26.212359 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0ac7fab0-0e2a-47db-a7e6-5ab36fb9cf58-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0ac7fab0-0e2a-47db-a7e6-5ab36fb9cf58" (UID: "0ac7fab0-0e2a-47db-a7e6-5ab36fb9cf58"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 21:57:26 crc kubenswrapper[4754]: I1005 21:57:26.251102 4754 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ac7fab0-0e2a-47db-a7e6-5ab36fb9cf58-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 21:57:26 crc kubenswrapper[4754]: I1005 21:57:26.251140 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sr69p\" (UniqueName: \"kubernetes.io/projected/0ac7fab0-0e2a-47db-a7e6-5ab36fb9cf58-kube-api-access-sr69p\") on node \"crc\" DevicePath \"\"" Oct 05 21:57:26 crc kubenswrapper[4754]: I1005 21:57:26.443410 4754 generic.go:334] "Generic (PLEG): container finished" podID="0ac7fab0-0e2a-47db-a7e6-5ab36fb9cf58" containerID="c8b23b428984f279a44ebff8fb61fc69a8d1927e4f0ab7e6451e57b057d30b6c" exitCode=0 Oct 05 21:57:26 crc kubenswrapper[4754]: I1005 21:57:26.443465 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dlznx" Oct 05 21:57:26 crc kubenswrapper[4754]: I1005 21:57:26.443459 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dlznx" event={"ID":"0ac7fab0-0e2a-47db-a7e6-5ab36fb9cf58","Type":"ContainerDied","Data":"c8b23b428984f279a44ebff8fb61fc69a8d1927e4f0ab7e6451e57b057d30b6c"} Oct 05 21:57:26 crc kubenswrapper[4754]: I1005 21:57:26.444674 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dlznx" event={"ID":"0ac7fab0-0e2a-47db-a7e6-5ab36fb9cf58","Type":"ContainerDied","Data":"17e12586db201b838241c3dc2bc4640a3600dcf2c28e4a0b4d9c3e4367842b60"} Oct 05 21:57:26 crc kubenswrapper[4754]: I1005 21:57:26.444771 4754 scope.go:117] "RemoveContainer" containerID="c8b23b428984f279a44ebff8fb61fc69a8d1927e4f0ab7e6451e57b057d30b6c" Oct 05 21:57:26 crc kubenswrapper[4754]: I1005 21:57:26.466978 4754 scope.go:117] "RemoveContainer" containerID="72afe7a2a495320bdd860846c948bcbe1bf39b6e2b96c57c7d8758e1387def23" Oct 05 21:57:26 crc kubenswrapper[4754]: I1005 21:57:26.472633 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-dlznx"] Oct 05 21:57:26 crc kubenswrapper[4754]: I1005 21:57:26.482855 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-dlznx"] Oct 05 21:57:26 crc kubenswrapper[4754]: I1005 21:57:26.498624 4754 scope.go:117] "RemoveContainer" containerID="38328d5301cf94ffa5ac5581e2773880da5b773a479df6f2e2a025c3eed9cc4d" Oct 05 21:57:26 crc kubenswrapper[4754]: I1005 21:57:26.532084 4754 scope.go:117] "RemoveContainer" containerID="c8b23b428984f279a44ebff8fb61fc69a8d1927e4f0ab7e6451e57b057d30b6c" Oct 05 21:57:26 crc kubenswrapper[4754]: E1005 21:57:26.532543 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c8b23b428984f279a44ebff8fb61fc69a8d1927e4f0ab7e6451e57b057d30b6c\": container with ID starting with c8b23b428984f279a44ebff8fb61fc69a8d1927e4f0ab7e6451e57b057d30b6c not found: ID does not exist" containerID="c8b23b428984f279a44ebff8fb61fc69a8d1927e4f0ab7e6451e57b057d30b6c" Oct 05 21:57:26 crc kubenswrapper[4754]: I1005 21:57:26.532575 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c8b23b428984f279a44ebff8fb61fc69a8d1927e4f0ab7e6451e57b057d30b6c"} err="failed to get container status \"c8b23b428984f279a44ebff8fb61fc69a8d1927e4f0ab7e6451e57b057d30b6c\": rpc error: code = NotFound desc = could not find container \"c8b23b428984f279a44ebff8fb61fc69a8d1927e4f0ab7e6451e57b057d30b6c\": container with ID starting with c8b23b428984f279a44ebff8fb61fc69a8d1927e4f0ab7e6451e57b057d30b6c not found: ID does not exist" Oct 05 21:57:26 crc kubenswrapper[4754]: I1005 21:57:26.532597 4754 scope.go:117] "RemoveContainer" containerID="72afe7a2a495320bdd860846c948bcbe1bf39b6e2b96c57c7d8758e1387def23" Oct 05 21:57:26 crc kubenswrapper[4754]: E1005 21:57:26.532858 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"72afe7a2a495320bdd860846c948bcbe1bf39b6e2b96c57c7d8758e1387def23\": container with ID starting with 72afe7a2a495320bdd860846c948bcbe1bf39b6e2b96c57c7d8758e1387def23 not found: ID does not exist" containerID="72afe7a2a495320bdd860846c948bcbe1bf39b6e2b96c57c7d8758e1387def23" Oct 05 21:57:26 crc kubenswrapper[4754]: I1005 21:57:26.532895 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"72afe7a2a495320bdd860846c948bcbe1bf39b6e2b96c57c7d8758e1387def23"} err="failed to get container status \"72afe7a2a495320bdd860846c948bcbe1bf39b6e2b96c57c7d8758e1387def23\": rpc error: code = NotFound desc = could not find container \"72afe7a2a495320bdd860846c948bcbe1bf39b6e2b96c57c7d8758e1387def23\": container with ID starting with 72afe7a2a495320bdd860846c948bcbe1bf39b6e2b96c57c7d8758e1387def23 not found: ID does not exist" Oct 05 21:57:26 crc kubenswrapper[4754]: I1005 21:57:26.532917 4754 scope.go:117] "RemoveContainer" containerID="38328d5301cf94ffa5ac5581e2773880da5b773a479df6f2e2a025c3eed9cc4d" Oct 05 21:57:26 crc kubenswrapper[4754]: E1005 21:57:26.533191 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"38328d5301cf94ffa5ac5581e2773880da5b773a479df6f2e2a025c3eed9cc4d\": container with ID starting with 38328d5301cf94ffa5ac5581e2773880da5b773a479df6f2e2a025c3eed9cc4d not found: ID does not exist" containerID="38328d5301cf94ffa5ac5581e2773880da5b773a479df6f2e2a025c3eed9cc4d" Oct 05 21:57:26 crc kubenswrapper[4754]: I1005 21:57:26.533216 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"38328d5301cf94ffa5ac5581e2773880da5b773a479df6f2e2a025c3eed9cc4d"} err="failed to get container status \"38328d5301cf94ffa5ac5581e2773880da5b773a479df6f2e2a025c3eed9cc4d\": rpc error: code = NotFound desc = could not find container \"38328d5301cf94ffa5ac5581e2773880da5b773a479df6f2e2a025c3eed9cc4d\": container with ID starting with 38328d5301cf94ffa5ac5581e2773880da5b773a479df6f2e2a025c3eed9cc4d not found: ID does not exist" Oct 05 21:57:26 crc kubenswrapper[4754]: I1005 21:57:26.852144 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0ac7fab0-0e2a-47db-a7e6-5ab36fb9cf58" path="/var/lib/kubelet/pods/0ac7fab0-0e2a-47db-a7e6-5ab36fb9cf58/volumes" Oct 05 21:57:34 crc kubenswrapper[4754]: I1005 21:57:34.837659 4754 scope.go:117] "RemoveContainer" containerID="f21169793903f3552e21e22d202f837dbee5ee4d48ced3d4666fe154c7e5b95c" Oct 05 21:57:34 crc kubenswrapper[4754]: E1005 21:57:34.839533 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 21:57:48 crc kubenswrapper[4754]: I1005 21:57:48.837430 4754 scope.go:117] "RemoveContainer" containerID="f21169793903f3552e21e22d202f837dbee5ee4d48ced3d4666fe154c7e5b95c" Oct 05 21:57:48 crc kubenswrapper[4754]: E1005 21:57:48.839060 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 21:58:02 crc kubenswrapper[4754]: I1005 21:58:02.838123 4754 scope.go:117] "RemoveContainer" containerID="f21169793903f3552e21e22d202f837dbee5ee4d48ced3d4666fe154c7e5b95c" Oct 05 21:58:02 crc kubenswrapper[4754]: E1005 21:58:02.839297 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 21:58:15 crc kubenswrapper[4754]: I1005 21:58:15.837730 4754 scope.go:117] "RemoveContainer" containerID="f21169793903f3552e21e22d202f837dbee5ee4d48ced3d4666fe154c7e5b95c" Oct 05 21:58:15 crc kubenswrapper[4754]: E1005 21:58:15.838483 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 21:58:26 crc kubenswrapper[4754]: I1005 21:58:26.844482 4754 scope.go:117] "RemoveContainer" containerID="f21169793903f3552e21e22d202f837dbee5ee4d48ced3d4666fe154c7e5b95c" Oct 05 21:58:26 crc kubenswrapper[4754]: E1005 21:58:26.845554 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 21:58:40 crc kubenswrapper[4754]: I1005 21:58:40.838004 4754 scope.go:117] "RemoveContainer" containerID="f21169793903f3552e21e22d202f837dbee5ee4d48ced3d4666fe154c7e5b95c" Oct 05 21:58:40 crc kubenswrapper[4754]: E1005 21:58:40.839018 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 21:58:51 crc kubenswrapper[4754]: I1005 21:58:51.837774 4754 scope.go:117] "RemoveContainer" containerID="f21169793903f3552e21e22d202f837dbee5ee4d48ced3d4666fe154c7e5b95c" Oct 05 21:58:51 crc kubenswrapper[4754]: E1005 21:58:51.838936 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 21:59:03 crc kubenswrapper[4754]: I1005 21:59:03.837458 4754 scope.go:117] "RemoveContainer" containerID="f21169793903f3552e21e22d202f837dbee5ee4d48ced3d4666fe154c7e5b95c" Oct 05 21:59:03 crc kubenswrapper[4754]: E1005 21:59:03.838347 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 21:59:15 crc kubenswrapper[4754]: I1005 21:59:15.837193 4754 scope.go:117] "RemoveContainer" containerID="f21169793903f3552e21e22d202f837dbee5ee4d48ced3d4666fe154c7e5b95c" Oct 05 21:59:15 crc kubenswrapper[4754]: E1005 21:59:15.837900 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 21:59:27 crc kubenswrapper[4754]: I1005 21:59:27.837187 4754 scope.go:117] "RemoveContainer" containerID="f21169793903f3552e21e22d202f837dbee5ee4d48ced3d4666fe154c7e5b95c" Oct 05 21:59:27 crc kubenswrapper[4754]: E1005 21:59:27.837824 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 21:59:38 crc kubenswrapper[4754]: I1005 21:59:38.837590 4754 scope.go:117] "RemoveContainer" containerID="f21169793903f3552e21e22d202f837dbee5ee4d48ced3d4666fe154c7e5b95c" Oct 05 21:59:38 crc kubenswrapper[4754]: E1005 21:59:38.838403 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 21:59:51 crc kubenswrapper[4754]: I1005 21:59:51.837565 4754 scope.go:117] "RemoveContainer" containerID="f21169793903f3552e21e22d202f837dbee5ee4d48ced3d4666fe154c7e5b95c" Oct 05 21:59:51 crc kubenswrapper[4754]: E1005 21:59:51.838238 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 22:00:00 crc kubenswrapper[4754]: I1005 22:00:00.189464 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29328360-hvgp4"] Oct 05 22:00:00 crc kubenswrapper[4754]: E1005 22:00:00.190305 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ac7fab0-0e2a-47db-a7e6-5ab36fb9cf58" containerName="extract-utilities" Oct 05 22:00:00 crc kubenswrapper[4754]: I1005 22:00:00.190318 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ac7fab0-0e2a-47db-a7e6-5ab36fb9cf58" containerName="extract-utilities" Oct 05 22:00:00 crc kubenswrapper[4754]: E1005 22:00:00.190325 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ac7fab0-0e2a-47db-a7e6-5ab36fb9cf58" containerName="registry-server" Oct 05 22:00:00 crc kubenswrapper[4754]: I1005 22:00:00.190331 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ac7fab0-0e2a-47db-a7e6-5ab36fb9cf58" containerName="registry-server" Oct 05 22:00:00 crc kubenswrapper[4754]: E1005 22:00:00.190352 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ac7fab0-0e2a-47db-a7e6-5ab36fb9cf58" containerName="extract-content" Oct 05 22:00:00 crc kubenswrapper[4754]: I1005 22:00:00.190359 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ac7fab0-0e2a-47db-a7e6-5ab36fb9cf58" containerName="extract-content" Oct 05 22:00:00 crc kubenswrapper[4754]: I1005 22:00:00.190546 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="0ac7fab0-0e2a-47db-a7e6-5ab36fb9cf58" containerName="registry-server" Oct 05 22:00:00 crc kubenswrapper[4754]: I1005 22:00:00.191127 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29328360-hvgp4" Oct 05 22:00:00 crc kubenswrapper[4754]: I1005 22:00:00.203393 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 05 22:00:00 crc kubenswrapper[4754]: I1005 22:00:00.204354 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 05 22:00:00 crc kubenswrapper[4754]: I1005 22:00:00.233444 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29328360-hvgp4"] Oct 05 22:00:00 crc kubenswrapper[4754]: I1005 22:00:00.272261 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2078f658-e9b1-40a2-a023-3afa2bff23cd-config-volume\") pod \"collect-profiles-29328360-hvgp4\" (UID: \"2078f658-e9b1-40a2-a023-3afa2bff23cd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29328360-hvgp4" Oct 05 22:00:00 crc kubenswrapper[4754]: I1005 22:00:00.272307 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nqf8w\" (UniqueName: \"kubernetes.io/projected/2078f658-e9b1-40a2-a023-3afa2bff23cd-kube-api-access-nqf8w\") pod \"collect-profiles-29328360-hvgp4\" (UID: \"2078f658-e9b1-40a2-a023-3afa2bff23cd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29328360-hvgp4" Oct 05 22:00:00 crc kubenswrapper[4754]: I1005 22:00:00.272330 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2078f658-e9b1-40a2-a023-3afa2bff23cd-secret-volume\") pod \"collect-profiles-29328360-hvgp4\" (UID: \"2078f658-e9b1-40a2-a023-3afa2bff23cd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29328360-hvgp4" Oct 05 22:00:00 crc kubenswrapper[4754]: I1005 22:00:00.374564 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2078f658-e9b1-40a2-a023-3afa2bff23cd-config-volume\") pod \"collect-profiles-29328360-hvgp4\" (UID: \"2078f658-e9b1-40a2-a023-3afa2bff23cd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29328360-hvgp4" Oct 05 22:00:00 crc kubenswrapper[4754]: I1005 22:00:00.374609 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nqf8w\" (UniqueName: \"kubernetes.io/projected/2078f658-e9b1-40a2-a023-3afa2bff23cd-kube-api-access-nqf8w\") pod \"collect-profiles-29328360-hvgp4\" (UID: \"2078f658-e9b1-40a2-a023-3afa2bff23cd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29328360-hvgp4" Oct 05 22:00:00 crc kubenswrapper[4754]: I1005 22:00:00.374641 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2078f658-e9b1-40a2-a023-3afa2bff23cd-secret-volume\") pod \"collect-profiles-29328360-hvgp4\" (UID: \"2078f658-e9b1-40a2-a023-3afa2bff23cd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29328360-hvgp4" Oct 05 22:00:00 crc kubenswrapper[4754]: I1005 22:00:00.375719 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2078f658-e9b1-40a2-a023-3afa2bff23cd-config-volume\") pod \"collect-profiles-29328360-hvgp4\" (UID: \"2078f658-e9b1-40a2-a023-3afa2bff23cd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29328360-hvgp4" Oct 05 22:00:00 crc kubenswrapper[4754]: I1005 22:00:00.394737 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2078f658-e9b1-40a2-a023-3afa2bff23cd-secret-volume\") pod \"collect-profiles-29328360-hvgp4\" (UID: \"2078f658-e9b1-40a2-a023-3afa2bff23cd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29328360-hvgp4" Oct 05 22:00:00 crc kubenswrapper[4754]: I1005 22:00:00.397044 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nqf8w\" (UniqueName: \"kubernetes.io/projected/2078f658-e9b1-40a2-a023-3afa2bff23cd-kube-api-access-nqf8w\") pod \"collect-profiles-29328360-hvgp4\" (UID: \"2078f658-e9b1-40a2-a023-3afa2bff23cd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29328360-hvgp4" Oct 05 22:00:00 crc kubenswrapper[4754]: I1005 22:00:00.512573 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29328360-hvgp4" Oct 05 22:00:01 crc kubenswrapper[4754]: I1005 22:00:01.523587 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29328360-hvgp4"] Oct 05 22:00:01 crc kubenswrapper[4754]: I1005 22:00:01.994835 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29328360-hvgp4" event={"ID":"2078f658-e9b1-40a2-a023-3afa2bff23cd","Type":"ContainerStarted","Data":"341430e0e94c6bb9650f62a6e6a974245ee847523650e2d237cf77de6350618c"} Oct 05 22:00:01 crc kubenswrapper[4754]: I1005 22:00:01.995116 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29328360-hvgp4" event={"ID":"2078f658-e9b1-40a2-a023-3afa2bff23cd","Type":"ContainerStarted","Data":"aca53620c688d3c48d38f8e08f5c1bf1162ba41032f7ba766a2eff7004b48a81"} Oct 05 22:00:03 crc kubenswrapper[4754]: I1005 22:00:03.004838 4754 generic.go:334] "Generic (PLEG): container finished" podID="2078f658-e9b1-40a2-a023-3afa2bff23cd" containerID="341430e0e94c6bb9650f62a6e6a974245ee847523650e2d237cf77de6350618c" exitCode=0 Oct 05 22:00:03 crc kubenswrapper[4754]: I1005 22:00:03.004891 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29328360-hvgp4" event={"ID":"2078f658-e9b1-40a2-a023-3afa2bff23cd","Type":"ContainerDied","Data":"341430e0e94c6bb9650f62a6e6a974245ee847523650e2d237cf77de6350618c"} Oct 05 22:00:04 crc kubenswrapper[4754]: I1005 22:00:04.457096 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29328360-hvgp4" Oct 05 22:00:04 crc kubenswrapper[4754]: I1005 22:00:04.559051 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2078f658-e9b1-40a2-a023-3afa2bff23cd-secret-volume\") pod \"2078f658-e9b1-40a2-a023-3afa2bff23cd\" (UID: \"2078f658-e9b1-40a2-a023-3afa2bff23cd\") " Oct 05 22:00:04 crc kubenswrapper[4754]: I1005 22:00:04.559355 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nqf8w\" (UniqueName: \"kubernetes.io/projected/2078f658-e9b1-40a2-a023-3afa2bff23cd-kube-api-access-nqf8w\") pod \"2078f658-e9b1-40a2-a023-3afa2bff23cd\" (UID: \"2078f658-e9b1-40a2-a023-3afa2bff23cd\") " Oct 05 22:00:04 crc kubenswrapper[4754]: I1005 22:00:04.559530 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2078f658-e9b1-40a2-a023-3afa2bff23cd-config-volume\") pod \"2078f658-e9b1-40a2-a023-3afa2bff23cd\" (UID: \"2078f658-e9b1-40a2-a023-3afa2bff23cd\") " Oct 05 22:00:04 crc kubenswrapper[4754]: I1005 22:00:04.560516 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2078f658-e9b1-40a2-a023-3afa2bff23cd-config-volume" (OuterVolumeSpecName: "config-volume") pod "2078f658-e9b1-40a2-a023-3afa2bff23cd" (UID: "2078f658-e9b1-40a2-a023-3afa2bff23cd"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 22:00:04 crc kubenswrapper[4754]: I1005 22:00:04.577820 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2078f658-e9b1-40a2-a023-3afa2bff23cd-kube-api-access-nqf8w" (OuterVolumeSpecName: "kube-api-access-nqf8w") pod "2078f658-e9b1-40a2-a023-3afa2bff23cd" (UID: "2078f658-e9b1-40a2-a023-3afa2bff23cd"). InnerVolumeSpecName "kube-api-access-nqf8w". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 22:00:04 crc kubenswrapper[4754]: I1005 22:00:04.581631 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2078f658-e9b1-40a2-a023-3afa2bff23cd-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "2078f658-e9b1-40a2-a023-3afa2bff23cd" (UID: "2078f658-e9b1-40a2-a023-3afa2bff23cd"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 22:00:04 crc kubenswrapper[4754]: I1005 22:00:04.607033 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29328315-mwzdd"] Oct 05 22:00:04 crc kubenswrapper[4754]: I1005 22:00:04.614113 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29328315-mwzdd"] Oct 05 22:00:04 crc kubenswrapper[4754]: I1005 22:00:04.661363 4754 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2078f658-e9b1-40a2-a023-3afa2bff23cd-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 05 22:00:04 crc kubenswrapper[4754]: I1005 22:00:04.661395 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nqf8w\" (UniqueName: \"kubernetes.io/projected/2078f658-e9b1-40a2-a023-3afa2bff23cd-kube-api-access-nqf8w\") on node \"crc\" DevicePath \"\"" Oct 05 22:00:04 crc kubenswrapper[4754]: I1005 22:00:04.661405 4754 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2078f658-e9b1-40a2-a023-3afa2bff23cd-config-volume\") on node \"crc\" DevicePath \"\"" Oct 05 22:00:04 crc kubenswrapper[4754]: I1005 22:00:04.855949 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="545fecb5-deec-43b9-9aac-303579f1e9c5" path="/var/lib/kubelet/pods/545fecb5-deec-43b9-9aac-303579f1e9c5/volumes" Oct 05 22:00:05 crc kubenswrapper[4754]: I1005 22:00:05.034332 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29328360-hvgp4" event={"ID":"2078f658-e9b1-40a2-a023-3afa2bff23cd","Type":"ContainerDied","Data":"aca53620c688d3c48d38f8e08f5c1bf1162ba41032f7ba766a2eff7004b48a81"} Oct 05 22:00:05 crc kubenswrapper[4754]: I1005 22:00:05.034817 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aca53620c688d3c48d38f8e08f5c1bf1162ba41032f7ba766a2eff7004b48a81" Oct 05 22:00:05 crc kubenswrapper[4754]: I1005 22:00:05.034630 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29328360-hvgp4" Oct 05 22:00:06 crc kubenswrapper[4754]: I1005 22:00:06.849928 4754 scope.go:117] "RemoveContainer" containerID="f21169793903f3552e21e22d202f837dbee5ee4d48ced3d4666fe154c7e5b95c" Oct 05 22:00:08 crc kubenswrapper[4754]: I1005 22:00:08.071988 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" event={"ID":"be95a413-6a59-45b4-84b7-b43cae694a26","Type":"ContainerStarted","Data":"7e8376fda04f81f5bc9127199e823c84456ce7d86c14da7ebaece2d9e07200e8"} Oct 05 22:00:29 crc kubenswrapper[4754]: I1005 22:00:29.788207 4754 scope.go:117] "RemoveContainer" containerID="fe2d1561f99f4ac4b0e2d046b6d8439f2b155c3f309f47b33d7b20f2ed80b389" Oct 05 22:01:00 crc kubenswrapper[4754]: I1005 22:01:00.146321 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29328361-wb6sz"] Oct 05 22:01:00 crc kubenswrapper[4754]: E1005 22:01:00.147382 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2078f658-e9b1-40a2-a023-3afa2bff23cd" containerName="collect-profiles" Oct 05 22:01:00 crc kubenswrapper[4754]: I1005 22:01:00.147397 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="2078f658-e9b1-40a2-a023-3afa2bff23cd" containerName="collect-profiles" Oct 05 22:01:00 crc kubenswrapper[4754]: I1005 22:01:00.147650 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="2078f658-e9b1-40a2-a023-3afa2bff23cd" containerName="collect-profiles" Oct 05 22:01:00 crc kubenswrapper[4754]: I1005 22:01:00.148306 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29328361-wb6sz" Oct 05 22:01:00 crc kubenswrapper[4754]: I1005 22:01:00.161647 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29328361-wb6sz"] Oct 05 22:01:00 crc kubenswrapper[4754]: I1005 22:01:00.276909 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc91c348-ca21-48c0-a694-f9850277e59b-config-data\") pod \"keystone-cron-29328361-wb6sz\" (UID: \"fc91c348-ca21-48c0-a694-f9850277e59b\") " pod="openstack/keystone-cron-29328361-wb6sz" Oct 05 22:01:00 crc kubenswrapper[4754]: I1005 22:01:00.277208 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/fc91c348-ca21-48c0-a694-f9850277e59b-fernet-keys\") pod \"keystone-cron-29328361-wb6sz\" (UID: \"fc91c348-ca21-48c0-a694-f9850277e59b\") " pod="openstack/keystone-cron-29328361-wb6sz" Oct 05 22:01:00 crc kubenswrapper[4754]: I1005 22:01:00.277240 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zrsxj\" (UniqueName: \"kubernetes.io/projected/fc91c348-ca21-48c0-a694-f9850277e59b-kube-api-access-zrsxj\") pod \"keystone-cron-29328361-wb6sz\" (UID: \"fc91c348-ca21-48c0-a694-f9850277e59b\") " pod="openstack/keystone-cron-29328361-wb6sz" Oct 05 22:01:00 crc kubenswrapper[4754]: I1005 22:01:00.277585 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc91c348-ca21-48c0-a694-f9850277e59b-combined-ca-bundle\") pod \"keystone-cron-29328361-wb6sz\" (UID: \"fc91c348-ca21-48c0-a694-f9850277e59b\") " pod="openstack/keystone-cron-29328361-wb6sz" Oct 05 22:01:00 crc kubenswrapper[4754]: I1005 22:01:00.380102 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zrsxj\" (UniqueName: \"kubernetes.io/projected/fc91c348-ca21-48c0-a694-f9850277e59b-kube-api-access-zrsxj\") pod \"keystone-cron-29328361-wb6sz\" (UID: \"fc91c348-ca21-48c0-a694-f9850277e59b\") " pod="openstack/keystone-cron-29328361-wb6sz" Oct 05 22:01:00 crc kubenswrapper[4754]: I1005 22:01:00.380360 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc91c348-ca21-48c0-a694-f9850277e59b-combined-ca-bundle\") pod \"keystone-cron-29328361-wb6sz\" (UID: \"fc91c348-ca21-48c0-a694-f9850277e59b\") " pod="openstack/keystone-cron-29328361-wb6sz" Oct 05 22:01:00 crc kubenswrapper[4754]: I1005 22:01:00.380452 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc91c348-ca21-48c0-a694-f9850277e59b-config-data\") pod \"keystone-cron-29328361-wb6sz\" (UID: \"fc91c348-ca21-48c0-a694-f9850277e59b\") " pod="openstack/keystone-cron-29328361-wb6sz" Oct 05 22:01:00 crc kubenswrapper[4754]: I1005 22:01:00.380574 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/fc91c348-ca21-48c0-a694-f9850277e59b-fernet-keys\") pod \"keystone-cron-29328361-wb6sz\" (UID: \"fc91c348-ca21-48c0-a694-f9850277e59b\") " pod="openstack/keystone-cron-29328361-wb6sz" Oct 05 22:01:00 crc kubenswrapper[4754]: I1005 22:01:00.387223 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc91c348-ca21-48c0-a694-f9850277e59b-combined-ca-bundle\") pod \"keystone-cron-29328361-wb6sz\" (UID: \"fc91c348-ca21-48c0-a694-f9850277e59b\") " pod="openstack/keystone-cron-29328361-wb6sz" Oct 05 22:01:00 crc kubenswrapper[4754]: I1005 22:01:00.389235 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/fc91c348-ca21-48c0-a694-f9850277e59b-fernet-keys\") pod \"keystone-cron-29328361-wb6sz\" (UID: \"fc91c348-ca21-48c0-a694-f9850277e59b\") " pod="openstack/keystone-cron-29328361-wb6sz" Oct 05 22:01:00 crc kubenswrapper[4754]: I1005 22:01:00.397779 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc91c348-ca21-48c0-a694-f9850277e59b-config-data\") pod \"keystone-cron-29328361-wb6sz\" (UID: \"fc91c348-ca21-48c0-a694-f9850277e59b\") " pod="openstack/keystone-cron-29328361-wb6sz" Oct 05 22:01:00 crc kubenswrapper[4754]: I1005 22:01:00.410690 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zrsxj\" (UniqueName: \"kubernetes.io/projected/fc91c348-ca21-48c0-a694-f9850277e59b-kube-api-access-zrsxj\") pod \"keystone-cron-29328361-wb6sz\" (UID: \"fc91c348-ca21-48c0-a694-f9850277e59b\") " pod="openstack/keystone-cron-29328361-wb6sz" Oct 05 22:01:00 crc kubenswrapper[4754]: I1005 22:01:00.466113 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29328361-wb6sz" Oct 05 22:01:00 crc kubenswrapper[4754]: I1005 22:01:00.976125 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29328361-wb6sz"] Oct 05 22:01:01 crc kubenswrapper[4754]: I1005 22:01:01.612043 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29328361-wb6sz" event={"ID":"fc91c348-ca21-48c0-a694-f9850277e59b","Type":"ContainerStarted","Data":"65d7438436282381ddfb453e45dbc5c6c0aced7bcd600701dc772f546f7529b4"} Oct 05 22:01:01 crc kubenswrapper[4754]: I1005 22:01:01.614256 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29328361-wb6sz" event={"ID":"fc91c348-ca21-48c0-a694-f9850277e59b","Type":"ContainerStarted","Data":"dca302552ad1499291c27d1d4fd08c4aea1ca438c63ba18cba022e16ce377c8a"} Oct 05 22:01:01 crc kubenswrapper[4754]: I1005 22:01:01.631012 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29328361-wb6sz" podStartSLOduration=1.6309872909999998 podStartE2EDuration="1.630987291s" podCreationTimestamp="2025-10-05 22:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 22:01:01.628288221 +0000 UTC m=+3985.532406961" watchObservedRunningTime="2025-10-05 22:01:01.630987291 +0000 UTC m=+3985.535106031" Oct 05 22:01:04 crc kubenswrapper[4754]: I1005 22:01:04.646137 4754 generic.go:334] "Generic (PLEG): container finished" podID="fc91c348-ca21-48c0-a694-f9850277e59b" containerID="65d7438436282381ddfb453e45dbc5c6c0aced7bcd600701dc772f546f7529b4" exitCode=0 Oct 05 22:01:04 crc kubenswrapper[4754]: I1005 22:01:04.646231 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29328361-wb6sz" event={"ID":"fc91c348-ca21-48c0-a694-f9850277e59b","Type":"ContainerDied","Data":"65d7438436282381ddfb453e45dbc5c6c0aced7bcd600701dc772f546f7529b4"} Oct 05 22:01:06 crc kubenswrapper[4754]: I1005 22:01:06.187038 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29328361-wb6sz" Oct 05 22:01:06 crc kubenswrapper[4754]: I1005 22:01:06.199703 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc91c348-ca21-48c0-a694-f9850277e59b-combined-ca-bundle\") pod \"fc91c348-ca21-48c0-a694-f9850277e59b\" (UID: \"fc91c348-ca21-48c0-a694-f9850277e59b\") " Oct 05 22:01:06 crc kubenswrapper[4754]: I1005 22:01:06.199766 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc91c348-ca21-48c0-a694-f9850277e59b-config-data\") pod \"fc91c348-ca21-48c0-a694-f9850277e59b\" (UID: \"fc91c348-ca21-48c0-a694-f9850277e59b\") " Oct 05 22:01:06 crc kubenswrapper[4754]: I1005 22:01:06.199798 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zrsxj\" (UniqueName: \"kubernetes.io/projected/fc91c348-ca21-48c0-a694-f9850277e59b-kube-api-access-zrsxj\") pod \"fc91c348-ca21-48c0-a694-f9850277e59b\" (UID: \"fc91c348-ca21-48c0-a694-f9850277e59b\") " Oct 05 22:01:06 crc kubenswrapper[4754]: I1005 22:01:06.200102 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/fc91c348-ca21-48c0-a694-f9850277e59b-fernet-keys\") pod \"fc91c348-ca21-48c0-a694-f9850277e59b\" (UID: \"fc91c348-ca21-48c0-a694-f9850277e59b\") " Oct 05 22:01:06 crc kubenswrapper[4754]: I1005 22:01:06.215634 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc91c348-ca21-48c0-a694-f9850277e59b-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "fc91c348-ca21-48c0-a694-f9850277e59b" (UID: "fc91c348-ca21-48c0-a694-f9850277e59b"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 22:01:06 crc kubenswrapper[4754]: I1005 22:01:06.238103 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fc91c348-ca21-48c0-a694-f9850277e59b-kube-api-access-zrsxj" (OuterVolumeSpecName: "kube-api-access-zrsxj") pod "fc91c348-ca21-48c0-a694-f9850277e59b" (UID: "fc91c348-ca21-48c0-a694-f9850277e59b"). InnerVolumeSpecName "kube-api-access-zrsxj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 22:01:06 crc kubenswrapper[4754]: I1005 22:01:06.248441 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc91c348-ca21-48c0-a694-f9850277e59b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fc91c348-ca21-48c0-a694-f9850277e59b" (UID: "fc91c348-ca21-48c0-a694-f9850277e59b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 22:01:06 crc kubenswrapper[4754]: I1005 22:01:06.302755 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc91c348-ca21-48c0-a694-f9850277e59b-config-data" (OuterVolumeSpecName: "config-data") pod "fc91c348-ca21-48c0-a694-f9850277e59b" (UID: "fc91c348-ca21-48c0-a694-f9850277e59b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 22:01:06 crc kubenswrapper[4754]: I1005 22:01:06.302950 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc91c348-ca21-48c0-a694-f9850277e59b-config-data\") pod \"fc91c348-ca21-48c0-a694-f9850277e59b\" (UID: \"fc91c348-ca21-48c0-a694-f9850277e59b\") " Oct 05 22:01:06 crc kubenswrapper[4754]: I1005 22:01:06.303488 4754 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc91c348-ca21-48c0-a694-f9850277e59b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 22:01:06 crc kubenswrapper[4754]: I1005 22:01:06.303520 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zrsxj\" (UniqueName: \"kubernetes.io/projected/fc91c348-ca21-48c0-a694-f9850277e59b-kube-api-access-zrsxj\") on node \"crc\" DevicePath \"\"" Oct 05 22:01:06 crc kubenswrapper[4754]: I1005 22:01:06.303532 4754 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/fc91c348-ca21-48c0-a694-f9850277e59b-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 05 22:01:06 crc kubenswrapper[4754]: W1005 22:01:06.304747 4754 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/fc91c348-ca21-48c0-a694-f9850277e59b/volumes/kubernetes.io~secret/config-data Oct 05 22:01:06 crc kubenswrapper[4754]: I1005 22:01:06.305256 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc91c348-ca21-48c0-a694-f9850277e59b-config-data" (OuterVolumeSpecName: "config-data") pod "fc91c348-ca21-48c0-a694-f9850277e59b" (UID: "fc91c348-ca21-48c0-a694-f9850277e59b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 22:01:06 crc kubenswrapper[4754]: I1005 22:01:06.405879 4754 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc91c348-ca21-48c0-a694-f9850277e59b-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 22:01:06 crc kubenswrapper[4754]: I1005 22:01:06.663255 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29328361-wb6sz" event={"ID":"fc91c348-ca21-48c0-a694-f9850277e59b","Type":"ContainerDied","Data":"dca302552ad1499291c27d1d4fd08c4aea1ca438c63ba18cba022e16ce377c8a"} Oct 05 22:01:06 crc kubenswrapper[4754]: I1005 22:01:06.663295 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dca302552ad1499291c27d1d4fd08c4aea1ca438c63ba18cba022e16ce377c8a" Oct 05 22:01:06 crc kubenswrapper[4754]: I1005 22:01:06.663330 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29328361-wb6sz" Oct 05 22:02:35 crc kubenswrapper[4754]: I1005 22:02:35.245101 4754 patch_prober.go:28] interesting pod/machine-config-daemon-b2h9k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 22:02:35 crc kubenswrapper[4754]: I1005 22:02:35.245574 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 22:03:05 crc kubenswrapper[4754]: I1005 22:03:05.244727 4754 patch_prober.go:28] interesting pod/machine-config-daemon-b2h9k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 22:03:05 crc kubenswrapper[4754]: I1005 22:03:05.245346 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 22:03:35 crc kubenswrapper[4754]: I1005 22:03:35.244929 4754 patch_prober.go:28] interesting pod/machine-config-daemon-b2h9k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 22:03:35 crc kubenswrapper[4754]: I1005 22:03:35.245464 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 22:03:35 crc kubenswrapper[4754]: I1005 22:03:35.245528 4754 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" Oct 05 22:03:35 crc kubenswrapper[4754]: I1005 22:03:35.246324 4754 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7e8376fda04f81f5bc9127199e823c84456ce7d86c14da7ebaece2d9e07200e8"} pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 05 22:03:35 crc kubenswrapper[4754]: I1005 22:03:35.246402 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" containerName="machine-config-daemon" containerID="cri-o://7e8376fda04f81f5bc9127199e823c84456ce7d86c14da7ebaece2d9e07200e8" gracePeriod=600 Oct 05 22:03:36 crc kubenswrapper[4754]: I1005 22:03:36.106338 4754 generic.go:334] "Generic (PLEG): container finished" podID="be95a413-6a59-45b4-84b7-b43cae694a26" containerID="7e8376fda04f81f5bc9127199e823c84456ce7d86c14da7ebaece2d9e07200e8" exitCode=0 Oct 05 22:03:36 crc kubenswrapper[4754]: I1005 22:03:36.106406 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" event={"ID":"be95a413-6a59-45b4-84b7-b43cae694a26","Type":"ContainerDied","Data":"7e8376fda04f81f5bc9127199e823c84456ce7d86c14da7ebaece2d9e07200e8"} Oct 05 22:03:36 crc kubenswrapper[4754]: I1005 22:03:36.106912 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" event={"ID":"be95a413-6a59-45b4-84b7-b43cae694a26","Type":"ContainerStarted","Data":"d925cdc080b01050e7b9377dba2f144d30f517d60778d239b53617b071498b40"} Oct 05 22:03:36 crc kubenswrapper[4754]: I1005 22:03:36.106939 4754 scope.go:117] "RemoveContainer" containerID="f21169793903f3552e21e22d202f837dbee5ee4d48ced3d4666fe154c7e5b95c" Oct 05 22:04:37 crc kubenswrapper[4754]: I1005 22:04:37.641762 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-2d2mr"] Oct 05 22:04:37 crc kubenswrapper[4754]: E1005 22:04:37.662484 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc91c348-ca21-48c0-a694-f9850277e59b" containerName="keystone-cron" Oct 05 22:04:37 crc kubenswrapper[4754]: I1005 22:04:37.662744 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc91c348-ca21-48c0-a694-f9850277e59b" containerName="keystone-cron" Oct 05 22:04:37 crc kubenswrapper[4754]: I1005 22:04:37.663303 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="fc91c348-ca21-48c0-a694-f9850277e59b" containerName="keystone-cron" Oct 05 22:04:37 crc kubenswrapper[4754]: I1005 22:04:37.666152 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2d2mr" Oct 05 22:04:37 crc kubenswrapper[4754]: I1005 22:04:37.668008 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-2d2mr"] Oct 05 22:04:37 crc kubenswrapper[4754]: I1005 22:04:37.707995 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p7grk\" (UniqueName: \"kubernetes.io/projected/dbf8aa85-c070-4f04-b788-b60dd6b8cc9c-kube-api-access-p7grk\") pod \"redhat-operators-2d2mr\" (UID: \"dbf8aa85-c070-4f04-b788-b60dd6b8cc9c\") " pod="openshift-marketplace/redhat-operators-2d2mr" Oct 05 22:04:37 crc kubenswrapper[4754]: I1005 22:04:37.708199 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dbf8aa85-c070-4f04-b788-b60dd6b8cc9c-utilities\") pod \"redhat-operators-2d2mr\" (UID: \"dbf8aa85-c070-4f04-b788-b60dd6b8cc9c\") " pod="openshift-marketplace/redhat-operators-2d2mr" Oct 05 22:04:37 crc kubenswrapper[4754]: I1005 22:04:37.708315 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dbf8aa85-c070-4f04-b788-b60dd6b8cc9c-catalog-content\") pod \"redhat-operators-2d2mr\" (UID: \"dbf8aa85-c070-4f04-b788-b60dd6b8cc9c\") " pod="openshift-marketplace/redhat-operators-2d2mr" Oct 05 22:04:37 crc kubenswrapper[4754]: I1005 22:04:37.810470 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p7grk\" (UniqueName: \"kubernetes.io/projected/dbf8aa85-c070-4f04-b788-b60dd6b8cc9c-kube-api-access-p7grk\") pod \"redhat-operators-2d2mr\" (UID: \"dbf8aa85-c070-4f04-b788-b60dd6b8cc9c\") " pod="openshift-marketplace/redhat-operators-2d2mr" Oct 05 22:04:37 crc kubenswrapper[4754]: I1005 22:04:37.810523 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dbf8aa85-c070-4f04-b788-b60dd6b8cc9c-utilities\") pod \"redhat-operators-2d2mr\" (UID: \"dbf8aa85-c070-4f04-b788-b60dd6b8cc9c\") " pod="openshift-marketplace/redhat-operators-2d2mr" Oct 05 22:04:37 crc kubenswrapper[4754]: I1005 22:04:37.810568 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dbf8aa85-c070-4f04-b788-b60dd6b8cc9c-catalog-content\") pod \"redhat-operators-2d2mr\" (UID: \"dbf8aa85-c070-4f04-b788-b60dd6b8cc9c\") " pod="openshift-marketplace/redhat-operators-2d2mr" Oct 05 22:04:37 crc kubenswrapper[4754]: I1005 22:04:37.813074 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dbf8aa85-c070-4f04-b788-b60dd6b8cc9c-catalog-content\") pod \"redhat-operators-2d2mr\" (UID: \"dbf8aa85-c070-4f04-b788-b60dd6b8cc9c\") " pod="openshift-marketplace/redhat-operators-2d2mr" Oct 05 22:04:37 crc kubenswrapper[4754]: I1005 22:04:37.813465 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dbf8aa85-c070-4f04-b788-b60dd6b8cc9c-utilities\") pod \"redhat-operators-2d2mr\" (UID: \"dbf8aa85-c070-4f04-b788-b60dd6b8cc9c\") " pod="openshift-marketplace/redhat-operators-2d2mr" Oct 05 22:04:37 crc kubenswrapper[4754]: I1005 22:04:37.835840 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p7grk\" (UniqueName: \"kubernetes.io/projected/dbf8aa85-c070-4f04-b788-b60dd6b8cc9c-kube-api-access-p7grk\") pod \"redhat-operators-2d2mr\" (UID: \"dbf8aa85-c070-4f04-b788-b60dd6b8cc9c\") " pod="openshift-marketplace/redhat-operators-2d2mr" Oct 05 22:04:38 crc kubenswrapper[4754]: I1005 22:04:38.001153 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2d2mr" Oct 05 22:04:38 crc kubenswrapper[4754]: I1005 22:04:38.479669 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-2d2mr"] Oct 05 22:04:38 crc kubenswrapper[4754]: I1005 22:04:38.749551 4754 generic.go:334] "Generic (PLEG): container finished" podID="dbf8aa85-c070-4f04-b788-b60dd6b8cc9c" containerID="864eefaec67a4ddbffc7825982742ea19bfdfd2ae0eac0214685f432aa4a7bcd" exitCode=0 Oct 05 22:04:38 crc kubenswrapper[4754]: I1005 22:04:38.749659 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2d2mr" event={"ID":"dbf8aa85-c070-4f04-b788-b60dd6b8cc9c","Type":"ContainerDied","Data":"864eefaec67a4ddbffc7825982742ea19bfdfd2ae0eac0214685f432aa4a7bcd"} Oct 05 22:04:38 crc kubenswrapper[4754]: I1005 22:04:38.749689 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2d2mr" event={"ID":"dbf8aa85-c070-4f04-b788-b60dd6b8cc9c","Type":"ContainerStarted","Data":"00c7731fb1feec266979ba1999aa641a6c3ebf02744fa8c7938590532ebe863a"} Oct 05 22:04:38 crc kubenswrapper[4754]: I1005 22:04:38.755442 4754 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 05 22:04:40 crc kubenswrapper[4754]: I1005 22:04:40.773829 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2d2mr" event={"ID":"dbf8aa85-c070-4f04-b788-b60dd6b8cc9c","Type":"ContainerStarted","Data":"d3a1254ee9a93016c7556f35688d705e4ebbeede45a25ddbda0cc28182b368f9"} Oct 05 22:04:41 crc kubenswrapper[4754]: I1005 22:04:41.995218 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-4rqkm"] Oct 05 22:04:42 crc kubenswrapper[4754]: I1005 22:04:42.000369 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4rqkm" Oct 05 22:04:42 crc kubenswrapper[4754]: I1005 22:04:42.103368 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4rqkm"] Oct 05 22:04:42 crc kubenswrapper[4754]: I1005 22:04:42.117955 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c8a91434-1adc-4cfe-af34-8490669681ae-utilities\") pod \"certified-operators-4rqkm\" (UID: \"c8a91434-1adc-4cfe-af34-8490669681ae\") " pod="openshift-marketplace/certified-operators-4rqkm" Oct 05 22:04:42 crc kubenswrapper[4754]: I1005 22:04:42.118031 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c8a91434-1adc-4cfe-af34-8490669681ae-catalog-content\") pod \"certified-operators-4rqkm\" (UID: \"c8a91434-1adc-4cfe-af34-8490669681ae\") " pod="openshift-marketplace/certified-operators-4rqkm" Oct 05 22:04:42 crc kubenswrapper[4754]: I1005 22:04:42.118068 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qn7br\" (UniqueName: \"kubernetes.io/projected/c8a91434-1adc-4cfe-af34-8490669681ae-kube-api-access-qn7br\") pod \"certified-operators-4rqkm\" (UID: \"c8a91434-1adc-4cfe-af34-8490669681ae\") " pod="openshift-marketplace/certified-operators-4rqkm" Oct 05 22:04:42 crc kubenswrapper[4754]: I1005 22:04:42.219956 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c8a91434-1adc-4cfe-af34-8490669681ae-utilities\") pod \"certified-operators-4rqkm\" (UID: \"c8a91434-1adc-4cfe-af34-8490669681ae\") " pod="openshift-marketplace/certified-operators-4rqkm" Oct 05 22:04:42 crc kubenswrapper[4754]: I1005 22:04:42.220051 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c8a91434-1adc-4cfe-af34-8490669681ae-catalog-content\") pod \"certified-operators-4rqkm\" (UID: \"c8a91434-1adc-4cfe-af34-8490669681ae\") " pod="openshift-marketplace/certified-operators-4rqkm" Oct 05 22:04:42 crc kubenswrapper[4754]: I1005 22:04:42.220087 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qn7br\" (UniqueName: \"kubernetes.io/projected/c8a91434-1adc-4cfe-af34-8490669681ae-kube-api-access-qn7br\") pod \"certified-operators-4rqkm\" (UID: \"c8a91434-1adc-4cfe-af34-8490669681ae\") " pod="openshift-marketplace/certified-operators-4rqkm" Oct 05 22:04:42 crc kubenswrapper[4754]: I1005 22:04:42.220545 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c8a91434-1adc-4cfe-af34-8490669681ae-utilities\") pod \"certified-operators-4rqkm\" (UID: \"c8a91434-1adc-4cfe-af34-8490669681ae\") " pod="openshift-marketplace/certified-operators-4rqkm" Oct 05 22:04:42 crc kubenswrapper[4754]: I1005 22:04:42.220582 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c8a91434-1adc-4cfe-af34-8490669681ae-catalog-content\") pod \"certified-operators-4rqkm\" (UID: \"c8a91434-1adc-4cfe-af34-8490669681ae\") " pod="openshift-marketplace/certified-operators-4rqkm" Oct 05 22:04:42 crc kubenswrapper[4754]: I1005 22:04:42.238614 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qn7br\" (UniqueName: \"kubernetes.io/projected/c8a91434-1adc-4cfe-af34-8490669681ae-kube-api-access-qn7br\") pod \"certified-operators-4rqkm\" (UID: \"c8a91434-1adc-4cfe-af34-8490669681ae\") " pod="openshift-marketplace/certified-operators-4rqkm" Oct 05 22:04:42 crc kubenswrapper[4754]: I1005 22:04:42.326930 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4rqkm" Oct 05 22:04:42 crc kubenswrapper[4754]: I1005 22:04:42.981255 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4rqkm"] Oct 05 22:04:42 crc kubenswrapper[4754]: W1005 22:04:42.993283 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc8a91434_1adc_4cfe_af34_8490669681ae.slice/crio-780b4106de1025162044c1f47574db07ee4cb635ce8d4f740cac9774e8662e52 WatchSource:0}: Error finding container 780b4106de1025162044c1f47574db07ee4cb635ce8d4f740cac9774e8662e52: Status 404 returned error can't find the container with id 780b4106de1025162044c1f47574db07ee4cb635ce8d4f740cac9774e8662e52 Oct 05 22:04:43 crc kubenswrapper[4754]: I1005 22:04:43.810618 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4rqkm" event={"ID":"c8a91434-1adc-4cfe-af34-8490669681ae","Type":"ContainerStarted","Data":"d07114eff1d8e3c7e3552dbf3ef216b369fceb4d8d9c1ffced02342dfa53a609"} Oct 05 22:04:43 crc kubenswrapper[4754]: I1005 22:04:43.810929 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4rqkm" event={"ID":"c8a91434-1adc-4cfe-af34-8490669681ae","Type":"ContainerStarted","Data":"780b4106de1025162044c1f47574db07ee4cb635ce8d4f740cac9774e8662e52"} Oct 05 22:04:44 crc kubenswrapper[4754]: I1005 22:04:44.830096 4754 generic.go:334] "Generic (PLEG): container finished" podID="c8a91434-1adc-4cfe-af34-8490669681ae" containerID="d07114eff1d8e3c7e3552dbf3ef216b369fceb4d8d9c1ffced02342dfa53a609" exitCode=0 Oct 05 22:04:44 crc kubenswrapper[4754]: I1005 22:04:44.830260 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4rqkm" event={"ID":"c8a91434-1adc-4cfe-af34-8490669681ae","Type":"ContainerDied","Data":"d07114eff1d8e3c7e3552dbf3ef216b369fceb4d8d9c1ffced02342dfa53a609"} Oct 05 22:04:47 crc kubenswrapper[4754]: I1005 22:04:47.876016 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4rqkm" event={"ID":"c8a91434-1adc-4cfe-af34-8490669681ae","Type":"ContainerStarted","Data":"afa2ef73105a6fe8bd2fd56be1d84bc54052656f6e35335cae6e2c0988692571"} Oct 05 22:04:53 crc kubenswrapper[4754]: I1005 22:04:53.964138 4754 generic.go:334] "Generic (PLEG): container finished" podID="dbf8aa85-c070-4f04-b788-b60dd6b8cc9c" containerID="d3a1254ee9a93016c7556f35688d705e4ebbeede45a25ddbda0cc28182b368f9" exitCode=0 Oct 05 22:04:53 crc kubenswrapper[4754]: I1005 22:04:53.964682 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2d2mr" event={"ID":"dbf8aa85-c070-4f04-b788-b60dd6b8cc9c","Type":"ContainerDied","Data":"d3a1254ee9a93016c7556f35688d705e4ebbeede45a25ddbda0cc28182b368f9"} Oct 05 22:04:54 crc kubenswrapper[4754]: I1005 22:04:54.974707 4754 generic.go:334] "Generic (PLEG): container finished" podID="c8a91434-1adc-4cfe-af34-8490669681ae" containerID="afa2ef73105a6fe8bd2fd56be1d84bc54052656f6e35335cae6e2c0988692571" exitCode=0 Oct 05 22:04:54 crc kubenswrapper[4754]: I1005 22:04:54.974812 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4rqkm" event={"ID":"c8a91434-1adc-4cfe-af34-8490669681ae","Type":"ContainerDied","Data":"afa2ef73105a6fe8bd2fd56be1d84bc54052656f6e35335cae6e2c0988692571"} Oct 05 22:04:55 crc kubenswrapper[4754]: I1005 22:04:55.986566 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4rqkm" event={"ID":"c8a91434-1adc-4cfe-af34-8490669681ae","Type":"ContainerStarted","Data":"5431706f66422aeae1fb42882e2b0b496168ea3464e094e51e0a28c622dff975"} Oct 05 22:04:55 crc kubenswrapper[4754]: I1005 22:04:55.993355 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2d2mr" event={"ID":"dbf8aa85-c070-4f04-b788-b60dd6b8cc9c","Type":"ContainerStarted","Data":"2e4abf8237938108950b4975f3e27a5c435dcfbcd3106a539e9922ef2c918309"} Oct 05 22:04:56 crc kubenswrapper[4754]: I1005 22:04:56.032996 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-4rqkm" podStartSLOduration=4.378051536 podStartE2EDuration="15.032977914s" podCreationTimestamp="2025-10-05 22:04:41 +0000 UTC" firstStartedPulling="2025-10-05 22:04:44.83365919 +0000 UTC m=+4208.737777940" lastFinishedPulling="2025-10-05 22:04:55.488585598 +0000 UTC m=+4219.392704318" observedRunningTime="2025-10-05 22:04:56.011286224 +0000 UTC m=+4219.915404934" watchObservedRunningTime="2025-10-05 22:04:56.032977914 +0000 UTC m=+4219.937096614" Oct 05 22:04:56 crc kubenswrapper[4754]: I1005 22:04:56.034488 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-2d2mr" podStartSLOduration=2.747063909 podStartE2EDuration="19.034481533s" podCreationTimestamp="2025-10-05 22:04:37 +0000 UTC" firstStartedPulling="2025-10-05 22:04:38.755259408 +0000 UTC m=+4202.659378118" lastFinishedPulling="2025-10-05 22:04:55.042677002 +0000 UTC m=+4218.946795742" observedRunningTime="2025-10-05 22:04:56.031640131 +0000 UTC m=+4219.935758841" watchObservedRunningTime="2025-10-05 22:04:56.034481533 +0000 UTC m=+4219.938600233" Oct 05 22:04:58 crc kubenswrapper[4754]: I1005 22:04:58.002004 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-2d2mr" Oct 05 22:04:58 crc kubenswrapper[4754]: I1005 22:04:58.002379 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-2d2mr" Oct 05 22:04:59 crc kubenswrapper[4754]: I1005 22:04:59.058454 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-2d2mr" podUID="dbf8aa85-c070-4f04-b788-b60dd6b8cc9c" containerName="registry-server" probeResult="failure" output=< Oct 05 22:04:59 crc kubenswrapper[4754]: timeout: failed to connect service ":50051" within 1s Oct 05 22:04:59 crc kubenswrapper[4754]: > Oct 05 22:05:02 crc kubenswrapper[4754]: I1005 22:05:02.327718 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-4rqkm" Oct 05 22:05:02 crc kubenswrapper[4754]: I1005 22:05:02.328155 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-4rqkm" Oct 05 22:05:02 crc kubenswrapper[4754]: I1005 22:05:02.382083 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-4rqkm" Oct 05 22:05:03 crc kubenswrapper[4754]: I1005 22:05:03.133418 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-4rqkm" Oct 05 22:05:03 crc kubenswrapper[4754]: I1005 22:05:03.197165 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-4rqkm"] Oct 05 22:05:05 crc kubenswrapper[4754]: I1005 22:05:05.097537 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-4rqkm" podUID="c8a91434-1adc-4cfe-af34-8490669681ae" containerName="registry-server" containerID="cri-o://5431706f66422aeae1fb42882e2b0b496168ea3464e094e51e0a28c622dff975" gracePeriod=2 Oct 05 22:05:05 crc kubenswrapper[4754]: I1005 22:05:05.700450 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4rqkm" Oct 05 22:05:05 crc kubenswrapper[4754]: I1005 22:05:05.884652 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qn7br\" (UniqueName: \"kubernetes.io/projected/c8a91434-1adc-4cfe-af34-8490669681ae-kube-api-access-qn7br\") pod \"c8a91434-1adc-4cfe-af34-8490669681ae\" (UID: \"c8a91434-1adc-4cfe-af34-8490669681ae\") " Oct 05 22:05:05 crc kubenswrapper[4754]: I1005 22:05:05.884697 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c8a91434-1adc-4cfe-af34-8490669681ae-utilities\") pod \"c8a91434-1adc-4cfe-af34-8490669681ae\" (UID: \"c8a91434-1adc-4cfe-af34-8490669681ae\") " Oct 05 22:05:05 crc kubenswrapper[4754]: I1005 22:05:05.884776 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c8a91434-1adc-4cfe-af34-8490669681ae-catalog-content\") pod \"c8a91434-1adc-4cfe-af34-8490669681ae\" (UID: \"c8a91434-1adc-4cfe-af34-8490669681ae\") " Oct 05 22:05:05 crc kubenswrapper[4754]: I1005 22:05:05.885954 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c8a91434-1adc-4cfe-af34-8490669681ae-utilities" (OuterVolumeSpecName: "utilities") pod "c8a91434-1adc-4cfe-af34-8490669681ae" (UID: "c8a91434-1adc-4cfe-af34-8490669681ae"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 22:05:05 crc kubenswrapper[4754]: I1005 22:05:05.892208 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c8a91434-1adc-4cfe-af34-8490669681ae-kube-api-access-qn7br" (OuterVolumeSpecName: "kube-api-access-qn7br") pod "c8a91434-1adc-4cfe-af34-8490669681ae" (UID: "c8a91434-1adc-4cfe-af34-8490669681ae"). InnerVolumeSpecName "kube-api-access-qn7br". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 22:05:05 crc kubenswrapper[4754]: I1005 22:05:05.945471 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c8a91434-1adc-4cfe-af34-8490669681ae-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c8a91434-1adc-4cfe-af34-8490669681ae" (UID: "c8a91434-1adc-4cfe-af34-8490669681ae"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 22:05:05 crc kubenswrapper[4754]: I1005 22:05:05.988452 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qn7br\" (UniqueName: \"kubernetes.io/projected/c8a91434-1adc-4cfe-af34-8490669681ae-kube-api-access-qn7br\") on node \"crc\" DevicePath \"\"" Oct 05 22:05:05 crc kubenswrapper[4754]: I1005 22:05:05.988492 4754 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c8a91434-1adc-4cfe-af34-8490669681ae-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 22:05:05 crc kubenswrapper[4754]: I1005 22:05:05.988519 4754 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c8a91434-1adc-4cfe-af34-8490669681ae-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 22:05:06 crc kubenswrapper[4754]: I1005 22:05:06.110871 4754 generic.go:334] "Generic (PLEG): container finished" podID="c8a91434-1adc-4cfe-af34-8490669681ae" containerID="5431706f66422aeae1fb42882e2b0b496168ea3464e094e51e0a28c622dff975" exitCode=0 Oct 05 22:05:06 crc kubenswrapper[4754]: I1005 22:05:06.110910 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4rqkm" event={"ID":"c8a91434-1adc-4cfe-af34-8490669681ae","Type":"ContainerDied","Data":"5431706f66422aeae1fb42882e2b0b496168ea3464e094e51e0a28c622dff975"} Oct 05 22:05:06 crc kubenswrapper[4754]: I1005 22:05:06.110988 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4rqkm" Oct 05 22:05:06 crc kubenswrapper[4754]: I1005 22:05:06.111186 4754 scope.go:117] "RemoveContainer" containerID="5431706f66422aeae1fb42882e2b0b496168ea3464e094e51e0a28c622dff975" Oct 05 22:05:06 crc kubenswrapper[4754]: I1005 22:05:06.111171 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4rqkm" event={"ID":"c8a91434-1adc-4cfe-af34-8490669681ae","Type":"ContainerDied","Data":"780b4106de1025162044c1f47574db07ee4cb635ce8d4f740cac9774e8662e52"} Oct 05 22:05:06 crc kubenswrapper[4754]: I1005 22:05:06.134640 4754 scope.go:117] "RemoveContainer" containerID="afa2ef73105a6fe8bd2fd56be1d84bc54052656f6e35335cae6e2c0988692571" Oct 05 22:05:06 crc kubenswrapper[4754]: I1005 22:05:06.154529 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-4rqkm"] Oct 05 22:05:06 crc kubenswrapper[4754]: I1005 22:05:06.160126 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-4rqkm"] Oct 05 22:05:06 crc kubenswrapper[4754]: I1005 22:05:06.175201 4754 scope.go:117] "RemoveContainer" containerID="d07114eff1d8e3c7e3552dbf3ef216b369fceb4d8d9c1ffced02342dfa53a609" Oct 05 22:05:06 crc kubenswrapper[4754]: I1005 22:05:06.230064 4754 scope.go:117] "RemoveContainer" containerID="5431706f66422aeae1fb42882e2b0b496168ea3464e094e51e0a28c622dff975" Oct 05 22:05:06 crc kubenswrapper[4754]: E1005 22:05:06.230584 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5431706f66422aeae1fb42882e2b0b496168ea3464e094e51e0a28c622dff975\": container with ID starting with 5431706f66422aeae1fb42882e2b0b496168ea3464e094e51e0a28c622dff975 not found: ID does not exist" containerID="5431706f66422aeae1fb42882e2b0b496168ea3464e094e51e0a28c622dff975" Oct 05 22:05:06 crc kubenswrapper[4754]: I1005 22:05:06.230632 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5431706f66422aeae1fb42882e2b0b496168ea3464e094e51e0a28c622dff975"} err="failed to get container status \"5431706f66422aeae1fb42882e2b0b496168ea3464e094e51e0a28c622dff975\": rpc error: code = NotFound desc = could not find container \"5431706f66422aeae1fb42882e2b0b496168ea3464e094e51e0a28c622dff975\": container with ID starting with 5431706f66422aeae1fb42882e2b0b496168ea3464e094e51e0a28c622dff975 not found: ID does not exist" Oct 05 22:05:06 crc kubenswrapper[4754]: I1005 22:05:06.230661 4754 scope.go:117] "RemoveContainer" containerID="afa2ef73105a6fe8bd2fd56be1d84bc54052656f6e35335cae6e2c0988692571" Oct 05 22:05:06 crc kubenswrapper[4754]: E1005 22:05:06.231194 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"afa2ef73105a6fe8bd2fd56be1d84bc54052656f6e35335cae6e2c0988692571\": container with ID starting with afa2ef73105a6fe8bd2fd56be1d84bc54052656f6e35335cae6e2c0988692571 not found: ID does not exist" containerID="afa2ef73105a6fe8bd2fd56be1d84bc54052656f6e35335cae6e2c0988692571" Oct 05 22:05:06 crc kubenswrapper[4754]: I1005 22:05:06.231378 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"afa2ef73105a6fe8bd2fd56be1d84bc54052656f6e35335cae6e2c0988692571"} err="failed to get container status \"afa2ef73105a6fe8bd2fd56be1d84bc54052656f6e35335cae6e2c0988692571\": rpc error: code = NotFound desc = could not find container \"afa2ef73105a6fe8bd2fd56be1d84bc54052656f6e35335cae6e2c0988692571\": container with ID starting with afa2ef73105a6fe8bd2fd56be1d84bc54052656f6e35335cae6e2c0988692571 not found: ID does not exist" Oct 05 22:05:06 crc kubenswrapper[4754]: I1005 22:05:06.231558 4754 scope.go:117] "RemoveContainer" containerID="d07114eff1d8e3c7e3552dbf3ef216b369fceb4d8d9c1ffced02342dfa53a609" Oct 05 22:05:06 crc kubenswrapper[4754]: E1005 22:05:06.232053 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d07114eff1d8e3c7e3552dbf3ef216b369fceb4d8d9c1ffced02342dfa53a609\": container with ID starting with d07114eff1d8e3c7e3552dbf3ef216b369fceb4d8d9c1ffced02342dfa53a609 not found: ID does not exist" containerID="d07114eff1d8e3c7e3552dbf3ef216b369fceb4d8d9c1ffced02342dfa53a609" Oct 05 22:05:06 crc kubenswrapper[4754]: I1005 22:05:06.232087 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d07114eff1d8e3c7e3552dbf3ef216b369fceb4d8d9c1ffced02342dfa53a609"} err="failed to get container status \"d07114eff1d8e3c7e3552dbf3ef216b369fceb4d8d9c1ffced02342dfa53a609\": rpc error: code = NotFound desc = could not find container \"d07114eff1d8e3c7e3552dbf3ef216b369fceb4d8d9c1ffced02342dfa53a609\": container with ID starting with d07114eff1d8e3c7e3552dbf3ef216b369fceb4d8d9c1ffced02342dfa53a609 not found: ID does not exist" Oct 05 22:05:06 crc kubenswrapper[4754]: I1005 22:05:06.850180 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c8a91434-1adc-4cfe-af34-8490669681ae" path="/var/lib/kubelet/pods/c8a91434-1adc-4cfe-af34-8490669681ae/volumes" Oct 05 22:05:09 crc kubenswrapper[4754]: I1005 22:05:09.065949 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-2d2mr" podUID="dbf8aa85-c070-4f04-b788-b60dd6b8cc9c" containerName="registry-server" probeResult="failure" output=< Oct 05 22:05:09 crc kubenswrapper[4754]: timeout: failed to connect service ":50051" within 1s Oct 05 22:05:09 crc kubenswrapper[4754]: > Oct 05 22:05:18 crc kubenswrapper[4754]: I1005 22:05:18.063159 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-2d2mr" Oct 05 22:05:18 crc kubenswrapper[4754]: I1005 22:05:18.134988 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-2d2mr" Oct 05 22:05:18 crc kubenswrapper[4754]: I1005 22:05:18.319266 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-2d2mr"] Oct 05 22:05:19 crc kubenswrapper[4754]: I1005 22:05:19.243799 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-2d2mr" podUID="dbf8aa85-c070-4f04-b788-b60dd6b8cc9c" containerName="registry-server" containerID="cri-o://2e4abf8237938108950b4975f3e27a5c435dcfbcd3106a539e9922ef2c918309" gracePeriod=2 Oct 05 22:05:20 crc kubenswrapper[4754]: I1005 22:05:20.253937 4754 generic.go:334] "Generic (PLEG): container finished" podID="dbf8aa85-c070-4f04-b788-b60dd6b8cc9c" containerID="2e4abf8237938108950b4975f3e27a5c435dcfbcd3106a539e9922ef2c918309" exitCode=0 Oct 05 22:05:20 crc kubenswrapper[4754]: I1005 22:05:20.254361 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2d2mr" event={"ID":"dbf8aa85-c070-4f04-b788-b60dd6b8cc9c","Type":"ContainerDied","Data":"2e4abf8237938108950b4975f3e27a5c435dcfbcd3106a539e9922ef2c918309"} Oct 05 22:05:20 crc kubenswrapper[4754]: I1005 22:05:20.254388 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2d2mr" event={"ID":"dbf8aa85-c070-4f04-b788-b60dd6b8cc9c","Type":"ContainerDied","Data":"00c7731fb1feec266979ba1999aa641a6c3ebf02744fa8c7938590532ebe863a"} Oct 05 22:05:20 crc kubenswrapper[4754]: I1005 22:05:20.254400 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="00c7731fb1feec266979ba1999aa641a6c3ebf02744fa8c7938590532ebe863a" Oct 05 22:05:20 crc kubenswrapper[4754]: I1005 22:05:20.295565 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2d2mr" Oct 05 22:05:20 crc kubenswrapper[4754]: I1005 22:05:20.378523 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dbf8aa85-c070-4f04-b788-b60dd6b8cc9c-catalog-content\") pod \"dbf8aa85-c070-4f04-b788-b60dd6b8cc9c\" (UID: \"dbf8aa85-c070-4f04-b788-b60dd6b8cc9c\") " Oct 05 22:05:20 crc kubenswrapper[4754]: I1005 22:05:20.378669 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p7grk\" (UniqueName: \"kubernetes.io/projected/dbf8aa85-c070-4f04-b788-b60dd6b8cc9c-kube-api-access-p7grk\") pod \"dbf8aa85-c070-4f04-b788-b60dd6b8cc9c\" (UID: \"dbf8aa85-c070-4f04-b788-b60dd6b8cc9c\") " Oct 05 22:05:20 crc kubenswrapper[4754]: I1005 22:05:20.378727 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dbf8aa85-c070-4f04-b788-b60dd6b8cc9c-utilities\") pod \"dbf8aa85-c070-4f04-b788-b60dd6b8cc9c\" (UID: \"dbf8aa85-c070-4f04-b788-b60dd6b8cc9c\") " Oct 05 22:05:20 crc kubenswrapper[4754]: I1005 22:05:20.379421 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dbf8aa85-c070-4f04-b788-b60dd6b8cc9c-utilities" (OuterVolumeSpecName: "utilities") pod "dbf8aa85-c070-4f04-b788-b60dd6b8cc9c" (UID: "dbf8aa85-c070-4f04-b788-b60dd6b8cc9c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 22:05:20 crc kubenswrapper[4754]: I1005 22:05:20.386719 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dbf8aa85-c070-4f04-b788-b60dd6b8cc9c-kube-api-access-p7grk" (OuterVolumeSpecName: "kube-api-access-p7grk") pod "dbf8aa85-c070-4f04-b788-b60dd6b8cc9c" (UID: "dbf8aa85-c070-4f04-b788-b60dd6b8cc9c"). InnerVolumeSpecName "kube-api-access-p7grk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 22:05:20 crc kubenswrapper[4754]: I1005 22:05:20.481552 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p7grk\" (UniqueName: \"kubernetes.io/projected/dbf8aa85-c070-4f04-b788-b60dd6b8cc9c-kube-api-access-p7grk\") on node \"crc\" DevicePath \"\"" Oct 05 22:05:20 crc kubenswrapper[4754]: I1005 22:05:20.481589 4754 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dbf8aa85-c070-4f04-b788-b60dd6b8cc9c-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 22:05:20 crc kubenswrapper[4754]: I1005 22:05:20.499032 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dbf8aa85-c070-4f04-b788-b60dd6b8cc9c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "dbf8aa85-c070-4f04-b788-b60dd6b8cc9c" (UID: "dbf8aa85-c070-4f04-b788-b60dd6b8cc9c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 22:05:20 crc kubenswrapper[4754]: I1005 22:05:20.583849 4754 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dbf8aa85-c070-4f04-b788-b60dd6b8cc9c-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 22:05:21 crc kubenswrapper[4754]: I1005 22:05:21.261800 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2d2mr" Oct 05 22:05:21 crc kubenswrapper[4754]: I1005 22:05:21.286606 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-2d2mr"] Oct 05 22:05:21 crc kubenswrapper[4754]: I1005 22:05:21.296812 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-2d2mr"] Oct 05 22:05:22 crc kubenswrapper[4754]: I1005 22:05:22.849321 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dbf8aa85-c070-4f04-b788-b60dd6b8cc9c" path="/var/lib/kubelet/pods/dbf8aa85-c070-4f04-b788-b60dd6b8cc9c/volumes" Oct 05 22:05:35 crc kubenswrapper[4754]: I1005 22:05:35.245154 4754 patch_prober.go:28] interesting pod/machine-config-daemon-b2h9k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 22:05:35 crc kubenswrapper[4754]: I1005 22:05:35.245843 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 22:06:05 crc kubenswrapper[4754]: I1005 22:06:05.245206 4754 patch_prober.go:28] interesting pod/machine-config-daemon-b2h9k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 22:06:05 crc kubenswrapper[4754]: I1005 22:06:05.247898 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 22:06:35 crc kubenswrapper[4754]: I1005 22:06:35.245644 4754 patch_prober.go:28] interesting pod/machine-config-daemon-b2h9k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 22:06:35 crc kubenswrapper[4754]: I1005 22:06:35.246275 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 22:06:35 crc kubenswrapper[4754]: I1005 22:06:35.246325 4754 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" Oct 05 22:06:35 crc kubenswrapper[4754]: I1005 22:06:35.247162 4754 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d925cdc080b01050e7b9377dba2f144d30f517d60778d239b53617b071498b40"} pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 05 22:06:35 crc kubenswrapper[4754]: I1005 22:06:35.247228 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" containerName="machine-config-daemon" containerID="cri-o://d925cdc080b01050e7b9377dba2f144d30f517d60778d239b53617b071498b40" gracePeriod=600 Oct 05 22:06:35 crc kubenswrapper[4754]: E1005 22:06:35.382096 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 22:06:36 crc kubenswrapper[4754]: I1005 22:06:36.046886 4754 generic.go:334] "Generic (PLEG): container finished" podID="be95a413-6a59-45b4-84b7-b43cae694a26" containerID="d925cdc080b01050e7b9377dba2f144d30f517d60778d239b53617b071498b40" exitCode=0 Oct 05 22:06:36 crc kubenswrapper[4754]: I1005 22:06:36.046979 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" event={"ID":"be95a413-6a59-45b4-84b7-b43cae694a26","Type":"ContainerDied","Data":"d925cdc080b01050e7b9377dba2f144d30f517d60778d239b53617b071498b40"} Oct 05 22:06:36 crc kubenswrapper[4754]: I1005 22:06:36.047185 4754 scope.go:117] "RemoveContainer" containerID="7e8376fda04f81f5bc9127199e823c84456ce7d86c14da7ebaece2d9e07200e8" Oct 05 22:06:36 crc kubenswrapper[4754]: I1005 22:06:36.047842 4754 scope.go:117] "RemoveContainer" containerID="d925cdc080b01050e7b9377dba2f144d30f517d60778d239b53617b071498b40" Oct 05 22:06:36 crc kubenswrapper[4754]: E1005 22:06:36.048167 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 22:06:50 crc kubenswrapper[4754]: I1005 22:06:50.837735 4754 scope.go:117] "RemoveContainer" containerID="d925cdc080b01050e7b9377dba2f144d30f517d60778d239b53617b071498b40" Oct 05 22:06:50 crc kubenswrapper[4754]: E1005 22:06:50.838339 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 22:07:03 crc kubenswrapper[4754]: I1005 22:07:03.837751 4754 scope.go:117] "RemoveContainer" containerID="d925cdc080b01050e7b9377dba2f144d30f517d60778d239b53617b071498b40" Oct 05 22:07:03 crc kubenswrapper[4754]: E1005 22:07:03.841312 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 22:07:14 crc kubenswrapper[4754]: I1005 22:07:14.839624 4754 scope.go:117] "RemoveContainer" containerID="d925cdc080b01050e7b9377dba2f144d30f517d60778d239b53617b071498b40" Oct 05 22:07:14 crc kubenswrapper[4754]: E1005 22:07:14.840554 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 22:07:16 crc kubenswrapper[4754]: I1005 22:07:16.521629 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-7wj22"] Oct 05 22:07:16 crc kubenswrapper[4754]: E1005 22:07:16.522282 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dbf8aa85-c070-4f04-b788-b60dd6b8cc9c" containerName="extract-utilities" Oct 05 22:07:16 crc kubenswrapper[4754]: I1005 22:07:16.522294 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="dbf8aa85-c070-4f04-b788-b60dd6b8cc9c" containerName="extract-utilities" Oct 05 22:07:16 crc kubenswrapper[4754]: E1005 22:07:16.522305 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8a91434-1adc-4cfe-af34-8490669681ae" containerName="extract-utilities" Oct 05 22:07:16 crc kubenswrapper[4754]: I1005 22:07:16.522311 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8a91434-1adc-4cfe-af34-8490669681ae" containerName="extract-utilities" Oct 05 22:07:16 crc kubenswrapper[4754]: E1005 22:07:16.522323 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8a91434-1adc-4cfe-af34-8490669681ae" containerName="registry-server" Oct 05 22:07:16 crc kubenswrapper[4754]: I1005 22:07:16.522329 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8a91434-1adc-4cfe-af34-8490669681ae" containerName="registry-server" Oct 05 22:07:16 crc kubenswrapper[4754]: E1005 22:07:16.522338 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8a91434-1adc-4cfe-af34-8490669681ae" containerName="extract-content" Oct 05 22:07:16 crc kubenswrapper[4754]: I1005 22:07:16.522344 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8a91434-1adc-4cfe-af34-8490669681ae" containerName="extract-content" Oct 05 22:07:16 crc kubenswrapper[4754]: E1005 22:07:16.522356 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dbf8aa85-c070-4f04-b788-b60dd6b8cc9c" containerName="extract-content" Oct 05 22:07:16 crc kubenswrapper[4754]: I1005 22:07:16.522361 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="dbf8aa85-c070-4f04-b788-b60dd6b8cc9c" containerName="extract-content" Oct 05 22:07:16 crc kubenswrapper[4754]: E1005 22:07:16.522371 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dbf8aa85-c070-4f04-b788-b60dd6b8cc9c" containerName="registry-server" Oct 05 22:07:16 crc kubenswrapper[4754]: I1005 22:07:16.522376 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="dbf8aa85-c070-4f04-b788-b60dd6b8cc9c" containerName="registry-server" Oct 05 22:07:16 crc kubenswrapper[4754]: I1005 22:07:16.522546 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8a91434-1adc-4cfe-af34-8490669681ae" containerName="registry-server" Oct 05 22:07:16 crc kubenswrapper[4754]: I1005 22:07:16.522557 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="dbf8aa85-c070-4f04-b788-b60dd6b8cc9c" containerName="registry-server" Oct 05 22:07:16 crc kubenswrapper[4754]: I1005 22:07:16.525851 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7wj22" Oct 05 22:07:16 crc kubenswrapper[4754]: I1005 22:07:16.546324 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-7wj22"] Oct 05 22:07:16 crc kubenswrapper[4754]: I1005 22:07:16.649917 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/86a96b26-33d0-423e-a98f-5ab78139da4f-utilities\") pod \"community-operators-7wj22\" (UID: \"86a96b26-33d0-423e-a98f-5ab78139da4f\") " pod="openshift-marketplace/community-operators-7wj22" Oct 05 22:07:16 crc kubenswrapper[4754]: I1005 22:07:16.650607 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/86a96b26-33d0-423e-a98f-5ab78139da4f-catalog-content\") pod \"community-operators-7wj22\" (UID: \"86a96b26-33d0-423e-a98f-5ab78139da4f\") " pod="openshift-marketplace/community-operators-7wj22" Oct 05 22:07:16 crc kubenswrapper[4754]: I1005 22:07:16.650788 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dwc5t\" (UniqueName: \"kubernetes.io/projected/86a96b26-33d0-423e-a98f-5ab78139da4f-kube-api-access-dwc5t\") pod \"community-operators-7wj22\" (UID: \"86a96b26-33d0-423e-a98f-5ab78139da4f\") " pod="openshift-marketplace/community-operators-7wj22" Oct 05 22:07:16 crc kubenswrapper[4754]: I1005 22:07:16.752458 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dwc5t\" (UniqueName: \"kubernetes.io/projected/86a96b26-33d0-423e-a98f-5ab78139da4f-kube-api-access-dwc5t\") pod \"community-operators-7wj22\" (UID: \"86a96b26-33d0-423e-a98f-5ab78139da4f\") " pod="openshift-marketplace/community-operators-7wj22" Oct 05 22:07:16 crc kubenswrapper[4754]: I1005 22:07:16.752568 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/86a96b26-33d0-423e-a98f-5ab78139da4f-utilities\") pod \"community-operators-7wj22\" (UID: \"86a96b26-33d0-423e-a98f-5ab78139da4f\") " pod="openshift-marketplace/community-operators-7wj22" Oct 05 22:07:16 crc kubenswrapper[4754]: I1005 22:07:16.752593 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/86a96b26-33d0-423e-a98f-5ab78139da4f-catalog-content\") pod \"community-operators-7wj22\" (UID: \"86a96b26-33d0-423e-a98f-5ab78139da4f\") " pod="openshift-marketplace/community-operators-7wj22" Oct 05 22:07:16 crc kubenswrapper[4754]: I1005 22:07:16.753021 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/86a96b26-33d0-423e-a98f-5ab78139da4f-catalog-content\") pod \"community-operators-7wj22\" (UID: \"86a96b26-33d0-423e-a98f-5ab78139da4f\") " pod="openshift-marketplace/community-operators-7wj22" Oct 05 22:07:16 crc kubenswrapper[4754]: I1005 22:07:16.753135 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/86a96b26-33d0-423e-a98f-5ab78139da4f-utilities\") pod \"community-operators-7wj22\" (UID: \"86a96b26-33d0-423e-a98f-5ab78139da4f\") " pod="openshift-marketplace/community-operators-7wj22" Oct 05 22:07:16 crc kubenswrapper[4754]: I1005 22:07:16.770778 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dwc5t\" (UniqueName: \"kubernetes.io/projected/86a96b26-33d0-423e-a98f-5ab78139da4f-kube-api-access-dwc5t\") pod \"community-operators-7wj22\" (UID: \"86a96b26-33d0-423e-a98f-5ab78139da4f\") " pod="openshift-marketplace/community-operators-7wj22" Oct 05 22:07:16 crc kubenswrapper[4754]: I1005 22:07:16.847371 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7wj22" Oct 05 22:07:17 crc kubenswrapper[4754]: I1005 22:07:17.478517 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-7wj22"] Oct 05 22:07:17 crc kubenswrapper[4754]: W1005 22:07:17.483051 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod86a96b26_33d0_423e_a98f_5ab78139da4f.slice/crio-b9acdd356dbaa81ecda9a72395ce7fa1825a664c9d60fe1625c5b86dbc66daa7 WatchSource:0}: Error finding container b9acdd356dbaa81ecda9a72395ce7fa1825a664c9d60fe1625c5b86dbc66daa7: Status 404 returned error can't find the container with id b9acdd356dbaa81ecda9a72395ce7fa1825a664c9d60fe1625c5b86dbc66daa7 Oct 05 22:07:17 crc kubenswrapper[4754]: I1005 22:07:17.520642 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-6rmz7"] Oct 05 22:07:17 crc kubenswrapper[4754]: I1005 22:07:17.574049 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6rmz7" Oct 05 22:07:17 crc kubenswrapper[4754]: I1005 22:07:17.574057 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-6rmz7"] Oct 05 22:07:17 crc kubenswrapper[4754]: I1005 22:07:17.678539 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gj5wz\" (UniqueName: \"kubernetes.io/projected/17e710fa-0b2b-4ed6-aed5-2a837e46df07-kube-api-access-gj5wz\") pod \"redhat-marketplace-6rmz7\" (UID: \"17e710fa-0b2b-4ed6-aed5-2a837e46df07\") " pod="openshift-marketplace/redhat-marketplace-6rmz7" Oct 05 22:07:17 crc kubenswrapper[4754]: I1005 22:07:17.678749 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/17e710fa-0b2b-4ed6-aed5-2a837e46df07-catalog-content\") pod \"redhat-marketplace-6rmz7\" (UID: \"17e710fa-0b2b-4ed6-aed5-2a837e46df07\") " pod="openshift-marketplace/redhat-marketplace-6rmz7" Oct 05 22:07:17 crc kubenswrapper[4754]: I1005 22:07:17.678984 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/17e710fa-0b2b-4ed6-aed5-2a837e46df07-utilities\") pod \"redhat-marketplace-6rmz7\" (UID: \"17e710fa-0b2b-4ed6-aed5-2a837e46df07\") " pod="openshift-marketplace/redhat-marketplace-6rmz7" Oct 05 22:07:17 crc kubenswrapper[4754]: I1005 22:07:17.780458 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/17e710fa-0b2b-4ed6-aed5-2a837e46df07-utilities\") pod \"redhat-marketplace-6rmz7\" (UID: \"17e710fa-0b2b-4ed6-aed5-2a837e46df07\") " pod="openshift-marketplace/redhat-marketplace-6rmz7" Oct 05 22:07:17 crc kubenswrapper[4754]: I1005 22:07:17.780598 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gj5wz\" (UniqueName: \"kubernetes.io/projected/17e710fa-0b2b-4ed6-aed5-2a837e46df07-kube-api-access-gj5wz\") pod \"redhat-marketplace-6rmz7\" (UID: \"17e710fa-0b2b-4ed6-aed5-2a837e46df07\") " pod="openshift-marketplace/redhat-marketplace-6rmz7" Oct 05 22:07:17 crc kubenswrapper[4754]: I1005 22:07:17.780623 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/17e710fa-0b2b-4ed6-aed5-2a837e46df07-catalog-content\") pod \"redhat-marketplace-6rmz7\" (UID: \"17e710fa-0b2b-4ed6-aed5-2a837e46df07\") " pod="openshift-marketplace/redhat-marketplace-6rmz7" Oct 05 22:07:17 crc kubenswrapper[4754]: I1005 22:07:17.780917 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/17e710fa-0b2b-4ed6-aed5-2a837e46df07-utilities\") pod \"redhat-marketplace-6rmz7\" (UID: \"17e710fa-0b2b-4ed6-aed5-2a837e46df07\") " pod="openshift-marketplace/redhat-marketplace-6rmz7" Oct 05 22:07:17 crc kubenswrapper[4754]: I1005 22:07:17.780958 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/17e710fa-0b2b-4ed6-aed5-2a837e46df07-catalog-content\") pod \"redhat-marketplace-6rmz7\" (UID: \"17e710fa-0b2b-4ed6-aed5-2a837e46df07\") " pod="openshift-marketplace/redhat-marketplace-6rmz7" Oct 05 22:07:17 crc kubenswrapper[4754]: I1005 22:07:17.798829 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gj5wz\" (UniqueName: \"kubernetes.io/projected/17e710fa-0b2b-4ed6-aed5-2a837e46df07-kube-api-access-gj5wz\") pod \"redhat-marketplace-6rmz7\" (UID: \"17e710fa-0b2b-4ed6-aed5-2a837e46df07\") " pod="openshift-marketplace/redhat-marketplace-6rmz7" Oct 05 22:07:17 crc kubenswrapper[4754]: I1005 22:07:17.944582 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6rmz7" Oct 05 22:07:18 crc kubenswrapper[4754]: I1005 22:07:18.371410 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-6rmz7"] Oct 05 22:07:18 crc kubenswrapper[4754]: W1005 22:07:18.386702 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod17e710fa_0b2b_4ed6_aed5_2a837e46df07.slice/crio-001c754043c79f031338c762171aefed89b42976da6580a23cc9cc661f182b6f WatchSource:0}: Error finding container 001c754043c79f031338c762171aefed89b42976da6580a23cc9cc661f182b6f: Status 404 returned error can't find the container with id 001c754043c79f031338c762171aefed89b42976da6580a23cc9cc661f182b6f Oct 05 22:07:18 crc kubenswrapper[4754]: I1005 22:07:18.481054 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6rmz7" event={"ID":"17e710fa-0b2b-4ed6-aed5-2a837e46df07","Type":"ContainerStarted","Data":"001c754043c79f031338c762171aefed89b42976da6580a23cc9cc661f182b6f"} Oct 05 22:07:18 crc kubenswrapper[4754]: I1005 22:07:18.482485 4754 generic.go:334] "Generic (PLEG): container finished" podID="86a96b26-33d0-423e-a98f-5ab78139da4f" containerID="a5a9944c21e6fe05637ac41609b07724132402cc499713be28e785cc7fb95925" exitCode=0 Oct 05 22:07:18 crc kubenswrapper[4754]: I1005 22:07:18.482545 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7wj22" event={"ID":"86a96b26-33d0-423e-a98f-5ab78139da4f","Type":"ContainerDied","Data":"a5a9944c21e6fe05637ac41609b07724132402cc499713be28e785cc7fb95925"} Oct 05 22:07:18 crc kubenswrapper[4754]: I1005 22:07:18.482572 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7wj22" event={"ID":"86a96b26-33d0-423e-a98f-5ab78139da4f","Type":"ContainerStarted","Data":"b9acdd356dbaa81ecda9a72395ce7fa1825a664c9d60fe1625c5b86dbc66daa7"} Oct 05 22:07:19 crc kubenswrapper[4754]: I1005 22:07:19.490092 4754 generic.go:334] "Generic (PLEG): container finished" podID="17e710fa-0b2b-4ed6-aed5-2a837e46df07" containerID="ea20209750207176fb19436e0a548f337d56bf3cdc301a71457c432ba5c33dad" exitCode=0 Oct 05 22:07:19 crc kubenswrapper[4754]: I1005 22:07:19.490192 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6rmz7" event={"ID":"17e710fa-0b2b-4ed6-aed5-2a837e46df07","Type":"ContainerDied","Data":"ea20209750207176fb19436e0a548f337d56bf3cdc301a71457c432ba5c33dad"} Oct 05 22:07:20 crc kubenswrapper[4754]: I1005 22:07:20.520841 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7wj22" event={"ID":"86a96b26-33d0-423e-a98f-5ab78139da4f","Type":"ContainerStarted","Data":"762c983cff71868008618da88e7833fa9d5f45a99796000785fc06bcae9c4dc0"} Oct 05 22:07:21 crc kubenswrapper[4754]: I1005 22:07:21.537293 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6rmz7" event={"ID":"17e710fa-0b2b-4ed6-aed5-2a837e46df07","Type":"ContainerStarted","Data":"b7877ab08784d556ed281df60796e1ca02d720884f309f42a30f1f4f9fa2abe1"} Oct 05 22:07:22 crc kubenswrapper[4754]: I1005 22:07:22.550921 4754 generic.go:334] "Generic (PLEG): container finished" podID="86a96b26-33d0-423e-a98f-5ab78139da4f" containerID="762c983cff71868008618da88e7833fa9d5f45a99796000785fc06bcae9c4dc0" exitCode=0 Oct 05 22:07:22 crc kubenswrapper[4754]: I1005 22:07:22.551016 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7wj22" event={"ID":"86a96b26-33d0-423e-a98f-5ab78139da4f","Type":"ContainerDied","Data":"762c983cff71868008618da88e7833fa9d5f45a99796000785fc06bcae9c4dc0"} Oct 05 22:07:22 crc kubenswrapper[4754]: I1005 22:07:22.554783 4754 generic.go:334] "Generic (PLEG): container finished" podID="17e710fa-0b2b-4ed6-aed5-2a837e46df07" containerID="b7877ab08784d556ed281df60796e1ca02d720884f309f42a30f1f4f9fa2abe1" exitCode=0 Oct 05 22:07:22 crc kubenswrapper[4754]: I1005 22:07:22.554825 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6rmz7" event={"ID":"17e710fa-0b2b-4ed6-aed5-2a837e46df07","Type":"ContainerDied","Data":"b7877ab08784d556ed281df60796e1ca02d720884f309f42a30f1f4f9fa2abe1"} Oct 05 22:07:23 crc kubenswrapper[4754]: I1005 22:07:23.565469 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6rmz7" event={"ID":"17e710fa-0b2b-4ed6-aed5-2a837e46df07","Type":"ContainerStarted","Data":"a6cbcdbb9d808a8bf7c89173dc82961850cfe9e0c76a884d6a141efd8c554aec"} Oct 05 22:07:23 crc kubenswrapper[4754]: I1005 22:07:23.567649 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7wj22" event={"ID":"86a96b26-33d0-423e-a98f-5ab78139da4f","Type":"ContainerStarted","Data":"0faf273f7b052c213d73e35169e2a57861b09f024b9e4b805718eb399fff7733"} Oct 05 22:07:23 crc kubenswrapper[4754]: I1005 22:07:23.588378 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-6rmz7" podStartSLOduration=3.116010766 podStartE2EDuration="6.58836394s" podCreationTimestamp="2025-10-05 22:07:17 +0000 UTC" firstStartedPulling="2025-10-05 22:07:19.491910677 +0000 UTC m=+4363.396029397" lastFinishedPulling="2025-10-05 22:07:22.964263861 +0000 UTC m=+4366.868382571" observedRunningTime="2025-10-05 22:07:23.586686748 +0000 UTC m=+4367.490805458" watchObservedRunningTime="2025-10-05 22:07:23.58836394 +0000 UTC m=+4367.492482640" Oct 05 22:07:23 crc kubenswrapper[4754]: I1005 22:07:23.613148 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-7wj22" podStartSLOduration=3.058800266 podStartE2EDuration="7.613128249s" podCreationTimestamp="2025-10-05 22:07:16 +0000 UTC" firstStartedPulling="2025-10-05 22:07:18.484520871 +0000 UTC m=+4362.388639581" lastFinishedPulling="2025-10-05 22:07:23.038848834 +0000 UTC m=+4366.942967564" observedRunningTime="2025-10-05 22:07:23.612029441 +0000 UTC m=+4367.516148151" watchObservedRunningTime="2025-10-05 22:07:23.613128249 +0000 UTC m=+4367.517246969" Oct 05 22:07:26 crc kubenswrapper[4754]: I1005 22:07:26.843792 4754 scope.go:117] "RemoveContainer" containerID="d925cdc080b01050e7b9377dba2f144d30f517d60778d239b53617b071498b40" Oct 05 22:07:26 crc kubenswrapper[4754]: E1005 22:07:26.845514 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 22:07:26 crc kubenswrapper[4754]: I1005 22:07:26.848442 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-7wj22" Oct 05 22:07:26 crc kubenswrapper[4754]: I1005 22:07:26.848561 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-7wj22" Oct 05 22:07:26 crc kubenswrapper[4754]: I1005 22:07:26.897130 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-7wj22" Oct 05 22:07:27 crc kubenswrapper[4754]: I1005 22:07:27.945700 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-6rmz7" Oct 05 22:07:27 crc kubenswrapper[4754]: I1005 22:07:27.945744 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-6rmz7" Oct 05 22:07:28 crc kubenswrapper[4754]: I1005 22:07:28.016384 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-6rmz7" Oct 05 22:07:28 crc kubenswrapper[4754]: I1005 22:07:28.678001 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-7wj22" Oct 05 22:07:28 crc kubenswrapper[4754]: I1005 22:07:28.689690 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-6rmz7" Oct 05 22:07:30 crc kubenswrapper[4754]: I1005 22:07:30.509587 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-7wj22"] Oct 05 22:07:30 crc kubenswrapper[4754]: I1005 22:07:30.637662 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-7wj22" podUID="86a96b26-33d0-423e-a98f-5ab78139da4f" containerName="registry-server" containerID="cri-o://0faf273f7b052c213d73e35169e2a57861b09f024b9e4b805718eb399fff7733" gracePeriod=2 Oct 05 22:07:31 crc kubenswrapper[4754]: I1005 22:07:31.106701 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-6rmz7"] Oct 05 22:07:31 crc kubenswrapper[4754]: I1005 22:07:31.107136 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-6rmz7" podUID="17e710fa-0b2b-4ed6-aed5-2a837e46df07" containerName="registry-server" containerID="cri-o://a6cbcdbb9d808a8bf7c89173dc82961850cfe9e0c76a884d6a141efd8c554aec" gracePeriod=2 Oct 05 22:07:31 crc kubenswrapper[4754]: I1005 22:07:31.297027 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7wj22" Oct 05 22:07:31 crc kubenswrapper[4754]: I1005 22:07:31.347559 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dwc5t\" (UniqueName: \"kubernetes.io/projected/86a96b26-33d0-423e-a98f-5ab78139da4f-kube-api-access-dwc5t\") pod \"86a96b26-33d0-423e-a98f-5ab78139da4f\" (UID: \"86a96b26-33d0-423e-a98f-5ab78139da4f\") " Oct 05 22:07:31 crc kubenswrapper[4754]: I1005 22:07:31.347766 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/86a96b26-33d0-423e-a98f-5ab78139da4f-utilities\") pod \"86a96b26-33d0-423e-a98f-5ab78139da4f\" (UID: \"86a96b26-33d0-423e-a98f-5ab78139da4f\") " Oct 05 22:07:31 crc kubenswrapper[4754]: I1005 22:07:31.347806 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/86a96b26-33d0-423e-a98f-5ab78139da4f-catalog-content\") pod \"86a96b26-33d0-423e-a98f-5ab78139da4f\" (UID: \"86a96b26-33d0-423e-a98f-5ab78139da4f\") " Oct 05 22:07:31 crc kubenswrapper[4754]: I1005 22:07:31.353902 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/86a96b26-33d0-423e-a98f-5ab78139da4f-utilities" (OuterVolumeSpecName: "utilities") pod "86a96b26-33d0-423e-a98f-5ab78139da4f" (UID: "86a96b26-33d0-423e-a98f-5ab78139da4f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 22:07:31 crc kubenswrapper[4754]: I1005 22:07:31.358466 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/86a96b26-33d0-423e-a98f-5ab78139da4f-kube-api-access-dwc5t" (OuterVolumeSpecName: "kube-api-access-dwc5t") pod "86a96b26-33d0-423e-a98f-5ab78139da4f" (UID: "86a96b26-33d0-423e-a98f-5ab78139da4f"). InnerVolumeSpecName "kube-api-access-dwc5t". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 22:07:31 crc kubenswrapper[4754]: I1005 22:07:31.406469 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/86a96b26-33d0-423e-a98f-5ab78139da4f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "86a96b26-33d0-423e-a98f-5ab78139da4f" (UID: "86a96b26-33d0-423e-a98f-5ab78139da4f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 22:07:31 crc kubenswrapper[4754]: I1005 22:07:31.450770 4754 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/86a96b26-33d0-423e-a98f-5ab78139da4f-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 22:07:31 crc kubenswrapper[4754]: I1005 22:07:31.450989 4754 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/86a96b26-33d0-423e-a98f-5ab78139da4f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 22:07:31 crc kubenswrapper[4754]: I1005 22:07:31.451004 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dwc5t\" (UniqueName: \"kubernetes.io/projected/86a96b26-33d0-423e-a98f-5ab78139da4f-kube-api-access-dwc5t\") on node \"crc\" DevicePath \"\"" Oct 05 22:07:31 crc kubenswrapper[4754]: I1005 22:07:31.673482 4754 generic.go:334] "Generic (PLEG): container finished" podID="86a96b26-33d0-423e-a98f-5ab78139da4f" containerID="0faf273f7b052c213d73e35169e2a57861b09f024b9e4b805718eb399fff7733" exitCode=0 Oct 05 22:07:31 crc kubenswrapper[4754]: I1005 22:07:31.673570 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7wj22" event={"ID":"86a96b26-33d0-423e-a98f-5ab78139da4f","Type":"ContainerDied","Data":"0faf273f7b052c213d73e35169e2a57861b09f024b9e4b805718eb399fff7733"} Oct 05 22:07:31 crc kubenswrapper[4754]: I1005 22:07:31.673599 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7wj22" event={"ID":"86a96b26-33d0-423e-a98f-5ab78139da4f","Type":"ContainerDied","Data":"b9acdd356dbaa81ecda9a72395ce7fa1825a664c9d60fe1625c5b86dbc66daa7"} Oct 05 22:07:31 crc kubenswrapper[4754]: I1005 22:07:31.673614 4754 scope.go:117] "RemoveContainer" containerID="0faf273f7b052c213d73e35169e2a57861b09f024b9e4b805718eb399fff7733" Oct 05 22:07:31 crc kubenswrapper[4754]: I1005 22:07:31.673729 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7wj22" Oct 05 22:07:31 crc kubenswrapper[4754]: I1005 22:07:31.696713 4754 generic.go:334] "Generic (PLEG): container finished" podID="17e710fa-0b2b-4ed6-aed5-2a837e46df07" containerID="a6cbcdbb9d808a8bf7c89173dc82961850cfe9e0c76a884d6a141efd8c554aec" exitCode=0 Oct 05 22:07:31 crc kubenswrapper[4754]: I1005 22:07:31.696755 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6rmz7" event={"ID":"17e710fa-0b2b-4ed6-aed5-2a837e46df07","Type":"ContainerDied","Data":"a6cbcdbb9d808a8bf7c89173dc82961850cfe9e0c76a884d6a141efd8c554aec"} Oct 05 22:07:31 crc kubenswrapper[4754]: I1005 22:07:31.721920 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-7wj22"] Oct 05 22:07:31 crc kubenswrapper[4754]: I1005 22:07:31.736288 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-7wj22"] Oct 05 22:07:32 crc kubenswrapper[4754]: I1005 22:07:32.041578 4754 scope.go:117] "RemoveContainer" containerID="762c983cff71868008618da88e7833fa9d5f45a99796000785fc06bcae9c4dc0" Oct 05 22:07:32 crc kubenswrapper[4754]: I1005 22:07:32.319206 4754 scope.go:117] "RemoveContainer" containerID="a5a9944c21e6fe05637ac41609b07724132402cc499713be28e785cc7fb95925" Oct 05 22:07:32 crc kubenswrapper[4754]: I1005 22:07:32.352997 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6rmz7" Oct 05 22:07:32 crc kubenswrapper[4754]: I1005 22:07:32.365013 4754 scope.go:117] "RemoveContainer" containerID="0faf273f7b052c213d73e35169e2a57861b09f024b9e4b805718eb399fff7733" Oct 05 22:07:32 crc kubenswrapper[4754]: E1005 22:07:32.365426 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0faf273f7b052c213d73e35169e2a57861b09f024b9e4b805718eb399fff7733\": container with ID starting with 0faf273f7b052c213d73e35169e2a57861b09f024b9e4b805718eb399fff7733 not found: ID does not exist" containerID="0faf273f7b052c213d73e35169e2a57861b09f024b9e4b805718eb399fff7733" Oct 05 22:07:32 crc kubenswrapper[4754]: I1005 22:07:32.365466 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0faf273f7b052c213d73e35169e2a57861b09f024b9e4b805718eb399fff7733"} err="failed to get container status \"0faf273f7b052c213d73e35169e2a57861b09f024b9e4b805718eb399fff7733\": rpc error: code = NotFound desc = could not find container \"0faf273f7b052c213d73e35169e2a57861b09f024b9e4b805718eb399fff7733\": container with ID starting with 0faf273f7b052c213d73e35169e2a57861b09f024b9e4b805718eb399fff7733 not found: ID does not exist" Oct 05 22:07:32 crc kubenswrapper[4754]: I1005 22:07:32.365507 4754 scope.go:117] "RemoveContainer" containerID="762c983cff71868008618da88e7833fa9d5f45a99796000785fc06bcae9c4dc0" Oct 05 22:07:32 crc kubenswrapper[4754]: E1005 22:07:32.365761 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"762c983cff71868008618da88e7833fa9d5f45a99796000785fc06bcae9c4dc0\": container with ID starting with 762c983cff71868008618da88e7833fa9d5f45a99796000785fc06bcae9c4dc0 not found: ID does not exist" containerID="762c983cff71868008618da88e7833fa9d5f45a99796000785fc06bcae9c4dc0" Oct 05 22:07:32 crc kubenswrapper[4754]: I1005 22:07:32.365837 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"762c983cff71868008618da88e7833fa9d5f45a99796000785fc06bcae9c4dc0"} err="failed to get container status \"762c983cff71868008618da88e7833fa9d5f45a99796000785fc06bcae9c4dc0\": rpc error: code = NotFound desc = could not find container \"762c983cff71868008618da88e7833fa9d5f45a99796000785fc06bcae9c4dc0\": container with ID starting with 762c983cff71868008618da88e7833fa9d5f45a99796000785fc06bcae9c4dc0 not found: ID does not exist" Oct 05 22:07:32 crc kubenswrapper[4754]: I1005 22:07:32.365909 4754 scope.go:117] "RemoveContainer" containerID="a5a9944c21e6fe05637ac41609b07724132402cc499713be28e785cc7fb95925" Oct 05 22:07:32 crc kubenswrapper[4754]: E1005 22:07:32.366255 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a5a9944c21e6fe05637ac41609b07724132402cc499713be28e785cc7fb95925\": container with ID starting with a5a9944c21e6fe05637ac41609b07724132402cc499713be28e785cc7fb95925 not found: ID does not exist" containerID="a5a9944c21e6fe05637ac41609b07724132402cc499713be28e785cc7fb95925" Oct 05 22:07:32 crc kubenswrapper[4754]: I1005 22:07:32.366282 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a5a9944c21e6fe05637ac41609b07724132402cc499713be28e785cc7fb95925"} err="failed to get container status \"a5a9944c21e6fe05637ac41609b07724132402cc499713be28e785cc7fb95925\": rpc error: code = NotFound desc = could not find container \"a5a9944c21e6fe05637ac41609b07724132402cc499713be28e785cc7fb95925\": container with ID starting with a5a9944c21e6fe05637ac41609b07724132402cc499713be28e785cc7fb95925 not found: ID does not exist" Oct 05 22:07:32 crc kubenswrapper[4754]: I1005 22:07:32.468134 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/17e710fa-0b2b-4ed6-aed5-2a837e46df07-utilities\") pod \"17e710fa-0b2b-4ed6-aed5-2a837e46df07\" (UID: \"17e710fa-0b2b-4ed6-aed5-2a837e46df07\") " Oct 05 22:07:32 crc kubenswrapper[4754]: I1005 22:07:32.468199 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/17e710fa-0b2b-4ed6-aed5-2a837e46df07-catalog-content\") pod \"17e710fa-0b2b-4ed6-aed5-2a837e46df07\" (UID: \"17e710fa-0b2b-4ed6-aed5-2a837e46df07\") " Oct 05 22:07:32 crc kubenswrapper[4754]: I1005 22:07:32.468244 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gj5wz\" (UniqueName: \"kubernetes.io/projected/17e710fa-0b2b-4ed6-aed5-2a837e46df07-kube-api-access-gj5wz\") pod \"17e710fa-0b2b-4ed6-aed5-2a837e46df07\" (UID: \"17e710fa-0b2b-4ed6-aed5-2a837e46df07\") " Oct 05 22:07:32 crc kubenswrapper[4754]: I1005 22:07:32.469193 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/17e710fa-0b2b-4ed6-aed5-2a837e46df07-utilities" (OuterVolumeSpecName: "utilities") pod "17e710fa-0b2b-4ed6-aed5-2a837e46df07" (UID: "17e710fa-0b2b-4ed6-aed5-2a837e46df07"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 22:07:32 crc kubenswrapper[4754]: I1005 22:07:32.472192 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/17e710fa-0b2b-4ed6-aed5-2a837e46df07-kube-api-access-gj5wz" (OuterVolumeSpecName: "kube-api-access-gj5wz") pod "17e710fa-0b2b-4ed6-aed5-2a837e46df07" (UID: "17e710fa-0b2b-4ed6-aed5-2a837e46df07"). InnerVolumeSpecName "kube-api-access-gj5wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 22:07:32 crc kubenswrapper[4754]: I1005 22:07:32.480154 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/17e710fa-0b2b-4ed6-aed5-2a837e46df07-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "17e710fa-0b2b-4ed6-aed5-2a837e46df07" (UID: "17e710fa-0b2b-4ed6-aed5-2a837e46df07"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 22:07:32 crc kubenswrapper[4754]: I1005 22:07:32.569840 4754 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/17e710fa-0b2b-4ed6-aed5-2a837e46df07-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 22:07:32 crc kubenswrapper[4754]: I1005 22:07:32.569884 4754 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/17e710fa-0b2b-4ed6-aed5-2a837e46df07-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 22:07:32 crc kubenswrapper[4754]: I1005 22:07:32.569898 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gj5wz\" (UniqueName: \"kubernetes.io/projected/17e710fa-0b2b-4ed6-aed5-2a837e46df07-kube-api-access-gj5wz\") on node \"crc\" DevicePath \"\"" Oct 05 22:07:32 crc kubenswrapper[4754]: I1005 22:07:32.710261 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6rmz7" event={"ID":"17e710fa-0b2b-4ed6-aed5-2a837e46df07","Type":"ContainerDied","Data":"001c754043c79f031338c762171aefed89b42976da6580a23cc9cc661f182b6f"} Oct 05 22:07:32 crc kubenswrapper[4754]: I1005 22:07:32.710331 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6rmz7" Oct 05 22:07:32 crc kubenswrapper[4754]: I1005 22:07:32.711056 4754 scope.go:117] "RemoveContainer" containerID="a6cbcdbb9d808a8bf7c89173dc82961850cfe9e0c76a884d6a141efd8c554aec" Oct 05 22:07:32 crc kubenswrapper[4754]: I1005 22:07:32.755911 4754 scope.go:117] "RemoveContainer" containerID="b7877ab08784d556ed281df60796e1ca02d720884f309f42a30f1f4f9fa2abe1" Oct 05 22:07:32 crc kubenswrapper[4754]: I1005 22:07:32.765755 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-6rmz7"] Oct 05 22:07:32 crc kubenswrapper[4754]: I1005 22:07:32.775682 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-6rmz7"] Oct 05 22:07:32 crc kubenswrapper[4754]: I1005 22:07:32.783748 4754 scope.go:117] "RemoveContainer" containerID="ea20209750207176fb19436e0a548f337d56bf3cdc301a71457c432ba5c33dad" Oct 05 22:07:32 crc kubenswrapper[4754]: I1005 22:07:32.849217 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="17e710fa-0b2b-4ed6-aed5-2a837e46df07" path="/var/lib/kubelet/pods/17e710fa-0b2b-4ed6-aed5-2a837e46df07/volumes" Oct 05 22:07:32 crc kubenswrapper[4754]: I1005 22:07:32.850614 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="86a96b26-33d0-423e-a98f-5ab78139da4f" path="/var/lib/kubelet/pods/86a96b26-33d0-423e-a98f-5ab78139da4f/volumes" Oct 05 22:07:39 crc kubenswrapper[4754]: I1005 22:07:39.839382 4754 scope.go:117] "RemoveContainer" containerID="d925cdc080b01050e7b9377dba2f144d30f517d60778d239b53617b071498b40" Oct 05 22:07:39 crc kubenswrapper[4754]: E1005 22:07:39.840144 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 22:07:52 crc kubenswrapper[4754]: I1005 22:07:52.837212 4754 scope.go:117] "RemoveContainer" containerID="d925cdc080b01050e7b9377dba2f144d30f517d60778d239b53617b071498b40" Oct 05 22:07:52 crc kubenswrapper[4754]: E1005 22:07:52.837992 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 22:08:07 crc kubenswrapper[4754]: I1005 22:08:07.838235 4754 scope.go:117] "RemoveContainer" containerID="d925cdc080b01050e7b9377dba2f144d30f517d60778d239b53617b071498b40" Oct 05 22:08:07 crc kubenswrapper[4754]: E1005 22:08:07.838884 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 22:08:20 crc kubenswrapper[4754]: I1005 22:08:20.837292 4754 scope.go:117] "RemoveContainer" containerID="d925cdc080b01050e7b9377dba2f144d30f517d60778d239b53617b071498b40" Oct 05 22:08:20 crc kubenswrapper[4754]: E1005 22:08:20.837966 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 22:08:33 crc kubenswrapper[4754]: I1005 22:08:33.837486 4754 scope.go:117] "RemoveContainer" containerID="d925cdc080b01050e7b9377dba2f144d30f517d60778d239b53617b071498b40" Oct 05 22:08:33 crc kubenswrapper[4754]: E1005 22:08:33.839462 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 22:08:46 crc kubenswrapper[4754]: I1005 22:08:46.849292 4754 scope.go:117] "RemoveContainer" containerID="d925cdc080b01050e7b9377dba2f144d30f517d60778d239b53617b071498b40" Oct 05 22:08:46 crc kubenswrapper[4754]: E1005 22:08:46.850714 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 22:08:57 crc kubenswrapper[4754]: I1005 22:08:57.838468 4754 scope.go:117] "RemoveContainer" containerID="d925cdc080b01050e7b9377dba2f144d30f517d60778d239b53617b071498b40" Oct 05 22:08:57 crc kubenswrapper[4754]: E1005 22:08:57.839286 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 22:09:10 crc kubenswrapper[4754]: I1005 22:09:10.837875 4754 scope.go:117] "RemoveContainer" containerID="d925cdc080b01050e7b9377dba2f144d30f517d60778d239b53617b071498b40" Oct 05 22:09:10 crc kubenswrapper[4754]: E1005 22:09:10.839010 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 22:09:18 crc kubenswrapper[4754]: I1005 22:09:18.889371 4754 generic.go:334] "Generic (PLEG): container finished" podID="5922ff4d-2ec1-4156-acc7-8e6427819b3e" containerID="a7fc76a07a2380639510193b4c0f81120c5ee9beae1678cdbb5acac37cb15675" exitCode=0 Oct 05 22:09:18 crc kubenswrapper[4754]: I1005 22:09:18.889458 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"5922ff4d-2ec1-4156-acc7-8e6427819b3e","Type":"ContainerDied","Data":"a7fc76a07a2380639510193b4c0f81120c5ee9beae1678cdbb5acac37cb15675"} Oct 05 22:09:20 crc kubenswrapper[4754]: I1005 22:09:20.267036 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Oct 05 22:09:20 crc kubenswrapper[4754]: I1005 22:09:20.425522 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/5922ff4d-2ec1-4156-acc7-8e6427819b3e-test-operator-ephemeral-temporary\") pod \"5922ff4d-2ec1-4156-acc7-8e6427819b3e\" (UID: \"5922ff4d-2ec1-4156-acc7-8e6427819b3e\") " Oct 05 22:09:20 crc kubenswrapper[4754]: I1005 22:09:20.425572 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"5922ff4d-2ec1-4156-acc7-8e6427819b3e\" (UID: \"5922ff4d-2ec1-4156-acc7-8e6427819b3e\") " Oct 05 22:09:20 crc kubenswrapper[4754]: I1005 22:09:20.425596 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/5922ff4d-2ec1-4156-acc7-8e6427819b3e-test-operator-ephemeral-workdir\") pod \"5922ff4d-2ec1-4156-acc7-8e6427819b3e\" (UID: \"5922ff4d-2ec1-4156-acc7-8e6427819b3e\") " Oct 05 22:09:20 crc kubenswrapper[4754]: I1005 22:09:20.425630 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/5922ff4d-2ec1-4156-acc7-8e6427819b3e-openstack-config-secret\") pod \"5922ff4d-2ec1-4156-acc7-8e6427819b3e\" (UID: \"5922ff4d-2ec1-4156-acc7-8e6427819b3e\") " Oct 05 22:09:20 crc kubenswrapper[4754]: I1005 22:09:20.425827 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5922ff4d-2ec1-4156-acc7-8e6427819b3e-config-data\") pod \"5922ff4d-2ec1-4156-acc7-8e6427819b3e\" (UID: \"5922ff4d-2ec1-4156-acc7-8e6427819b3e\") " Oct 05 22:09:20 crc kubenswrapper[4754]: I1005 22:09:20.425868 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5922ff4d-2ec1-4156-acc7-8e6427819b3e-ssh-key\") pod \"5922ff4d-2ec1-4156-acc7-8e6427819b3e\" (UID: \"5922ff4d-2ec1-4156-acc7-8e6427819b3e\") " Oct 05 22:09:20 crc kubenswrapper[4754]: I1005 22:09:20.425894 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/5922ff4d-2ec1-4156-acc7-8e6427819b3e-openstack-config\") pod \"5922ff4d-2ec1-4156-acc7-8e6427819b3e\" (UID: \"5922ff4d-2ec1-4156-acc7-8e6427819b3e\") " Oct 05 22:09:20 crc kubenswrapper[4754]: I1005 22:09:20.425990 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x5j8w\" (UniqueName: \"kubernetes.io/projected/5922ff4d-2ec1-4156-acc7-8e6427819b3e-kube-api-access-x5j8w\") pod \"5922ff4d-2ec1-4156-acc7-8e6427819b3e\" (UID: \"5922ff4d-2ec1-4156-acc7-8e6427819b3e\") " Oct 05 22:09:20 crc kubenswrapper[4754]: I1005 22:09:20.426063 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/5922ff4d-2ec1-4156-acc7-8e6427819b3e-ca-certs\") pod \"5922ff4d-2ec1-4156-acc7-8e6427819b3e\" (UID: \"5922ff4d-2ec1-4156-acc7-8e6427819b3e\") " Oct 05 22:09:20 crc kubenswrapper[4754]: I1005 22:09:20.428850 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5922ff4d-2ec1-4156-acc7-8e6427819b3e-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "5922ff4d-2ec1-4156-acc7-8e6427819b3e" (UID: "5922ff4d-2ec1-4156-acc7-8e6427819b3e"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 22:09:20 crc kubenswrapper[4754]: I1005 22:09:20.429554 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5922ff4d-2ec1-4156-acc7-8e6427819b3e-config-data" (OuterVolumeSpecName: "config-data") pod "5922ff4d-2ec1-4156-acc7-8e6427819b3e" (UID: "5922ff4d-2ec1-4156-acc7-8e6427819b3e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 22:09:20 crc kubenswrapper[4754]: I1005 22:09:20.432863 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5922ff4d-2ec1-4156-acc7-8e6427819b3e-kube-api-access-x5j8w" (OuterVolumeSpecName: "kube-api-access-x5j8w") pod "5922ff4d-2ec1-4156-acc7-8e6427819b3e" (UID: "5922ff4d-2ec1-4156-acc7-8e6427819b3e"). InnerVolumeSpecName "kube-api-access-x5j8w". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 22:09:20 crc kubenswrapper[4754]: I1005 22:09:20.433303 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5922ff4d-2ec1-4156-acc7-8e6427819b3e-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "5922ff4d-2ec1-4156-acc7-8e6427819b3e" (UID: "5922ff4d-2ec1-4156-acc7-8e6427819b3e"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 22:09:20 crc kubenswrapper[4754]: I1005 22:09:20.435010 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "test-operator-logs") pod "5922ff4d-2ec1-4156-acc7-8e6427819b3e" (UID: "5922ff4d-2ec1-4156-acc7-8e6427819b3e"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 05 22:09:20 crc kubenswrapper[4754]: I1005 22:09:20.457286 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5922ff4d-2ec1-4156-acc7-8e6427819b3e-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "5922ff4d-2ec1-4156-acc7-8e6427819b3e" (UID: "5922ff4d-2ec1-4156-acc7-8e6427819b3e"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 22:09:20 crc kubenswrapper[4754]: I1005 22:09:20.459231 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5922ff4d-2ec1-4156-acc7-8e6427819b3e-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "5922ff4d-2ec1-4156-acc7-8e6427819b3e" (UID: "5922ff4d-2ec1-4156-acc7-8e6427819b3e"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 22:09:20 crc kubenswrapper[4754]: I1005 22:09:20.460130 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5922ff4d-2ec1-4156-acc7-8e6427819b3e-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "5922ff4d-2ec1-4156-acc7-8e6427819b3e" (UID: "5922ff4d-2ec1-4156-acc7-8e6427819b3e"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 22:09:20 crc kubenswrapper[4754]: I1005 22:09:20.528813 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x5j8w\" (UniqueName: \"kubernetes.io/projected/5922ff4d-2ec1-4156-acc7-8e6427819b3e-kube-api-access-x5j8w\") on node \"crc\" DevicePath \"\"" Oct 05 22:09:20 crc kubenswrapper[4754]: I1005 22:09:20.528846 4754 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/5922ff4d-2ec1-4156-acc7-8e6427819b3e-ca-certs\") on node \"crc\" DevicePath \"\"" Oct 05 22:09:20 crc kubenswrapper[4754]: I1005 22:09:20.528859 4754 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/5922ff4d-2ec1-4156-acc7-8e6427819b3e-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Oct 05 22:09:20 crc kubenswrapper[4754]: I1005 22:09:20.530227 4754 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Oct 05 22:09:20 crc kubenswrapper[4754]: I1005 22:09:20.530254 4754 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/5922ff4d-2ec1-4156-acc7-8e6427819b3e-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Oct 05 22:09:20 crc kubenswrapper[4754]: I1005 22:09:20.530268 4754 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/5922ff4d-2ec1-4156-acc7-8e6427819b3e-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Oct 05 22:09:20 crc kubenswrapper[4754]: I1005 22:09:20.530284 4754 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5922ff4d-2ec1-4156-acc7-8e6427819b3e-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 22:09:20 crc kubenswrapper[4754]: I1005 22:09:20.530295 4754 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5922ff4d-2ec1-4156-acc7-8e6427819b3e-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 05 22:09:20 crc kubenswrapper[4754]: I1005 22:09:20.920652 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Oct 05 22:09:21 crc kubenswrapper[4754]: I1005 22:09:21.105962 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5922ff4d-2ec1-4156-acc7-8e6427819b3e-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "5922ff4d-2ec1-4156-acc7-8e6427819b3e" (UID: "5922ff4d-2ec1-4156-acc7-8e6427819b3e"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 22:09:21 crc kubenswrapper[4754]: I1005 22:09:21.129402 4754 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Oct 05 22:09:21 crc kubenswrapper[4754]: I1005 22:09:21.144067 4754 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Oct 05 22:09:21 crc kubenswrapper[4754]: I1005 22:09:21.144116 4754 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/5922ff4d-2ec1-4156-acc7-8e6427819b3e-openstack-config\") on node \"crc\" DevicePath \"\"" Oct 05 22:09:21 crc kubenswrapper[4754]: I1005 22:09:21.199427 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"5922ff4d-2ec1-4156-acc7-8e6427819b3e","Type":"ContainerDied","Data":"e2712f6081e90acaba2b35a7652cc92d1a43dea8c02a3267d9c8e5c3f5438972"} Oct 05 22:09:21 crc kubenswrapper[4754]: I1005 22:09:21.199614 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e2712f6081e90acaba2b35a7652cc92d1a43dea8c02a3267d9c8e5c3f5438972" Oct 05 22:09:25 crc kubenswrapper[4754]: I1005 22:09:25.837679 4754 scope.go:117] "RemoveContainer" containerID="d925cdc080b01050e7b9377dba2f144d30f517d60778d239b53617b071498b40" Oct 05 22:09:25 crc kubenswrapper[4754]: E1005 22:09:25.838623 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 22:09:33 crc kubenswrapper[4754]: I1005 22:09:33.039245 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Oct 05 22:09:33 crc kubenswrapper[4754]: E1005 22:09:33.040433 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86a96b26-33d0-423e-a98f-5ab78139da4f" containerName="extract-utilities" Oct 05 22:09:33 crc kubenswrapper[4754]: I1005 22:09:33.040458 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="86a96b26-33d0-423e-a98f-5ab78139da4f" containerName="extract-utilities" Oct 05 22:09:33 crc kubenswrapper[4754]: E1005 22:09:33.040488 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17e710fa-0b2b-4ed6-aed5-2a837e46df07" containerName="registry-server" Oct 05 22:09:33 crc kubenswrapper[4754]: I1005 22:09:33.040523 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="17e710fa-0b2b-4ed6-aed5-2a837e46df07" containerName="registry-server" Oct 05 22:09:33 crc kubenswrapper[4754]: E1005 22:09:33.040547 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17e710fa-0b2b-4ed6-aed5-2a837e46df07" containerName="extract-utilities" Oct 05 22:09:33 crc kubenswrapper[4754]: I1005 22:09:33.040559 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="17e710fa-0b2b-4ed6-aed5-2a837e46df07" containerName="extract-utilities" Oct 05 22:09:33 crc kubenswrapper[4754]: E1005 22:09:33.040586 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86a96b26-33d0-423e-a98f-5ab78139da4f" containerName="extract-content" Oct 05 22:09:33 crc kubenswrapper[4754]: I1005 22:09:33.040598 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="86a96b26-33d0-423e-a98f-5ab78139da4f" containerName="extract-content" Oct 05 22:09:33 crc kubenswrapper[4754]: E1005 22:09:33.040628 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5922ff4d-2ec1-4156-acc7-8e6427819b3e" containerName="tempest-tests-tempest-tests-runner" Oct 05 22:09:33 crc kubenswrapper[4754]: I1005 22:09:33.040641 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="5922ff4d-2ec1-4156-acc7-8e6427819b3e" containerName="tempest-tests-tempest-tests-runner" Oct 05 22:09:33 crc kubenswrapper[4754]: E1005 22:09:33.040661 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17e710fa-0b2b-4ed6-aed5-2a837e46df07" containerName="extract-content" Oct 05 22:09:33 crc kubenswrapper[4754]: I1005 22:09:33.040673 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="17e710fa-0b2b-4ed6-aed5-2a837e46df07" containerName="extract-content" Oct 05 22:09:33 crc kubenswrapper[4754]: E1005 22:09:33.040705 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86a96b26-33d0-423e-a98f-5ab78139da4f" containerName="registry-server" Oct 05 22:09:33 crc kubenswrapper[4754]: I1005 22:09:33.040717 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="86a96b26-33d0-423e-a98f-5ab78139da4f" containerName="registry-server" Oct 05 22:09:33 crc kubenswrapper[4754]: I1005 22:09:33.041010 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="5922ff4d-2ec1-4156-acc7-8e6427819b3e" containerName="tempest-tests-tempest-tests-runner" Oct 05 22:09:33 crc kubenswrapper[4754]: I1005 22:09:33.041057 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="17e710fa-0b2b-4ed6-aed5-2a837e46df07" containerName="registry-server" Oct 05 22:09:33 crc kubenswrapper[4754]: I1005 22:09:33.041086 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="86a96b26-33d0-423e-a98f-5ab78139da4f" containerName="registry-server" Oct 05 22:09:33 crc kubenswrapper[4754]: I1005 22:09:33.042104 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 05 22:09:33 crc kubenswrapper[4754]: I1005 22:09:33.061185 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-dx5w8" Oct 05 22:09:33 crc kubenswrapper[4754]: I1005 22:09:33.063903 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Oct 05 22:09:33 crc kubenswrapper[4754]: I1005 22:09:33.114479 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cq278\" (UniqueName: \"kubernetes.io/projected/ea8cb295-85b1-4afe-8afe-6d5f19b25f28-kube-api-access-cq278\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"ea8cb295-85b1-4afe-8afe-6d5f19b25f28\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 05 22:09:33 crc kubenswrapper[4754]: I1005 22:09:33.114660 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"ea8cb295-85b1-4afe-8afe-6d5f19b25f28\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 05 22:09:33 crc kubenswrapper[4754]: I1005 22:09:33.216891 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cq278\" (UniqueName: \"kubernetes.io/projected/ea8cb295-85b1-4afe-8afe-6d5f19b25f28-kube-api-access-cq278\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"ea8cb295-85b1-4afe-8afe-6d5f19b25f28\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 05 22:09:33 crc kubenswrapper[4754]: I1005 22:09:33.217010 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"ea8cb295-85b1-4afe-8afe-6d5f19b25f28\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 05 22:09:33 crc kubenswrapper[4754]: I1005 22:09:33.218978 4754 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"ea8cb295-85b1-4afe-8afe-6d5f19b25f28\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 05 22:09:33 crc kubenswrapper[4754]: I1005 22:09:33.245392 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cq278\" (UniqueName: \"kubernetes.io/projected/ea8cb295-85b1-4afe-8afe-6d5f19b25f28-kube-api-access-cq278\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"ea8cb295-85b1-4afe-8afe-6d5f19b25f28\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 05 22:09:33 crc kubenswrapper[4754]: I1005 22:09:33.263288 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"ea8cb295-85b1-4afe-8afe-6d5f19b25f28\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 05 22:09:33 crc kubenswrapper[4754]: I1005 22:09:33.383778 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 05 22:09:33 crc kubenswrapper[4754]: I1005 22:09:33.907447 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Oct 05 22:09:34 crc kubenswrapper[4754]: I1005 22:09:34.090698 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"ea8cb295-85b1-4afe-8afe-6d5f19b25f28","Type":"ContainerStarted","Data":"16f1fd39131e357b47f0ea6ba8c407f4aefd9efe4e78eb692110c90de32fe736"} Oct 05 22:09:36 crc kubenswrapper[4754]: I1005 22:09:36.115775 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"ea8cb295-85b1-4afe-8afe-6d5f19b25f28","Type":"ContainerStarted","Data":"00e0e941e9b7de0d408933a9be05c473cfc56b7dcfeed89a74a5f7091717041c"} Oct 05 22:09:36 crc kubenswrapper[4754]: I1005 22:09:36.133478 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=1.443197764 podStartE2EDuration="3.133461172s" podCreationTimestamp="2025-10-05 22:09:33 +0000 UTC" firstStartedPulling="2025-10-05 22:09:33.915322377 +0000 UTC m=+4497.819441127" lastFinishedPulling="2025-10-05 22:09:35.605585785 +0000 UTC m=+4499.509704535" observedRunningTime="2025-10-05 22:09:36.127826989 +0000 UTC m=+4500.031945699" watchObservedRunningTime="2025-10-05 22:09:36.133461172 +0000 UTC m=+4500.037579892" Oct 05 22:09:38 crc kubenswrapper[4754]: I1005 22:09:38.837813 4754 scope.go:117] "RemoveContainer" containerID="d925cdc080b01050e7b9377dba2f144d30f517d60778d239b53617b071498b40" Oct 05 22:09:38 crc kubenswrapper[4754]: E1005 22:09:38.838686 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 22:09:50 crc kubenswrapper[4754]: I1005 22:09:50.838480 4754 scope.go:117] "RemoveContainer" containerID="d925cdc080b01050e7b9377dba2f144d30f517d60778d239b53617b071498b40" Oct 05 22:09:50 crc kubenswrapper[4754]: E1005 22:09:50.839927 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 22:09:52 crc kubenswrapper[4754]: I1005 22:09:52.573662 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-44bn6/must-gather-l24vb"] Oct 05 22:09:52 crc kubenswrapper[4754]: I1005 22:09:52.577054 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-44bn6/must-gather-l24vb" Oct 05 22:09:52 crc kubenswrapper[4754]: I1005 22:09:52.581655 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-44bn6/must-gather-l24vb"] Oct 05 22:09:52 crc kubenswrapper[4754]: I1005 22:09:52.583310 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-44bn6"/"openshift-service-ca.crt" Oct 05 22:09:52 crc kubenswrapper[4754]: I1005 22:09:52.583467 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-44bn6"/"kube-root-ca.crt" Oct 05 22:09:52 crc kubenswrapper[4754]: I1005 22:09:52.583609 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-44bn6"/"default-dockercfg-kpxsz" Oct 05 22:09:52 crc kubenswrapper[4754]: I1005 22:09:52.631339 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wrxg2\" (UniqueName: \"kubernetes.io/projected/5d9dc270-0dc5-4682-a4cf-1b61e0e3390b-kube-api-access-wrxg2\") pod \"must-gather-l24vb\" (UID: \"5d9dc270-0dc5-4682-a4cf-1b61e0e3390b\") " pod="openshift-must-gather-44bn6/must-gather-l24vb" Oct 05 22:09:52 crc kubenswrapper[4754]: I1005 22:09:52.631525 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/5d9dc270-0dc5-4682-a4cf-1b61e0e3390b-must-gather-output\") pod \"must-gather-l24vb\" (UID: \"5d9dc270-0dc5-4682-a4cf-1b61e0e3390b\") " pod="openshift-must-gather-44bn6/must-gather-l24vb" Oct 05 22:09:52 crc kubenswrapper[4754]: I1005 22:09:52.732963 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/5d9dc270-0dc5-4682-a4cf-1b61e0e3390b-must-gather-output\") pod \"must-gather-l24vb\" (UID: \"5d9dc270-0dc5-4682-a4cf-1b61e0e3390b\") " pod="openshift-must-gather-44bn6/must-gather-l24vb" Oct 05 22:09:52 crc kubenswrapper[4754]: I1005 22:09:52.733129 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wrxg2\" (UniqueName: \"kubernetes.io/projected/5d9dc270-0dc5-4682-a4cf-1b61e0e3390b-kube-api-access-wrxg2\") pod \"must-gather-l24vb\" (UID: \"5d9dc270-0dc5-4682-a4cf-1b61e0e3390b\") " pod="openshift-must-gather-44bn6/must-gather-l24vb" Oct 05 22:09:52 crc kubenswrapper[4754]: I1005 22:09:52.733384 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/5d9dc270-0dc5-4682-a4cf-1b61e0e3390b-must-gather-output\") pod \"must-gather-l24vb\" (UID: \"5d9dc270-0dc5-4682-a4cf-1b61e0e3390b\") " pod="openshift-must-gather-44bn6/must-gather-l24vb" Oct 05 22:09:52 crc kubenswrapper[4754]: I1005 22:09:52.749142 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wrxg2\" (UniqueName: \"kubernetes.io/projected/5d9dc270-0dc5-4682-a4cf-1b61e0e3390b-kube-api-access-wrxg2\") pod \"must-gather-l24vb\" (UID: \"5d9dc270-0dc5-4682-a4cf-1b61e0e3390b\") " pod="openshift-must-gather-44bn6/must-gather-l24vb" Oct 05 22:09:52 crc kubenswrapper[4754]: I1005 22:09:52.905101 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-44bn6/must-gather-l24vb" Oct 05 22:09:53 crc kubenswrapper[4754]: I1005 22:09:53.536112 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-44bn6/must-gather-l24vb"] Oct 05 22:09:53 crc kubenswrapper[4754]: W1005 22:09:53.543725 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5d9dc270_0dc5_4682_a4cf_1b61e0e3390b.slice/crio-95cef2e33c4505a3d6f721cedc4969342c294330be961eee67d951b32beb3c2b WatchSource:0}: Error finding container 95cef2e33c4505a3d6f721cedc4969342c294330be961eee67d951b32beb3c2b: Status 404 returned error can't find the container with id 95cef2e33c4505a3d6f721cedc4969342c294330be961eee67d951b32beb3c2b Oct 05 22:09:53 crc kubenswrapper[4754]: I1005 22:09:53.546341 4754 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 05 22:09:54 crc kubenswrapper[4754]: I1005 22:09:54.361820 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-44bn6/must-gather-l24vb" event={"ID":"5d9dc270-0dc5-4682-a4cf-1b61e0e3390b","Type":"ContainerStarted","Data":"95cef2e33c4505a3d6f721cedc4969342c294330be961eee67d951b32beb3c2b"} Oct 05 22:09:59 crc kubenswrapper[4754]: I1005 22:09:59.405662 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-44bn6/must-gather-l24vb" event={"ID":"5d9dc270-0dc5-4682-a4cf-1b61e0e3390b","Type":"ContainerStarted","Data":"1eae1877e73e7e6bc401df6113094f06180449c379bbdd6e8951cce4dfa749a3"} Oct 05 22:09:59 crc kubenswrapper[4754]: I1005 22:09:59.406396 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-44bn6/must-gather-l24vb" event={"ID":"5d9dc270-0dc5-4682-a4cf-1b61e0e3390b","Type":"ContainerStarted","Data":"79c7a645c7b390dae565f4af0e443592f8f1620a90d1c32ab67f31639e2c89f3"} Oct 05 22:09:59 crc kubenswrapper[4754]: I1005 22:09:59.429730 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-44bn6/must-gather-l24vb" podStartSLOduration=3.367084488 podStartE2EDuration="7.429711642s" podCreationTimestamp="2025-10-05 22:09:52 +0000 UTC" firstStartedPulling="2025-10-05 22:09:53.546123774 +0000 UTC m=+4517.450242484" lastFinishedPulling="2025-10-05 22:09:57.608750928 +0000 UTC m=+4521.512869638" observedRunningTime="2025-10-05 22:09:59.428656325 +0000 UTC m=+4523.332775075" watchObservedRunningTime="2025-10-05 22:09:59.429711642 +0000 UTC m=+4523.333830352" Oct 05 22:10:03 crc kubenswrapper[4754]: I1005 22:10:03.837355 4754 scope.go:117] "RemoveContainer" containerID="d925cdc080b01050e7b9377dba2f144d30f517d60778d239b53617b071498b40" Oct 05 22:10:03 crc kubenswrapper[4754]: E1005 22:10:03.837863 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 22:10:03 crc kubenswrapper[4754]: I1005 22:10:03.937131 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-44bn6/crc-debug-cspl7"] Oct 05 22:10:03 crc kubenswrapper[4754]: I1005 22:10:03.938175 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-44bn6/crc-debug-cspl7" Oct 05 22:10:04 crc kubenswrapper[4754]: I1005 22:10:04.052472 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/fecb08dd-672e-4857-ba0c-198a6efa07d5-host\") pod \"crc-debug-cspl7\" (UID: \"fecb08dd-672e-4857-ba0c-198a6efa07d5\") " pod="openshift-must-gather-44bn6/crc-debug-cspl7" Oct 05 22:10:04 crc kubenswrapper[4754]: I1005 22:10:04.052546 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ldl2v\" (UniqueName: \"kubernetes.io/projected/fecb08dd-672e-4857-ba0c-198a6efa07d5-kube-api-access-ldl2v\") pod \"crc-debug-cspl7\" (UID: \"fecb08dd-672e-4857-ba0c-198a6efa07d5\") " pod="openshift-must-gather-44bn6/crc-debug-cspl7" Oct 05 22:10:04 crc kubenswrapper[4754]: I1005 22:10:04.154704 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/fecb08dd-672e-4857-ba0c-198a6efa07d5-host\") pod \"crc-debug-cspl7\" (UID: \"fecb08dd-672e-4857-ba0c-198a6efa07d5\") " pod="openshift-must-gather-44bn6/crc-debug-cspl7" Oct 05 22:10:04 crc kubenswrapper[4754]: I1005 22:10:04.154982 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ldl2v\" (UniqueName: \"kubernetes.io/projected/fecb08dd-672e-4857-ba0c-198a6efa07d5-kube-api-access-ldl2v\") pod \"crc-debug-cspl7\" (UID: \"fecb08dd-672e-4857-ba0c-198a6efa07d5\") " pod="openshift-must-gather-44bn6/crc-debug-cspl7" Oct 05 22:10:04 crc kubenswrapper[4754]: I1005 22:10:04.154876 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/fecb08dd-672e-4857-ba0c-198a6efa07d5-host\") pod \"crc-debug-cspl7\" (UID: \"fecb08dd-672e-4857-ba0c-198a6efa07d5\") " pod="openshift-must-gather-44bn6/crc-debug-cspl7" Oct 05 22:10:04 crc kubenswrapper[4754]: I1005 22:10:04.174502 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ldl2v\" (UniqueName: \"kubernetes.io/projected/fecb08dd-672e-4857-ba0c-198a6efa07d5-kube-api-access-ldl2v\") pod \"crc-debug-cspl7\" (UID: \"fecb08dd-672e-4857-ba0c-198a6efa07d5\") " pod="openshift-must-gather-44bn6/crc-debug-cspl7" Oct 05 22:10:04 crc kubenswrapper[4754]: I1005 22:10:04.252425 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-44bn6/crc-debug-cspl7" Oct 05 22:10:04 crc kubenswrapper[4754]: I1005 22:10:04.456381 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-44bn6/crc-debug-cspl7" event={"ID":"fecb08dd-672e-4857-ba0c-198a6efa07d5","Type":"ContainerStarted","Data":"c32c4e9ae117860f28c917869c986a2b739d898cd4b2f274f9713ede5852452e"} Oct 05 22:10:14 crc kubenswrapper[4754]: I1005 22:10:14.837708 4754 scope.go:117] "RemoveContainer" containerID="d925cdc080b01050e7b9377dba2f144d30f517d60778d239b53617b071498b40" Oct 05 22:10:14 crc kubenswrapper[4754]: E1005 22:10:14.838417 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 22:10:15 crc kubenswrapper[4754]: I1005 22:10:15.554692 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-44bn6/crc-debug-cspl7" event={"ID":"fecb08dd-672e-4857-ba0c-198a6efa07d5","Type":"ContainerStarted","Data":"af5377fba7a759eff76ab1fcecce4f748f0fa0b79bd410cb77230eef9b5204e5"} Oct 05 22:10:15 crc kubenswrapper[4754]: I1005 22:10:15.566417 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-44bn6/crc-debug-cspl7" podStartSLOduration=1.88044522 podStartE2EDuration="12.566402713s" podCreationTimestamp="2025-10-05 22:10:03 +0000 UTC" firstStartedPulling="2025-10-05 22:10:04.281097234 +0000 UTC m=+4528.185215944" lastFinishedPulling="2025-10-05 22:10:14.967054727 +0000 UTC m=+4538.871173437" observedRunningTime="2025-10-05 22:10:15.565096929 +0000 UTC m=+4539.469215649" watchObservedRunningTime="2025-10-05 22:10:15.566402713 +0000 UTC m=+4539.470521423" Oct 05 22:10:26 crc kubenswrapper[4754]: I1005 22:10:26.844824 4754 scope.go:117] "RemoveContainer" containerID="d925cdc080b01050e7b9377dba2f144d30f517d60778d239b53617b071498b40" Oct 05 22:10:26 crc kubenswrapper[4754]: E1005 22:10:26.845697 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 22:10:41 crc kubenswrapper[4754]: I1005 22:10:41.837813 4754 scope.go:117] "RemoveContainer" containerID="d925cdc080b01050e7b9377dba2f144d30f517d60778d239b53617b071498b40" Oct 05 22:10:41 crc kubenswrapper[4754]: E1005 22:10:41.838532 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 22:10:55 crc kubenswrapper[4754]: I1005 22:10:55.837383 4754 scope.go:117] "RemoveContainer" containerID="d925cdc080b01050e7b9377dba2f144d30f517d60778d239b53617b071498b40" Oct 05 22:10:55 crc kubenswrapper[4754]: E1005 22:10:55.838975 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 22:11:09 crc kubenswrapper[4754]: I1005 22:11:09.837389 4754 scope.go:117] "RemoveContainer" containerID="d925cdc080b01050e7b9377dba2f144d30f517d60778d239b53617b071498b40" Oct 05 22:11:09 crc kubenswrapper[4754]: E1005 22:11:09.838100 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 22:11:22 crc kubenswrapper[4754]: I1005 22:11:22.837170 4754 scope.go:117] "RemoveContainer" containerID="d925cdc080b01050e7b9377dba2f144d30f517d60778d239b53617b071498b40" Oct 05 22:11:22 crc kubenswrapper[4754]: E1005 22:11:22.837894 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 22:11:27 crc kubenswrapper[4754]: I1005 22:11:27.947125 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-765c6857c8-q7glx_bd23550d-8815-4211-b8be-10ee8ab95c1f/barbican-api/0.log" Oct 05 22:11:28 crc kubenswrapper[4754]: I1005 22:11:28.013047 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-765c6857c8-q7glx_bd23550d-8815-4211-b8be-10ee8ab95c1f/barbican-api-log/0.log" Oct 05 22:11:28 crc kubenswrapper[4754]: I1005 22:11:28.187997 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-9dc4bfc9d-cfm98_358e33ff-b4ce-46e3-b077-e1df373f2a75/barbican-keystone-listener/0.log" Oct 05 22:11:28 crc kubenswrapper[4754]: I1005 22:11:28.340565 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-9dc4bfc9d-cfm98_358e33ff-b4ce-46e3-b077-e1df373f2a75/barbican-keystone-listener-log/0.log" Oct 05 22:11:28 crc kubenswrapper[4754]: I1005 22:11:28.491448 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-54b7554787-vqrt4_346d8b39-98ea-442d-910e-b40a6e4b85c4/barbican-worker/0.log" Oct 05 22:11:28 crc kubenswrapper[4754]: I1005 22:11:28.583680 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-54b7554787-vqrt4_346d8b39-98ea-442d-910e-b40a6e4b85c4/barbican-worker-log/0.log" Oct 05 22:11:28 crc kubenswrapper[4754]: I1005 22:11:28.789170 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-6cdph_31cbd0a7-7e3d-4bd6-b84c-816431fdc633/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Oct 05 22:11:28 crc kubenswrapper[4754]: I1005 22:11:28.970848 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_b5010900-7b57-48e9-9621-73940c2aefa1/ceilometer-central-agent/0.log" Oct 05 22:11:29 crc kubenswrapper[4754]: I1005 22:11:29.040095 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_b5010900-7b57-48e9-9621-73940c2aefa1/ceilometer-notification-agent/0.log" Oct 05 22:11:29 crc kubenswrapper[4754]: I1005 22:11:29.066688 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_b5010900-7b57-48e9-9621-73940c2aefa1/proxy-httpd/0.log" Oct 05 22:11:29 crc kubenswrapper[4754]: I1005 22:11:29.184172 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_b5010900-7b57-48e9-9621-73940c2aefa1/sg-core/0.log" Oct 05 22:11:29 crc kubenswrapper[4754]: I1005 22:11:29.339325 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_90277df3-c655-423b-b6a6-8f3142ab046e/cinder-api/0.log" Oct 05 22:11:29 crc kubenswrapper[4754]: I1005 22:11:29.409665 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_90277df3-c655-423b-b6a6-8f3142ab046e/cinder-api-log/0.log" Oct 05 22:11:29 crc kubenswrapper[4754]: I1005 22:11:29.555959 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_8da021ab-a4b1-4a42-bf58-2973a0578d8e/cinder-scheduler/0.log" Oct 05 22:11:29 crc kubenswrapper[4754]: I1005 22:11:29.652622 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_8da021ab-a4b1-4a42-bf58-2973a0578d8e/probe/0.log" Oct 05 22:11:29 crc kubenswrapper[4754]: I1005 22:11:29.810725 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-7xfdc_5421db8c-f81e-483e-ac15-c0b725a2c277/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Oct 05 22:11:30 crc kubenswrapper[4754]: I1005 22:11:30.116270 4754 scope.go:117] "RemoveContainer" containerID="864eefaec67a4ddbffc7825982742ea19bfdfd2ae0eac0214685f432aa4a7bcd" Oct 05 22:11:30 crc kubenswrapper[4754]: I1005 22:11:30.156430 4754 scope.go:117] "RemoveContainer" containerID="2e4abf8237938108950b4975f3e27a5c435dcfbcd3106a539e9922ef2c918309" Oct 05 22:11:30 crc kubenswrapper[4754]: I1005 22:11:30.198962 4754 scope.go:117] "RemoveContainer" containerID="d3a1254ee9a93016c7556f35688d705e4ebbeede45a25ddbda0cc28182b368f9" Oct 05 22:11:30 crc kubenswrapper[4754]: I1005 22:11:30.446533 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-k54sd_77f69424-dcfc-4915-b704-3447e3cf31eb/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 05 22:11:30 crc kubenswrapper[4754]: I1005 22:11:30.455405 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-p2gbk_50e8c589-a949-4193-94a1-022fea47e2ba/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 05 22:11:30 crc kubenswrapper[4754]: I1005 22:11:30.685317 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-79dc84bdb7-ngcb6_f031b1f6-304e-4419-a4b2-97a8415a530a/init/0.log" Oct 05 22:11:30 crc kubenswrapper[4754]: I1005 22:11:30.900238 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-79dc84bdb7-ngcb6_f031b1f6-304e-4419-a4b2-97a8415a530a/init/0.log" Oct 05 22:11:31 crc kubenswrapper[4754]: I1005 22:11:31.053999 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-79dc84bdb7-ngcb6_f031b1f6-304e-4419-a4b2-97a8415a530a/dnsmasq-dns/0.log" Oct 05 22:11:31 crc kubenswrapper[4754]: I1005 22:11:31.193639 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-5x9f9_920f98be-147d-4b28-a82c-fa77c86fd8f1/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Oct 05 22:11:31 crc kubenswrapper[4754]: I1005 22:11:31.253745 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_49d68627-98f5-4eec-b544-1de274b6e786/glance-httpd/0.log" Oct 05 22:11:32 crc kubenswrapper[4754]: I1005 22:11:32.026383 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_49d68627-98f5-4eec-b544-1de274b6e786/glance-log/0.log" Oct 05 22:11:32 crc kubenswrapper[4754]: I1005 22:11:32.062748 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_b378ce06-065f-4d9e-80d2-d98d834a2556/glance-log/0.log" Oct 05 22:11:32 crc kubenswrapper[4754]: I1005 22:11:32.064534 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_b378ce06-065f-4d9e-80d2-d98d834a2556/glance-httpd/0.log" Oct 05 22:11:32 crc kubenswrapper[4754]: I1005 22:11:32.266655 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-574454d6cb-vbnk6_39d1aac3-14a1-45bf-9a9d-ba4cdb0afcb1/horizon/3.log" Oct 05 22:11:32 crc kubenswrapper[4754]: I1005 22:11:32.315212 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-574454d6cb-vbnk6_39d1aac3-14a1-45bf-9a9d-ba4cdb0afcb1/horizon/2.log" Oct 05 22:11:32 crc kubenswrapper[4754]: I1005 22:11:32.553790 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-ch8jw_27e32907-804b-43ba-ad6b-741e84eee1ab/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Oct 05 22:11:32 crc kubenswrapper[4754]: I1005 22:11:32.689604 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-574454d6cb-vbnk6_39d1aac3-14a1-45bf-9a9d-ba4cdb0afcb1/horizon-log/0.log" Oct 05 22:11:32 crc kubenswrapper[4754]: I1005 22:11:32.789348 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-csfdg_35d09fe3-0746-4e6c-82b2-cc5bef030483/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 05 22:11:33 crc kubenswrapper[4754]: I1005 22:11:33.067828 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29328361-wb6sz_fc91c348-ca21-48c0-a694-f9850277e59b/keystone-cron/0.log" Oct 05 22:11:33 crc kubenswrapper[4754]: I1005 22:11:33.161796 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_b54ee63c-a198-4855-a04c-fc753939dbc0/kube-state-metrics/0.log" Oct 05 22:11:33 crc kubenswrapper[4754]: I1005 22:11:33.219782 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-55bcb5dcdd-jhgrw_d4c2091b-e2d2-4861-8190-79d548039f18/keystone-api/0.log" Oct 05 22:11:33 crc kubenswrapper[4754]: I1005 22:11:33.363257 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-c7xwk_43f1b88f-48f4-4ee0-9fcd-8f53bee2f461/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Oct 05 22:11:33 crc kubenswrapper[4754]: I1005 22:11:33.996655 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-5d6f8484b5-vstqw_cb41212f-892e-4fe5-ae36-7ba898943277/neutron-httpd/0.log" Oct 05 22:11:34 crc kubenswrapper[4754]: I1005 22:11:34.154534 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-6rhpl_61d4efe0-4a19-4b36-9509-67354a6fd537/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Oct 05 22:11:34 crc kubenswrapper[4754]: I1005 22:11:34.190596 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-5d6f8484b5-vstqw_cb41212f-892e-4fe5-ae36-7ba898943277/neutron-api/0.log" Oct 05 22:11:35 crc kubenswrapper[4754]: I1005 22:11:35.128901 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_b8e2a011-2922-4e21-8e47-b6fcebbc21e9/nova-cell0-conductor-conductor/0.log" Oct 05 22:11:35 crc kubenswrapper[4754]: I1005 22:11:35.570164 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_527adc21-0816-4fba-b7a1-22b3db2e1d73/nova-api-log/0.log" Oct 05 22:11:35 crc kubenswrapper[4754]: I1005 22:11:35.704433 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_9cecb9f9-02d8-42fc-b9ff-b38d3a5aea1b/nova-cell1-conductor-conductor/0.log" Oct 05 22:11:35 crc kubenswrapper[4754]: I1005 22:11:35.877274 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_527adc21-0816-4fba-b7a1-22b3db2e1d73/nova-api-api/0.log" Oct 05 22:11:36 crc kubenswrapper[4754]: I1005 22:11:36.056667 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_eaf71ba6-38b3-4dd1-98e7-e3e5e5d9e428/nova-cell1-novncproxy-novncproxy/0.log" Oct 05 22:11:36 crc kubenswrapper[4754]: I1005 22:11:36.233982 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-fx4cg_542410e6-102a-4e41-b3a5-528f37501180/nova-edpm-deployment-openstack-edpm-ipam/0.log" Oct 05 22:11:36 crc kubenswrapper[4754]: I1005 22:11:36.637468 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_bfc1abc2-68ba-4fc4-9dac-5e15ce2f52cd/nova-metadata-log/0.log" Oct 05 22:11:36 crc kubenswrapper[4754]: I1005 22:11:36.845329 4754 scope.go:117] "RemoveContainer" containerID="d925cdc080b01050e7b9377dba2f144d30f517d60778d239b53617b071498b40" Oct 05 22:11:37 crc kubenswrapper[4754]: I1005 22:11:37.145774 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_d97f6b34-ef35-4b21-81eb-9e7009945fa6/mysql-bootstrap/0.log" Oct 05 22:11:37 crc kubenswrapper[4754]: I1005 22:11:37.340822 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" event={"ID":"be95a413-6a59-45b4-84b7-b43cae694a26","Type":"ContainerStarted","Data":"c50ffee964171eed8b7ee0cd69364b26a808f447930df84f79e3a9a2b8b5bab8"} Oct 05 22:11:37 crc kubenswrapper[4754]: I1005 22:11:37.414141 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_d97f6b34-ef35-4b21-81eb-9e7009945fa6/mysql-bootstrap/0.log" Oct 05 22:11:37 crc kubenswrapper[4754]: I1005 22:11:37.578756 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_3e33c8f1-8161-4919-98e0-971f46b33b05/nova-scheduler-scheduler/0.log" Oct 05 22:11:37 crc kubenswrapper[4754]: I1005 22:11:37.657527 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_d97f6b34-ef35-4b21-81eb-9e7009945fa6/galera/0.log" Oct 05 22:11:37 crc kubenswrapper[4754]: I1005 22:11:37.981267 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_d311946e-91c6-4be6-9d48-1b823ee77607/mysql-bootstrap/0.log" Oct 05 22:11:38 crc kubenswrapper[4754]: I1005 22:11:38.324772 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_d311946e-91c6-4be6-9d48-1b823ee77607/mysql-bootstrap/0.log" Oct 05 22:11:38 crc kubenswrapper[4754]: I1005 22:11:38.367315 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_d311946e-91c6-4be6-9d48-1b823ee77607/galera/0.log" Oct 05 22:11:38 crc kubenswrapper[4754]: I1005 22:11:38.588170 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_95ef109d-c48b-40ea-b491-3e730d4d651b/openstackclient/0.log" Oct 05 22:11:38 crc kubenswrapper[4754]: I1005 22:11:38.895350 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_bfc1abc2-68ba-4fc4-9dac-5e15ce2f52cd/nova-metadata-metadata/0.log" Oct 05 22:11:39 crc kubenswrapper[4754]: I1005 22:11:39.192028 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-8xfw8_06242eb6-ddde-49ea-b4a1-c61aad6f6402/openstack-network-exporter/0.log" Oct 05 22:11:39 crc kubenswrapper[4754]: I1005 22:11:39.485154 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-2krf2_1283a11a-7581-4146-87db-b62d6d5d6dc9/ovsdb-server-init/0.log" Oct 05 22:11:39 crc kubenswrapper[4754]: I1005 22:11:39.679538 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-2krf2_1283a11a-7581-4146-87db-b62d6d5d6dc9/ovsdb-server-init/0.log" Oct 05 22:11:39 crc kubenswrapper[4754]: I1005 22:11:39.703120 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-2krf2_1283a11a-7581-4146-87db-b62d6d5d6dc9/ovs-vswitchd/0.log" Oct 05 22:11:39 crc kubenswrapper[4754]: I1005 22:11:39.766847 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-2krf2_1283a11a-7581-4146-87db-b62d6d5d6dc9/ovsdb-server/0.log" Oct 05 22:11:40 crc kubenswrapper[4754]: I1005 22:11:40.022671 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-pp728_e3785031-57bd-44ab-b30e-b609cca6dcaf/ovn-controller/0.log" Oct 05 22:11:40 crc kubenswrapper[4754]: I1005 22:11:40.823877 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_316121e1-b760-4eaf-9b8e-ec5bacc6b117/openstack-network-exporter/0.log" Oct 05 22:11:40 crc kubenswrapper[4754]: I1005 22:11:40.898617 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-cwth8_c26e6366-e15d-4c7b-a370-a601e2cee56f/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Oct 05 22:11:40 crc kubenswrapper[4754]: I1005 22:11:40.991376 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_316121e1-b760-4eaf-9b8e-ec5bacc6b117/ovn-northd/0.log" Oct 05 22:11:41 crc kubenswrapper[4754]: I1005 22:11:41.156647 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_a5fdaf77-8a4c-4267-a32e-22a1c6be8f97/openstack-network-exporter/0.log" Oct 05 22:11:41 crc kubenswrapper[4754]: I1005 22:11:41.264276 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_a5fdaf77-8a4c-4267-a32e-22a1c6be8f97/ovsdbserver-nb/0.log" Oct 05 22:11:41 crc kubenswrapper[4754]: I1005 22:11:41.406463 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_3c38a205-1079-46c2-9db5-c380d19399e3/openstack-network-exporter/0.log" Oct 05 22:11:41 crc kubenswrapper[4754]: I1005 22:11:41.492458 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_3c38a205-1079-46c2-9db5-c380d19399e3/ovsdbserver-sb/0.log" Oct 05 22:11:41 crc kubenswrapper[4754]: I1005 22:11:41.873880 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-66d84555d4-xzsp4_ac596dcc-74ee-4144-bd90-3e0967d0478d/placement-api/0.log" Oct 05 22:11:41 crc kubenswrapper[4754]: I1005 22:11:41.963350 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-66d84555d4-xzsp4_ac596dcc-74ee-4144-bd90-3e0967d0478d/placement-log/0.log" Oct 05 22:11:42 crc kubenswrapper[4754]: I1005 22:11:42.104159 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_7eb0b89b-e87d-4d5e-9939-d3a0f97f94db/setup-container/0.log" Oct 05 22:11:42 crc kubenswrapper[4754]: I1005 22:11:42.420857 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_7eb0b89b-e87d-4d5e-9939-d3a0f97f94db/setup-container/0.log" Oct 05 22:11:42 crc kubenswrapper[4754]: I1005 22:11:42.460758 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_7eb0b89b-e87d-4d5e-9939-d3a0f97f94db/rabbitmq/0.log" Oct 05 22:11:43 crc kubenswrapper[4754]: I1005 22:11:43.176182 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_3a6824aa-cdc6-4598-b296-1db23f99b156/setup-container/0.log" Oct 05 22:11:43 crc kubenswrapper[4754]: I1005 22:11:43.422135 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_3a6824aa-cdc6-4598-b296-1db23f99b156/rabbitmq/0.log" Oct 05 22:11:43 crc kubenswrapper[4754]: I1005 22:11:43.455405 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_3a6824aa-cdc6-4598-b296-1db23f99b156/setup-container/0.log" Oct 05 22:11:43 crc kubenswrapper[4754]: I1005 22:11:43.683639 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-4srnb_66ef090a-4d9c-41f2-b225-a3bce5bb58c1/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 05 22:11:43 crc kubenswrapper[4754]: I1005 22:11:43.947976 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-xbw7s_fb37244c-6380-4683-a2ed-a96dd32525c6/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Oct 05 22:11:44 crc kubenswrapper[4754]: I1005 22:11:44.149056 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-bwlth_be10d33f-8bb1-4157-9a5f-1f5bbee194f0/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Oct 05 22:11:44 crc kubenswrapper[4754]: I1005 22:11:44.267212 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-m5p76_b30dea9b-eb5b-46c2-88ed-e5b6df21356a/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 05 22:11:44 crc kubenswrapper[4754]: I1005 22:11:44.481261 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-mx4cz_50b5cacd-e284-4f29-acb5-1908246d11d6/ssh-known-hosts-edpm-deployment/0.log" Oct 05 22:11:44 crc kubenswrapper[4754]: I1005 22:11:44.685327 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-646d45cbfc-6sjtl_e4c2921a-fff3-4c97-98df-206a94054d24/proxy-server/0.log" Oct 05 22:11:44 crc kubenswrapper[4754]: I1005 22:11:44.915448 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-646d45cbfc-6sjtl_e4c2921a-fff3-4c97-98df-206a94054d24/proxy-httpd/0.log" Oct 05 22:11:45 crc kubenswrapper[4754]: I1005 22:11:45.095658 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-ljsff_89a55f59-6693-4e0d-b8f8-dfb258d7a078/swift-ring-rebalance/0.log" Oct 05 22:11:45 crc kubenswrapper[4754]: I1005 22:11:45.331781 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9af24378-0013-4510-aee6-d2fe2172935c/account-reaper/0.log" Oct 05 22:11:45 crc kubenswrapper[4754]: I1005 22:11:45.360657 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9af24378-0013-4510-aee6-d2fe2172935c/account-auditor/0.log" Oct 05 22:11:45 crc kubenswrapper[4754]: I1005 22:11:45.603834 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9af24378-0013-4510-aee6-d2fe2172935c/account-replicator/0.log" Oct 05 22:11:45 crc kubenswrapper[4754]: I1005 22:11:45.610311 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9af24378-0013-4510-aee6-d2fe2172935c/account-server/0.log" Oct 05 22:11:45 crc kubenswrapper[4754]: I1005 22:11:45.660215 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9af24378-0013-4510-aee6-d2fe2172935c/container-auditor/0.log" Oct 05 22:11:45 crc kubenswrapper[4754]: I1005 22:11:45.823700 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9af24378-0013-4510-aee6-d2fe2172935c/container-server/0.log" Oct 05 22:11:45 crc kubenswrapper[4754]: I1005 22:11:45.936717 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9af24378-0013-4510-aee6-d2fe2172935c/container-updater/0.log" Oct 05 22:11:45 crc kubenswrapper[4754]: I1005 22:11:45.939343 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9af24378-0013-4510-aee6-d2fe2172935c/container-replicator/0.log" Oct 05 22:11:46 crc kubenswrapper[4754]: I1005 22:11:46.109983 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9af24378-0013-4510-aee6-d2fe2172935c/object-auditor/0.log" Oct 05 22:11:46 crc kubenswrapper[4754]: I1005 22:11:46.175519 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9af24378-0013-4510-aee6-d2fe2172935c/object-expirer/0.log" Oct 05 22:11:46 crc kubenswrapper[4754]: I1005 22:11:46.243434 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9af24378-0013-4510-aee6-d2fe2172935c/object-replicator/0.log" Oct 05 22:11:46 crc kubenswrapper[4754]: I1005 22:11:46.358518 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9af24378-0013-4510-aee6-d2fe2172935c/object-server/0.log" Oct 05 22:11:46 crc kubenswrapper[4754]: I1005 22:11:46.787715 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9af24378-0013-4510-aee6-d2fe2172935c/object-updater/0.log" Oct 05 22:11:46 crc kubenswrapper[4754]: I1005 22:11:46.860702 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9af24378-0013-4510-aee6-d2fe2172935c/rsync/0.log" Oct 05 22:11:46 crc kubenswrapper[4754]: I1005 22:11:46.928587 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9af24378-0013-4510-aee6-d2fe2172935c/swift-recon-cron/0.log" Oct 05 22:11:47 crc kubenswrapper[4754]: I1005 22:11:47.196477 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-s2xqb_7c094190-76ba-4310-ae5d-1e82d3caeac2/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Oct 05 22:11:47 crc kubenswrapper[4754]: I1005 22:11:47.421350 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_5922ff4d-2ec1-4156-acc7-8e6427819b3e/tempest-tests-tempest-tests-runner/0.log" Oct 05 22:11:47 crc kubenswrapper[4754]: I1005 22:11:47.539208 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_ea8cb295-85b1-4afe-8afe-6d5f19b25f28/test-operator-logs-container/0.log" Oct 05 22:11:47 crc kubenswrapper[4754]: I1005 22:11:47.919836 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-thlqz_773b7e70-b0ed-4b02-98fe-eeaf81925900/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Oct 05 22:11:56 crc kubenswrapper[4754]: I1005 22:11:56.101736 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_bdbca489-8f9b-420e-a306-65575175af99/memcached/0.log" Oct 05 22:12:19 crc kubenswrapper[4754]: I1005 22:12:19.748823 4754 generic.go:334] "Generic (PLEG): container finished" podID="fecb08dd-672e-4857-ba0c-198a6efa07d5" containerID="af5377fba7a759eff76ab1fcecce4f748f0fa0b79bd410cb77230eef9b5204e5" exitCode=0 Oct 05 22:12:19 crc kubenswrapper[4754]: I1005 22:12:19.748911 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-44bn6/crc-debug-cspl7" event={"ID":"fecb08dd-672e-4857-ba0c-198a6efa07d5","Type":"ContainerDied","Data":"af5377fba7a759eff76ab1fcecce4f748f0fa0b79bd410cb77230eef9b5204e5"} Oct 05 22:12:20 crc kubenswrapper[4754]: I1005 22:12:20.881979 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-44bn6/crc-debug-cspl7" Oct 05 22:12:20 crc kubenswrapper[4754]: I1005 22:12:20.936250 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-44bn6/crc-debug-cspl7"] Oct 05 22:12:20 crc kubenswrapper[4754]: I1005 22:12:20.956990 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ldl2v\" (UniqueName: \"kubernetes.io/projected/fecb08dd-672e-4857-ba0c-198a6efa07d5-kube-api-access-ldl2v\") pod \"fecb08dd-672e-4857-ba0c-198a6efa07d5\" (UID: \"fecb08dd-672e-4857-ba0c-198a6efa07d5\") " Oct 05 22:12:20 crc kubenswrapper[4754]: I1005 22:12:20.958161 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/fecb08dd-672e-4857-ba0c-198a6efa07d5-host\") pod \"fecb08dd-672e-4857-ba0c-198a6efa07d5\" (UID: \"fecb08dd-672e-4857-ba0c-198a6efa07d5\") " Oct 05 22:12:20 crc kubenswrapper[4754]: I1005 22:12:20.958951 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fecb08dd-672e-4857-ba0c-198a6efa07d5-host" (OuterVolumeSpecName: "host") pod "fecb08dd-672e-4857-ba0c-198a6efa07d5" (UID: "fecb08dd-672e-4857-ba0c-198a6efa07d5"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 05 22:12:20 crc kubenswrapper[4754]: I1005 22:12:20.959207 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-44bn6/crc-debug-cspl7"] Oct 05 22:12:20 crc kubenswrapper[4754]: I1005 22:12:20.968799 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fecb08dd-672e-4857-ba0c-198a6efa07d5-kube-api-access-ldl2v" (OuterVolumeSpecName: "kube-api-access-ldl2v") pod "fecb08dd-672e-4857-ba0c-198a6efa07d5" (UID: "fecb08dd-672e-4857-ba0c-198a6efa07d5"). InnerVolumeSpecName "kube-api-access-ldl2v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 22:12:21 crc kubenswrapper[4754]: I1005 22:12:21.061316 4754 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/fecb08dd-672e-4857-ba0c-198a6efa07d5-host\") on node \"crc\" DevicePath \"\"" Oct 05 22:12:21 crc kubenswrapper[4754]: I1005 22:12:21.061353 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ldl2v\" (UniqueName: \"kubernetes.io/projected/fecb08dd-672e-4857-ba0c-198a6efa07d5-kube-api-access-ldl2v\") on node \"crc\" DevicePath \"\"" Oct 05 22:12:21 crc kubenswrapper[4754]: I1005 22:12:21.776907 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c32c4e9ae117860f28c917869c986a2b739d898cd4b2f274f9713ede5852452e" Oct 05 22:12:21 crc kubenswrapper[4754]: I1005 22:12:21.776969 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-44bn6/crc-debug-cspl7" Oct 05 22:12:22 crc kubenswrapper[4754]: I1005 22:12:22.140184 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-44bn6/crc-debug-tdzbb"] Oct 05 22:12:22 crc kubenswrapper[4754]: E1005 22:12:22.141046 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fecb08dd-672e-4857-ba0c-198a6efa07d5" containerName="container-00" Oct 05 22:12:22 crc kubenswrapper[4754]: I1005 22:12:22.141060 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="fecb08dd-672e-4857-ba0c-198a6efa07d5" containerName="container-00" Oct 05 22:12:22 crc kubenswrapper[4754]: I1005 22:12:22.141264 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="fecb08dd-672e-4857-ba0c-198a6efa07d5" containerName="container-00" Oct 05 22:12:22 crc kubenswrapper[4754]: I1005 22:12:22.142112 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-44bn6/crc-debug-tdzbb" Oct 05 22:12:22 crc kubenswrapper[4754]: I1005 22:12:22.287475 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9f9th\" (UniqueName: \"kubernetes.io/projected/fa851898-82a1-4310-b74b-b6b087682f49-kube-api-access-9f9th\") pod \"crc-debug-tdzbb\" (UID: \"fa851898-82a1-4310-b74b-b6b087682f49\") " pod="openshift-must-gather-44bn6/crc-debug-tdzbb" Oct 05 22:12:22 crc kubenswrapper[4754]: I1005 22:12:22.287547 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/fa851898-82a1-4310-b74b-b6b087682f49-host\") pod \"crc-debug-tdzbb\" (UID: \"fa851898-82a1-4310-b74b-b6b087682f49\") " pod="openshift-must-gather-44bn6/crc-debug-tdzbb" Oct 05 22:12:22 crc kubenswrapper[4754]: I1005 22:12:22.389768 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9f9th\" (UniqueName: \"kubernetes.io/projected/fa851898-82a1-4310-b74b-b6b087682f49-kube-api-access-9f9th\") pod \"crc-debug-tdzbb\" (UID: \"fa851898-82a1-4310-b74b-b6b087682f49\") " pod="openshift-must-gather-44bn6/crc-debug-tdzbb" Oct 05 22:12:22 crc kubenswrapper[4754]: I1005 22:12:22.389876 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/fa851898-82a1-4310-b74b-b6b087682f49-host\") pod \"crc-debug-tdzbb\" (UID: \"fa851898-82a1-4310-b74b-b6b087682f49\") " pod="openshift-must-gather-44bn6/crc-debug-tdzbb" Oct 05 22:12:22 crc kubenswrapper[4754]: I1005 22:12:22.390091 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/fa851898-82a1-4310-b74b-b6b087682f49-host\") pod \"crc-debug-tdzbb\" (UID: \"fa851898-82a1-4310-b74b-b6b087682f49\") " pod="openshift-must-gather-44bn6/crc-debug-tdzbb" Oct 05 22:12:22 crc kubenswrapper[4754]: I1005 22:12:22.420427 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9f9th\" (UniqueName: \"kubernetes.io/projected/fa851898-82a1-4310-b74b-b6b087682f49-kube-api-access-9f9th\") pod \"crc-debug-tdzbb\" (UID: \"fa851898-82a1-4310-b74b-b6b087682f49\") " pod="openshift-must-gather-44bn6/crc-debug-tdzbb" Oct 05 22:12:22 crc kubenswrapper[4754]: I1005 22:12:22.464986 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-44bn6/crc-debug-tdzbb" Oct 05 22:12:22 crc kubenswrapper[4754]: I1005 22:12:22.786919 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-44bn6/crc-debug-tdzbb" event={"ID":"fa851898-82a1-4310-b74b-b6b087682f49","Type":"ContainerStarted","Data":"98c8eb1868eeae834122a92fa4c4fcbd2c64c1e430a55d02f2cfa2d4a3447b4d"} Oct 05 22:12:22 crc kubenswrapper[4754]: I1005 22:12:22.787187 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-44bn6/crc-debug-tdzbb" event={"ID":"fa851898-82a1-4310-b74b-b6b087682f49","Type":"ContainerStarted","Data":"d2f21ce4ae3373b61afce9d1a56ff9e738e04665ce458aec68fe322a38ac4b4a"} Oct 05 22:12:22 crc kubenswrapper[4754]: I1005 22:12:22.852812 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fecb08dd-672e-4857-ba0c-198a6efa07d5" path="/var/lib/kubelet/pods/fecb08dd-672e-4857-ba0c-198a6efa07d5/volumes" Oct 05 22:12:23 crc kubenswrapper[4754]: I1005 22:12:23.802262 4754 generic.go:334] "Generic (PLEG): container finished" podID="fa851898-82a1-4310-b74b-b6b087682f49" containerID="98c8eb1868eeae834122a92fa4c4fcbd2c64c1e430a55d02f2cfa2d4a3447b4d" exitCode=0 Oct 05 22:12:23 crc kubenswrapper[4754]: I1005 22:12:23.803044 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-44bn6/crc-debug-tdzbb" event={"ID":"fa851898-82a1-4310-b74b-b6b087682f49","Type":"ContainerDied","Data":"98c8eb1868eeae834122a92fa4c4fcbd2c64c1e430a55d02f2cfa2d4a3447b4d"} Oct 05 22:12:24 crc kubenswrapper[4754]: I1005 22:12:24.903729 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-44bn6/crc-debug-tdzbb" Oct 05 22:12:24 crc kubenswrapper[4754]: I1005 22:12:24.967638 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9f9th\" (UniqueName: \"kubernetes.io/projected/fa851898-82a1-4310-b74b-b6b087682f49-kube-api-access-9f9th\") pod \"fa851898-82a1-4310-b74b-b6b087682f49\" (UID: \"fa851898-82a1-4310-b74b-b6b087682f49\") " Oct 05 22:12:24 crc kubenswrapper[4754]: I1005 22:12:24.967904 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/fa851898-82a1-4310-b74b-b6b087682f49-host\") pod \"fa851898-82a1-4310-b74b-b6b087682f49\" (UID: \"fa851898-82a1-4310-b74b-b6b087682f49\") " Oct 05 22:12:24 crc kubenswrapper[4754]: I1005 22:12:24.968289 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fa851898-82a1-4310-b74b-b6b087682f49-host" (OuterVolumeSpecName: "host") pod "fa851898-82a1-4310-b74b-b6b087682f49" (UID: "fa851898-82a1-4310-b74b-b6b087682f49"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 05 22:12:24 crc kubenswrapper[4754]: I1005 22:12:24.972545 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fa851898-82a1-4310-b74b-b6b087682f49-kube-api-access-9f9th" (OuterVolumeSpecName: "kube-api-access-9f9th") pod "fa851898-82a1-4310-b74b-b6b087682f49" (UID: "fa851898-82a1-4310-b74b-b6b087682f49"). InnerVolumeSpecName "kube-api-access-9f9th". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 22:12:25 crc kubenswrapper[4754]: I1005 22:12:25.068764 4754 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/fa851898-82a1-4310-b74b-b6b087682f49-host\") on node \"crc\" DevicePath \"\"" Oct 05 22:12:25 crc kubenswrapper[4754]: I1005 22:12:25.069000 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9f9th\" (UniqueName: \"kubernetes.io/projected/fa851898-82a1-4310-b74b-b6b087682f49-kube-api-access-9f9th\") on node \"crc\" DevicePath \"\"" Oct 05 22:12:25 crc kubenswrapper[4754]: I1005 22:12:25.832868 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-44bn6/crc-debug-tdzbb" event={"ID":"fa851898-82a1-4310-b74b-b6b087682f49","Type":"ContainerDied","Data":"d2f21ce4ae3373b61afce9d1a56ff9e738e04665ce458aec68fe322a38ac4b4a"} Oct 05 22:12:25 crc kubenswrapper[4754]: I1005 22:12:25.833202 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-44bn6/crc-debug-tdzbb" Oct 05 22:12:25 crc kubenswrapper[4754]: I1005 22:12:25.833210 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d2f21ce4ae3373b61afce9d1a56ff9e738e04665ce458aec68fe322a38ac4b4a" Oct 05 22:12:30 crc kubenswrapper[4754]: I1005 22:12:30.018317 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-44bn6/crc-debug-tdzbb"] Oct 05 22:12:30 crc kubenswrapper[4754]: I1005 22:12:30.025318 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-44bn6/crc-debug-tdzbb"] Oct 05 22:12:30 crc kubenswrapper[4754]: I1005 22:12:30.855749 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fa851898-82a1-4310-b74b-b6b087682f49" path="/var/lib/kubelet/pods/fa851898-82a1-4310-b74b-b6b087682f49/volumes" Oct 05 22:12:31 crc kubenswrapper[4754]: I1005 22:12:31.211755 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-44bn6/crc-debug-x8zhf"] Oct 05 22:12:31 crc kubenswrapper[4754]: E1005 22:12:31.212125 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa851898-82a1-4310-b74b-b6b087682f49" containerName="container-00" Oct 05 22:12:31 crc kubenswrapper[4754]: I1005 22:12:31.212136 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa851898-82a1-4310-b74b-b6b087682f49" containerName="container-00" Oct 05 22:12:31 crc kubenswrapper[4754]: I1005 22:12:31.212303 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa851898-82a1-4310-b74b-b6b087682f49" containerName="container-00" Oct 05 22:12:31 crc kubenswrapper[4754]: I1005 22:12:31.212951 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-44bn6/crc-debug-x8zhf" Oct 05 22:12:31 crc kubenswrapper[4754]: I1005 22:12:31.395597 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-98lz6\" (UniqueName: \"kubernetes.io/projected/9c78155f-27d5-4073-9080-56b65a9f6280-kube-api-access-98lz6\") pod \"crc-debug-x8zhf\" (UID: \"9c78155f-27d5-4073-9080-56b65a9f6280\") " pod="openshift-must-gather-44bn6/crc-debug-x8zhf" Oct 05 22:12:31 crc kubenswrapper[4754]: I1005 22:12:31.395751 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9c78155f-27d5-4073-9080-56b65a9f6280-host\") pod \"crc-debug-x8zhf\" (UID: \"9c78155f-27d5-4073-9080-56b65a9f6280\") " pod="openshift-must-gather-44bn6/crc-debug-x8zhf" Oct 05 22:12:31 crc kubenswrapper[4754]: I1005 22:12:31.497720 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-98lz6\" (UniqueName: \"kubernetes.io/projected/9c78155f-27d5-4073-9080-56b65a9f6280-kube-api-access-98lz6\") pod \"crc-debug-x8zhf\" (UID: \"9c78155f-27d5-4073-9080-56b65a9f6280\") " pod="openshift-must-gather-44bn6/crc-debug-x8zhf" Oct 05 22:12:31 crc kubenswrapper[4754]: I1005 22:12:31.497812 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9c78155f-27d5-4073-9080-56b65a9f6280-host\") pod \"crc-debug-x8zhf\" (UID: \"9c78155f-27d5-4073-9080-56b65a9f6280\") " pod="openshift-must-gather-44bn6/crc-debug-x8zhf" Oct 05 22:12:31 crc kubenswrapper[4754]: I1005 22:12:31.498095 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9c78155f-27d5-4073-9080-56b65a9f6280-host\") pod \"crc-debug-x8zhf\" (UID: \"9c78155f-27d5-4073-9080-56b65a9f6280\") " pod="openshift-must-gather-44bn6/crc-debug-x8zhf" Oct 05 22:12:31 crc kubenswrapper[4754]: I1005 22:12:31.524004 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-98lz6\" (UniqueName: \"kubernetes.io/projected/9c78155f-27d5-4073-9080-56b65a9f6280-kube-api-access-98lz6\") pod \"crc-debug-x8zhf\" (UID: \"9c78155f-27d5-4073-9080-56b65a9f6280\") " pod="openshift-must-gather-44bn6/crc-debug-x8zhf" Oct 05 22:12:31 crc kubenswrapper[4754]: I1005 22:12:31.533029 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-44bn6/crc-debug-x8zhf" Oct 05 22:12:31 crc kubenswrapper[4754]: W1005 22:12:31.595997 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9c78155f_27d5_4073_9080_56b65a9f6280.slice/crio-6a0d3a2a05d44e2c1b2a9d9ac54cc2d1b099e639c6fe28fc00aacc45fc1b94c0 WatchSource:0}: Error finding container 6a0d3a2a05d44e2c1b2a9d9ac54cc2d1b099e639c6fe28fc00aacc45fc1b94c0: Status 404 returned error can't find the container with id 6a0d3a2a05d44e2c1b2a9d9ac54cc2d1b099e639c6fe28fc00aacc45fc1b94c0 Oct 05 22:12:31 crc kubenswrapper[4754]: I1005 22:12:31.884115 4754 generic.go:334] "Generic (PLEG): container finished" podID="9c78155f-27d5-4073-9080-56b65a9f6280" containerID="a679f30ee81760f053cdd411d89ae372145fc3a626b2b16323529b17d604a4c2" exitCode=0 Oct 05 22:12:31 crc kubenswrapper[4754]: I1005 22:12:31.884202 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-44bn6/crc-debug-x8zhf" event={"ID":"9c78155f-27d5-4073-9080-56b65a9f6280","Type":"ContainerDied","Data":"a679f30ee81760f053cdd411d89ae372145fc3a626b2b16323529b17d604a4c2"} Oct 05 22:12:31 crc kubenswrapper[4754]: I1005 22:12:31.884537 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-44bn6/crc-debug-x8zhf" event={"ID":"9c78155f-27d5-4073-9080-56b65a9f6280","Type":"ContainerStarted","Data":"6a0d3a2a05d44e2c1b2a9d9ac54cc2d1b099e639c6fe28fc00aacc45fc1b94c0"} Oct 05 22:12:31 crc kubenswrapper[4754]: I1005 22:12:31.931902 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-44bn6/crc-debug-x8zhf"] Oct 05 22:12:31 crc kubenswrapper[4754]: I1005 22:12:31.939868 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-44bn6/crc-debug-x8zhf"] Oct 05 22:12:32 crc kubenswrapper[4754]: I1005 22:12:32.984035 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-44bn6/crc-debug-x8zhf" Oct 05 22:12:33 crc kubenswrapper[4754]: I1005 22:12:33.127184 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9c78155f-27d5-4073-9080-56b65a9f6280-host\") pod \"9c78155f-27d5-4073-9080-56b65a9f6280\" (UID: \"9c78155f-27d5-4073-9080-56b65a9f6280\") " Oct 05 22:12:33 crc kubenswrapper[4754]: I1005 22:12:33.127268 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-98lz6\" (UniqueName: \"kubernetes.io/projected/9c78155f-27d5-4073-9080-56b65a9f6280-kube-api-access-98lz6\") pod \"9c78155f-27d5-4073-9080-56b65a9f6280\" (UID: \"9c78155f-27d5-4073-9080-56b65a9f6280\") " Oct 05 22:12:33 crc kubenswrapper[4754]: I1005 22:12:33.127287 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9c78155f-27d5-4073-9080-56b65a9f6280-host" (OuterVolumeSpecName: "host") pod "9c78155f-27d5-4073-9080-56b65a9f6280" (UID: "9c78155f-27d5-4073-9080-56b65a9f6280"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 05 22:12:33 crc kubenswrapper[4754]: I1005 22:12:33.127943 4754 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9c78155f-27d5-4073-9080-56b65a9f6280-host\") on node \"crc\" DevicePath \"\"" Oct 05 22:12:33 crc kubenswrapper[4754]: I1005 22:12:33.148745 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9c78155f-27d5-4073-9080-56b65a9f6280-kube-api-access-98lz6" (OuterVolumeSpecName: "kube-api-access-98lz6") pod "9c78155f-27d5-4073-9080-56b65a9f6280" (UID: "9c78155f-27d5-4073-9080-56b65a9f6280"). InnerVolumeSpecName "kube-api-access-98lz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 22:12:33 crc kubenswrapper[4754]: I1005 22:12:33.230072 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-98lz6\" (UniqueName: \"kubernetes.io/projected/9c78155f-27d5-4073-9080-56b65a9f6280-kube-api-access-98lz6\") on node \"crc\" DevicePath \"\"" Oct 05 22:12:33 crc kubenswrapper[4754]: I1005 22:12:33.899760 4754 scope.go:117] "RemoveContainer" containerID="a679f30ee81760f053cdd411d89ae372145fc3a626b2b16323529b17d604a4c2" Oct 05 22:12:33 crc kubenswrapper[4754]: I1005 22:12:33.899877 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-44bn6/crc-debug-x8zhf" Oct 05 22:12:34 crc kubenswrapper[4754]: I1005 22:12:34.015230 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_aded1cecd2f966ec868aa0a6ebe9c364f1aced79b6d6348d02479f11239rr6r_5f6867a7-6e2b-4f7a-bcdd-a0ff2035eee8/util/0.log" Oct 05 22:12:34 crc kubenswrapper[4754]: I1005 22:12:34.244980 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_aded1cecd2f966ec868aa0a6ebe9c364f1aced79b6d6348d02479f11239rr6r_5f6867a7-6e2b-4f7a-bcdd-a0ff2035eee8/pull/0.log" Oct 05 22:12:34 crc kubenswrapper[4754]: I1005 22:12:34.251500 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_aded1cecd2f966ec868aa0a6ebe9c364f1aced79b6d6348d02479f11239rr6r_5f6867a7-6e2b-4f7a-bcdd-a0ff2035eee8/pull/0.log" Oct 05 22:12:34 crc kubenswrapper[4754]: I1005 22:12:34.278176 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_aded1cecd2f966ec868aa0a6ebe9c364f1aced79b6d6348d02479f11239rr6r_5f6867a7-6e2b-4f7a-bcdd-a0ff2035eee8/util/0.log" Oct 05 22:12:34 crc kubenswrapper[4754]: I1005 22:12:34.434620 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_aded1cecd2f966ec868aa0a6ebe9c364f1aced79b6d6348d02479f11239rr6r_5f6867a7-6e2b-4f7a-bcdd-a0ff2035eee8/pull/0.log" Oct 05 22:12:34 crc kubenswrapper[4754]: I1005 22:12:34.481627 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_aded1cecd2f966ec868aa0a6ebe9c364f1aced79b6d6348d02479f11239rr6r_5f6867a7-6e2b-4f7a-bcdd-a0ff2035eee8/extract/0.log" Oct 05 22:12:34 crc kubenswrapper[4754]: I1005 22:12:34.511190 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_aded1cecd2f966ec868aa0a6ebe9c364f1aced79b6d6348d02479f11239rr6r_5f6867a7-6e2b-4f7a-bcdd-a0ff2035eee8/util/0.log" Oct 05 22:12:34 crc kubenswrapper[4754]: I1005 22:12:34.585277 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-5f7c849b98-n9ksx_9a1cc24e-8371-44cb-bd37-5b765853fdff/kube-rbac-proxy/0.log" Oct 05 22:12:34 crc kubenswrapper[4754]: I1005 22:12:34.735819 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-5f7c849b98-n9ksx_9a1cc24e-8371-44cb-bd37-5b765853fdff/manager/0.log" Oct 05 22:12:34 crc kubenswrapper[4754]: I1005 22:12:34.803376 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-7d4d4f8d-kcxvd_47cafc06-0f75-48ee-bae5-4e30352f4572/kube-rbac-proxy/0.log" Oct 05 22:12:34 crc kubenswrapper[4754]: I1005 22:12:34.850581 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9c78155f-27d5-4073-9080-56b65a9f6280" path="/var/lib/kubelet/pods/9c78155f-27d5-4073-9080-56b65a9f6280/volumes" Oct 05 22:12:34 crc kubenswrapper[4754]: I1005 22:12:34.854939 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-7d4d4f8d-kcxvd_47cafc06-0f75-48ee-bae5-4e30352f4572/manager/0.log" Oct 05 22:12:34 crc kubenswrapper[4754]: I1005 22:12:34.980785 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-75dfd9b554-fxx5s_3ec442d5-220a-4d2f-8f62-a13ff86a3229/kube-rbac-proxy/0.log" Oct 05 22:12:35 crc kubenswrapper[4754]: I1005 22:12:35.035706 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-75dfd9b554-fxx5s_3ec442d5-220a-4d2f-8f62-a13ff86a3229/manager/0.log" Oct 05 22:12:35 crc kubenswrapper[4754]: I1005 22:12:35.152555 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-5568b5d68-f5rjh_62159d2d-ec19-4b44-89df-846faa5757dc/kube-rbac-proxy/0.log" Oct 05 22:12:35 crc kubenswrapper[4754]: I1005 22:12:35.250775 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-5568b5d68-f5rjh_62159d2d-ec19-4b44-89df-846faa5757dc/manager/0.log" Oct 05 22:12:35 crc kubenswrapper[4754]: I1005 22:12:35.308043 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-8f58bc9db-xsrhv_82184642-9d7b-4ada-b7b2-efde6c91cb14/kube-rbac-proxy/0.log" Oct 05 22:12:35 crc kubenswrapper[4754]: I1005 22:12:35.342154 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-8f58bc9db-xsrhv_82184642-9d7b-4ada-b7b2-efde6c91cb14/manager/0.log" Oct 05 22:12:35 crc kubenswrapper[4754]: I1005 22:12:35.471776 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-54876c876f-svzwt_f560133d-b3b5-4791-b69a-a700367f0d96/kube-rbac-proxy/0.log" Oct 05 22:12:35 crc kubenswrapper[4754]: I1005 22:12:35.516142 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-54876c876f-svzwt_f560133d-b3b5-4791-b69a-a700367f0d96/manager/0.log" Oct 05 22:12:35 crc kubenswrapper[4754]: I1005 22:12:35.635250 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-658588b8c9-z7tq9_4bb20421-ce34-44ee-8740-82eedd4716f3/kube-rbac-proxy/0.log" Oct 05 22:12:35 crc kubenswrapper[4754]: I1005 22:12:35.789244 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-658588b8c9-z7tq9_4bb20421-ce34-44ee-8740-82eedd4716f3/manager/0.log" Oct 05 22:12:35 crc kubenswrapper[4754]: I1005 22:12:35.839314 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-649675d675-v7285_3ddbeeee-8a3e-4f3c-97e0-c7d8d5929345/kube-rbac-proxy/0.log" Oct 05 22:12:35 crc kubenswrapper[4754]: I1005 22:12:35.906735 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-649675d675-v7285_3ddbeeee-8a3e-4f3c-97e0-c7d8d5929345/manager/0.log" Oct 05 22:12:35 crc kubenswrapper[4754]: I1005 22:12:35.996776 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-655d88ccb9-l56qv_2735f261-cad9-4dff-9136-931a45c85ac5/kube-rbac-proxy/0.log" Oct 05 22:12:36 crc kubenswrapper[4754]: I1005 22:12:36.089615 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-655d88ccb9-l56qv_2735f261-cad9-4dff-9136-931a45c85ac5/manager/0.log" Oct 05 22:12:36 crc kubenswrapper[4754]: I1005 22:12:36.235793 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-65d89cfd9f-msh87_5757c596-103f-4a00-ac16-ed9e6c9e4719/kube-rbac-proxy/0.log" Oct 05 22:12:36 crc kubenswrapper[4754]: I1005 22:12:36.239323 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-65d89cfd9f-msh87_5757c596-103f-4a00-ac16-ed9e6c9e4719/manager/0.log" Oct 05 22:12:36 crc kubenswrapper[4754]: I1005 22:12:36.427225 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-6cd6d7bdf5-f5jxx_5a1312fe-c6d6-41de-a442-b044db904a6d/kube-rbac-proxy/0.log" Oct 05 22:12:36 crc kubenswrapper[4754]: I1005 22:12:36.486074 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-6cd6d7bdf5-f5jxx_5a1312fe-c6d6-41de-a442-b044db904a6d/manager/0.log" Oct 05 22:12:36 crc kubenswrapper[4754]: I1005 22:12:36.510302 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-8d984cc4d-7xzhb_ff3ba6ea-c2d8-4a3e-b0a0-58fd6df5b196/kube-rbac-proxy/0.log" Oct 05 22:12:36 crc kubenswrapper[4754]: I1005 22:12:36.647762 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-8d984cc4d-7xzhb_ff3ba6ea-c2d8-4a3e-b0a0-58fd6df5b196/manager/0.log" Oct 05 22:12:36 crc kubenswrapper[4754]: I1005 22:12:36.701462 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-7c7fc454ff-4fb5b_e0b701ef-8b8d-4717-aa27-07233f6b6c15/kube-rbac-proxy/0.log" Oct 05 22:12:36 crc kubenswrapper[4754]: I1005 22:12:36.817635 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-7c7fc454ff-4fb5b_e0b701ef-8b8d-4717-aa27-07233f6b6c15/manager/0.log" Oct 05 22:12:36 crc kubenswrapper[4754]: I1005 22:12:36.947470 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-7468f855d8-w8z85_cb4a4798-9283-4633-9b3b-e2d72faa221f/kube-rbac-proxy/0.log" Oct 05 22:12:36 crc kubenswrapper[4754]: I1005 22:12:36.948530 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-7468f855d8-w8z85_cb4a4798-9283-4633-9b3b-e2d72faa221f/manager/0.log" Oct 05 22:12:37 crc kubenswrapper[4754]: I1005 22:12:37.789270 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-5dfbbd665cfq946_c336d534-f819-4859-8e15-5f15b68a36ad/kube-rbac-proxy/0.log" Oct 05 22:12:37 crc kubenswrapper[4754]: I1005 22:12:37.826387 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-5dfbbd665cfq946_c336d534-f819-4859-8e15-5f15b68a36ad/manager/0.log" Oct 05 22:12:37 crc kubenswrapper[4754]: I1005 22:12:37.883882 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-55895b89bc-zzbvh_f131e747-362b-4442-8377-0c00aedfd8ae/kube-rbac-proxy/0.log" Oct 05 22:12:38 crc kubenswrapper[4754]: I1005 22:12:38.024367 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-56b64c5fcb-gzvmj_21fc1c35-c245-457c-a66f-97536332ed52/kube-rbac-proxy/0.log" Oct 05 22:12:38 crc kubenswrapper[4754]: I1005 22:12:38.165510 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-56b64c5fcb-gzvmj_21fc1c35-c245-457c-a66f-97536332ed52/operator/0.log" Oct 05 22:12:38 crc kubenswrapper[4754]: I1005 22:12:38.306449 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-pgjjb_506033e6-79e8-4b4e-acb8-97e63316a15b/registry-server/0.log" Oct 05 22:12:38 crc kubenswrapper[4754]: I1005 22:12:38.436458 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-579449c7d5-zcct6_b18dee24-0d55-4805-8d76-623180998686/kube-rbac-proxy/0.log" Oct 05 22:12:38 crc kubenswrapper[4754]: I1005 22:12:38.543551 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-579449c7d5-zcct6_b18dee24-0d55-4805-8d76-623180998686/manager/0.log" Oct 05 22:12:38 crc kubenswrapper[4754]: I1005 22:12:38.657200 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-54689d9f88-4lrzz_6316859b-8ab7-4b46-b1fa-c7f893d39a95/manager/0.log" Oct 05 22:12:38 crc kubenswrapper[4754]: I1005 22:12:38.675810 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-54689d9f88-4lrzz_6316859b-8ab7-4b46-b1fa-c7f893d39a95/kube-rbac-proxy/0.log" Oct 05 22:12:38 crc kubenswrapper[4754]: I1005 22:12:38.804196 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-5f97d8c699-hkb96_97f32c3b-569a-4f15-9f7c-fb18299dec30/operator/0.log" Oct 05 22:12:39 crc kubenswrapper[4754]: I1005 22:12:39.033979 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-55895b89bc-zzbvh_f131e747-362b-4442-8377-0c00aedfd8ae/manager/0.log" Oct 05 22:12:39 crc kubenswrapper[4754]: I1005 22:12:39.306538 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-6859f9b676-hgkv4_2b729715-7fd9-4ca5-b4dd-cb0eb0034aac/manager/0.log" Oct 05 22:12:39 crc kubenswrapper[4754]: I1005 22:12:39.329537 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-5d4d74dd89-h44bd_f48f3071-cd94-49dd-bfa0-1ec74f495c72/kube-rbac-proxy/0.log" Oct 05 22:12:39 crc kubenswrapper[4754]: I1005 22:12:39.329635 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-6859f9b676-hgkv4_2b729715-7fd9-4ca5-b4dd-cb0eb0034aac/kube-rbac-proxy/0.log" Oct 05 22:12:39 crc kubenswrapper[4754]: I1005 22:12:39.434249 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-5d4d74dd89-h44bd_f48f3071-cd94-49dd-bfa0-1ec74f495c72/manager/0.log" Oct 05 22:12:39 crc kubenswrapper[4754]: I1005 22:12:39.557940 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5cd5cb47d7-fx25z_b1cc32e7-f619-4a0e-b0f9-195e36c82d01/kube-rbac-proxy/0.log" Oct 05 22:12:39 crc kubenswrapper[4754]: I1005 22:12:39.617384 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5cd5cb47d7-fx25z_b1cc32e7-f619-4a0e-b0f9-195e36c82d01/manager/0.log" Oct 05 22:12:39 crc kubenswrapper[4754]: I1005 22:12:39.670278 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-6cbc6dd547-zq79b_cea6487f-b088-4f9e-a42d-6a3517df7669/kube-rbac-proxy/0.log" Oct 05 22:12:39 crc kubenswrapper[4754]: I1005 22:12:39.696165 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-6cbc6dd547-zq79b_cea6487f-b088-4f9e-a42d-6a3517df7669/manager/0.log" Oct 05 22:12:57 crc kubenswrapper[4754]: I1005 22:12:57.177876 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-5wx56_d1b2b973-8b95-4f6a-83f1-468f84fb006d/control-plane-machine-set-operator/0.log" Oct 05 22:12:57 crc kubenswrapper[4754]: I1005 22:12:57.327100 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-p5wqw_c463939d-d2d1-4049-b438-3754a118ff4b/machine-api-operator/0.log" Oct 05 22:12:57 crc kubenswrapper[4754]: I1005 22:12:57.330354 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-p5wqw_c463939d-d2d1-4049-b438-3754a118ff4b/kube-rbac-proxy/0.log" Oct 05 22:13:11 crc kubenswrapper[4754]: I1005 22:13:11.314870 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-c5gxb_e31db265-82f2-431c-af44-9a44bcb4762c/cert-manager-controller/0.log" Oct 05 22:13:11 crc kubenswrapper[4754]: I1005 22:13:11.396880 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-pqw5s_f060400d-9b95-4e44-b94a-ba61d23e307c/cert-manager-cainjector/0.log" Oct 05 22:13:11 crc kubenswrapper[4754]: I1005 22:13:11.489765 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-r2txk_577f8ef5-2c2c-4b67-91b5-55ce19fc52d2/cert-manager-webhook/0.log" Oct 05 22:13:25 crc kubenswrapper[4754]: I1005 22:13:25.225527 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-6b874cbd85-pvfm8_785a344e-fb90-43ec-a951-962783b9ccbf/nmstate-console-plugin/0.log" Oct 05 22:13:25 crc kubenswrapper[4754]: I1005 22:13:25.239253 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-d4qnm_76cf164c-fcc4-49e8-866f-e4ec435a8044/nmstate-handler/0.log" Oct 05 22:13:25 crc kubenswrapper[4754]: I1005 22:13:25.684214 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-7469b_b381c1ba-7339-4ab8-bc4a-d3887197d1ea/kube-rbac-proxy/0.log" Oct 05 22:13:25 crc kubenswrapper[4754]: I1005 22:13:25.744705 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-7469b_b381c1ba-7339-4ab8-bc4a-d3887197d1ea/nmstate-metrics/0.log" Oct 05 22:13:25 crc kubenswrapper[4754]: I1005 22:13:25.956518 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-858ddd8f98-z2c5l_262b4fe5-27b8-41fe-8d8c-aed9ca555a6e/nmstate-operator/0.log" Oct 05 22:13:25 crc kubenswrapper[4754]: I1005 22:13:25.995414 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-6cdbc54649-nds52_30f05ad2-a252-4b9c-890d-892b90aaa568/nmstate-webhook/0.log" Oct 05 22:13:40 crc kubenswrapper[4754]: I1005 22:13:40.917766 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-78nmb_22f6b8d5-a7f7-4db9-bd76-eb9e5e23ea79/controller/0.log" Oct 05 22:13:40 crc kubenswrapper[4754]: I1005 22:13:40.945430 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-78nmb_22f6b8d5-a7f7-4db9-bd76-eb9e5e23ea79/kube-rbac-proxy/0.log" Oct 05 22:13:41 crc kubenswrapper[4754]: I1005 22:13:41.106480 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tfh49_dbf73baa-e798-43fa-95de-daec87f5809a/cp-frr-files/0.log" Oct 05 22:13:41 crc kubenswrapper[4754]: I1005 22:13:41.272475 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tfh49_dbf73baa-e798-43fa-95de-daec87f5809a/cp-frr-files/0.log" Oct 05 22:13:41 crc kubenswrapper[4754]: I1005 22:13:41.308964 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tfh49_dbf73baa-e798-43fa-95de-daec87f5809a/cp-reloader/0.log" Oct 05 22:13:41 crc kubenswrapper[4754]: I1005 22:13:41.337172 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tfh49_dbf73baa-e798-43fa-95de-daec87f5809a/cp-metrics/0.log" Oct 05 22:13:41 crc kubenswrapper[4754]: I1005 22:13:41.375669 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tfh49_dbf73baa-e798-43fa-95de-daec87f5809a/cp-reloader/0.log" Oct 05 22:13:41 crc kubenswrapper[4754]: I1005 22:13:41.469262 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tfh49_dbf73baa-e798-43fa-95de-daec87f5809a/cp-frr-files/0.log" Oct 05 22:13:41 crc kubenswrapper[4754]: I1005 22:13:41.568532 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tfh49_dbf73baa-e798-43fa-95de-daec87f5809a/cp-reloader/0.log" Oct 05 22:13:41 crc kubenswrapper[4754]: I1005 22:13:41.577871 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tfh49_dbf73baa-e798-43fa-95de-daec87f5809a/cp-metrics/0.log" Oct 05 22:13:41 crc kubenswrapper[4754]: I1005 22:13:41.598331 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tfh49_dbf73baa-e798-43fa-95de-daec87f5809a/cp-metrics/0.log" Oct 05 22:13:41 crc kubenswrapper[4754]: I1005 22:13:41.798316 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tfh49_dbf73baa-e798-43fa-95de-daec87f5809a/cp-metrics/0.log" Oct 05 22:13:41 crc kubenswrapper[4754]: I1005 22:13:41.806084 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tfh49_dbf73baa-e798-43fa-95de-daec87f5809a/cp-frr-files/0.log" Oct 05 22:13:41 crc kubenswrapper[4754]: I1005 22:13:41.838047 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tfh49_dbf73baa-e798-43fa-95de-daec87f5809a/cp-reloader/0.log" Oct 05 22:13:41 crc kubenswrapper[4754]: I1005 22:13:41.900116 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tfh49_dbf73baa-e798-43fa-95de-daec87f5809a/controller/0.log" Oct 05 22:13:42 crc kubenswrapper[4754]: I1005 22:13:42.015195 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tfh49_dbf73baa-e798-43fa-95de-daec87f5809a/frr-metrics/0.log" Oct 05 22:13:42 crc kubenswrapper[4754]: I1005 22:13:42.116736 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tfh49_dbf73baa-e798-43fa-95de-daec87f5809a/kube-rbac-proxy/0.log" Oct 05 22:13:42 crc kubenswrapper[4754]: I1005 22:13:42.162813 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tfh49_dbf73baa-e798-43fa-95de-daec87f5809a/kube-rbac-proxy-frr/0.log" Oct 05 22:13:42 crc kubenswrapper[4754]: I1005 22:13:42.256695 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tfh49_dbf73baa-e798-43fa-95de-daec87f5809a/reloader/0.log" Oct 05 22:13:42 crc kubenswrapper[4754]: I1005 22:13:42.487219 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-64bf5d555-zt7sz_69687995-5b0d-4aae-9551-1b1b9cb2bec3/frr-k8s-webhook-server/0.log" Oct 05 22:13:42 crc kubenswrapper[4754]: I1005 22:13:42.708881 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-645fb4cd96-7qplt_697d9fce-0466-4ba2-a532-214dfd58a20e/manager/0.log" Oct 05 22:13:42 crc kubenswrapper[4754]: I1005 22:13:42.853754 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-5ff795db68-8xphz_56375d90-5d62-4d6f-897f-28ca18bf3682/webhook-server/0.log" Oct 05 22:13:43 crc kubenswrapper[4754]: I1005 22:13:43.034706 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-hkvd8_34ab3bd8-96ff-4f52-ae2e-e8cbd748827d/kube-rbac-proxy/0.log" Oct 05 22:13:43 crc kubenswrapper[4754]: I1005 22:13:43.352211 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tfh49_dbf73baa-e798-43fa-95de-daec87f5809a/frr/0.log" Oct 05 22:13:43 crc kubenswrapper[4754]: I1005 22:13:43.468531 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-hkvd8_34ab3bd8-96ff-4f52-ae2e-e8cbd748827d/speaker/0.log" Oct 05 22:13:45 crc kubenswrapper[4754]: E1005 22:13:45.977122 4754 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/system.slice/rpm-ostreed.service\": RecentStats: unable to find data in memory cache]" Oct 05 22:13:58 crc kubenswrapper[4754]: I1005 22:13:58.401834 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2kzp4h_3da3d4d5-ef4b-4dde-8b69-0d8e566703c1/util/0.log" Oct 05 22:13:58 crc kubenswrapper[4754]: I1005 22:13:58.570368 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2kzp4h_3da3d4d5-ef4b-4dde-8b69-0d8e566703c1/util/0.log" Oct 05 22:13:58 crc kubenswrapper[4754]: I1005 22:13:58.645026 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2kzp4h_3da3d4d5-ef4b-4dde-8b69-0d8e566703c1/pull/0.log" Oct 05 22:13:58 crc kubenswrapper[4754]: I1005 22:13:58.748610 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2kzp4h_3da3d4d5-ef4b-4dde-8b69-0d8e566703c1/pull/0.log" Oct 05 22:13:58 crc kubenswrapper[4754]: I1005 22:13:58.905899 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2kzp4h_3da3d4d5-ef4b-4dde-8b69-0d8e566703c1/pull/0.log" Oct 05 22:13:58 crc kubenswrapper[4754]: I1005 22:13:58.937203 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2kzp4h_3da3d4d5-ef4b-4dde-8b69-0d8e566703c1/util/0.log" Oct 05 22:13:58 crc kubenswrapper[4754]: I1005 22:13:58.961411 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2kzp4h_3da3d4d5-ef4b-4dde-8b69-0d8e566703c1/extract/0.log" Oct 05 22:13:59 crc kubenswrapper[4754]: I1005 22:13:59.097015 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-m7sf7_5a34915c-4494-448a-b835-6f1276b105fa/extract-utilities/0.log" Oct 05 22:13:59 crc kubenswrapper[4754]: I1005 22:13:59.257232 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-m7sf7_5a34915c-4494-448a-b835-6f1276b105fa/extract-utilities/0.log" Oct 05 22:13:59 crc kubenswrapper[4754]: I1005 22:13:59.303837 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-m7sf7_5a34915c-4494-448a-b835-6f1276b105fa/extract-content/0.log" Oct 05 22:13:59 crc kubenswrapper[4754]: I1005 22:13:59.320622 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-m7sf7_5a34915c-4494-448a-b835-6f1276b105fa/extract-content/0.log" Oct 05 22:13:59 crc kubenswrapper[4754]: I1005 22:13:59.443275 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-m7sf7_5a34915c-4494-448a-b835-6f1276b105fa/extract-utilities/0.log" Oct 05 22:13:59 crc kubenswrapper[4754]: I1005 22:13:59.476961 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-m7sf7_5a34915c-4494-448a-b835-6f1276b105fa/extract-content/0.log" Oct 05 22:13:59 crc kubenswrapper[4754]: I1005 22:13:59.715795 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-pkf48_095d3b9f-145a-4bc8-adca-a1831739e193/extract-utilities/0.log" Oct 05 22:13:59 crc kubenswrapper[4754]: I1005 22:13:59.944523 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-pkf48_095d3b9f-145a-4bc8-adca-a1831739e193/extract-utilities/0.log" Oct 05 22:13:59 crc kubenswrapper[4754]: I1005 22:13:59.995294 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-pkf48_095d3b9f-145a-4bc8-adca-a1831739e193/extract-content/0.log" Oct 05 22:13:59 crc kubenswrapper[4754]: I1005 22:13:59.997413 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-m7sf7_5a34915c-4494-448a-b835-6f1276b105fa/registry-server/0.log" Oct 05 22:14:00 crc kubenswrapper[4754]: I1005 22:14:00.025706 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-pkf48_095d3b9f-145a-4bc8-adca-a1831739e193/extract-content/0.log" Oct 05 22:14:00 crc kubenswrapper[4754]: I1005 22:14:00.211713 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-pkf48_095d3b9f-145a-4bc8-adca-a1831739e193/extract-content/0.log" Oct 05 22:14:00 crc kubenswrapper[4754]: I1005 22:14:00.252197 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-pkf48_095d3b9f-145a-4bc8-adca-a1831739e193/extract-utilities/0.log" Oct 05 22:14:00 crc kubenswrapper[4754]: I1005 22:14:00.495459 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cpbjks_f7edcdad-0804-4ae8-bb6d-1118b6d0f620/util/0.log" Oct 05 22:14:00 crc kubenswrapper[4754]: I1005 22:14:00.758669 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-pkf48_095d3b9f-145a-4bc8-adca-a1831739e193/registry-server/0.log" Oct 05 22:14:00 crc kubenswrapper[4754]: I1005 22:14:00.762996 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cpbjks_f7edcdad-0804-4ae8-bb6d-1118b6d0f620/pull/0.log" Oct 05 22:14:00 crc kubenswrapper[4754]: I1005 22:14:00.812323 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cpbjks_f7edcdad-0804-4ae8-bb6d-1118b6d0f620/util/0.log" Oct 05 22:14:00 crc kubenswrapper[4754]: I1005 22:14:00.813484 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cpbjks_f7edcdad-0804-4ae8-bb6d-1118b6d0f620/pull/0.log" Oct 05 22:14:00 crc kubenswrapper[4754]: I1005 22:14:00.963420 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cpbjks_f7edcdad-0804-4ae8-bb6d-1118b6d0f620/pull/0.log" Oct 05 22:14:00 crc kubenswrapper[4754]: I1005 22:14:00.999444 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cpbjks_f7edcdad-0804-4ae8-bb6d-1118b6d0f620/util/0.log" Oct 05 22:14:01 crc kubenswrapper[4754]: I1005 22:14:01.011616 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cpbjks_f7edcdad-0804-4ae8-bb6d-1118b6d0f620/extract/0.log" Oct 05 22:14:01 crc kubenswrapper[4754]: I1005 22:14:01.157387 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-xftcm_ed62afda-bbd2-4326-8101-0abe44ecb2f5/marketplace-operator/0.log" Oct 05 22:14:01 crc kubenswrapper[4754]: I1005 22:14:01.242115 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-qdsv6_f5951a44-c13f-4cef-a2cf-6803479067f8/extract-utilities/0.log" Oct 05 22:14:01 crc kubenswrapper[4754]: I1005 22:14:01.437984 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-qdsv6_f5951a44-c13f-4cef-a2cf-6803479067f8/extract-utilities/0.log" Oct 05 22:14:01 crc kubenswrapper[4754]: I1005 22:14:01.439036 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-qdsv6_f5951a44-c13f-4cef-a2cf-6803479067f8/extract-content/0.log" Oct 05 22:14:01 crc kubenswrapper[4754]: I1005 22:14:01.473103 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-qdsv6_f5951a44-c13f-4cef-a2cf-6803479067f8/extract-content/0.log" Oct 05 22:14:01 crc kubenswrapper[4754]: I1005 22:14:01.571302 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-qdsv6_f5951a44-c13f-4cef-a2cf-6803479067f8/extract-content/0.log" Oct 05 22:14:01 crc kubenswrapper[4754]: I1005 22:14:01.586420 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-qdsv6_f5951a44-c13f-4cef-a2cf-6803479067f8/extract-utilities/0.log" Oct 05 22:14:01 crc kubenswrapper[4754]: I1005 22:14:01.716348 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-qdsv6_f5951a44-c13f-4cef-a2cf-6803479067f8/registry-server/0.log" Oct 05 22:14:01 crc kubenswrapper[4754]: I1005 22:14:01.795027 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-2lvpm_b27b18ff-2e0a-446d-b605-5b74f825260b/extract-utilities/0.log" Oct 05 22:14:01 crc kubenswrapper[4754]: I1005 22:14:01.963701 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-2lvpm_b27b18ff-2e0a-446d-b605-5b74f825260b/extract-content/0.log" Oct 05 22:14:01 crc kubenswrapper[4754]: I1005 22:14:01.968099 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-2lvpm_b27b18ff-2e0a-446d-b605-5b74f825260b/extract-utilities/0.log" Oct 05 22:14:01 crc kubenswrapper[4754]: I1005 22:14:01.980575 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-2lvpm_b27b18ff-2e0a-446d-b605-5b74f825260b/extract-content/0.log" Oct 05 22:14:02 crc kubenswrapper[4754]: I1005 22:14:02.122916 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-2lvpm_b27b18ff-2e0a-446d-b605-5b74f825260b/extract-content/0.log" Oct 05 22:14:02 crc kubenswrapper[4754]: I1005 22:14:02.126191 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-2lvpm_b27b18ff-2e0a-446d-b605-5b74f825260b/extract-utilities/0.log" Oct 05 22:14:02 crc kubenswrapper[4754]: I1005 22:14:02.551879 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-2lvpm_b27b18ff-2e0a-446d-b605-5b74f825260b/registry-server/0.log" Oct 05 22:14:05 crc kubenswrapper[4754]: I1005 22:14:05.244861 4754 patch_prober.go:28] interesting pod/machine-config-daemon-b2h9k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 22:14:05 crc kubenswrapper[4754]: I1005 22:14:05.245424 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 22:14:35 crc kubenswrapper[4754]: I1005 22:14:35.245037 4754 patch_prober.go:28] interesting pod/machine-config-daemon-b2h9k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 22:14:35 crc kubenswrapper[4754]: I1005 22:14:35.245478 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 22:14:36 crc kubenswrapper[4754]: E1005 22:14:36.727116 4754 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.20:49922->38.102.83.20:44369: write tcp 38.102.83.20:49922->38.102.83.20:44369: write: broken pipe Oct 05 22:15:00 crc kubenswrapper[4754]: I1005 22:15:00.154799 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29328375-qkxnn"] Oct 05 22:15:00 crc kubenswrapper[4754]: E1005 22:15:00.155772 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c78155f-27d5-4073-9080-56b65a9f6280" containerName="container-00" Oct 05 22:15:00 crc kubenswrapper[4754]: I1005 22:15:00.155787 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c78155f-27d5-4073-9080-56b65a9f6280" containerName="container-00" Oct 05 22:15:00 crc kubenswrapper[4754]: I1005 22:15:00.155999 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c78155f-27d5-4073-9080-56b65a9f6280" containerName="container-00" Oct 05 22:15:00 crc kubenswrapper[4754]: I1005 22:15:00.156628 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29328375-qkxnn" Oct 05 22:15:00 crc kubenswrapper[4754]: I1005 22:15:00.160890 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 05 22:15:00 crc kubenswrapper[4754]: I1005 22:15:00.160940 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 05 22:15:00 crc kubenswrapper[4754]: I1005 22:15:00.174363 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29328375-qkxnn"] Oct 05 22:15:00 crc kubenswrapper[4754]: I1005 22:15:00.285556 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qbtvf\" (UniqueName: \"kubernetes.io/projected/9456a56b-1294-42a9-8f1e-8932bae6e4ba-kube-api-access-qbtvf\") pod \"collect-profiles-29328375-qkxnn\" (UID: \"9456a56b-1294-42a9-8f1e-8932bae6e4ba\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29328375-qkxnn" Oct 05 22:15:00 crc kubenswrapper[4754]: I1005 22:15:00.285609 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9456a56b-1294-42a9-8f1e-8932bae6e4ba-config-volume\") pod \"collect-profiles-29328375-qkxnn\" (UID: \"9456a56b-1294-42a9-8f1e-8932bae6e4ba\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29328375-qkxnn" Oct 05 22:15:00 crc kubenswrapper[4754]: I1005 22:15:00.285858 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9456a56b-1294-42a9-8f1e-8932bae6e4ba-secret-volume\") pod \"collect-profiles-29328375-qkxnn\" (UID: \"9456a56b-1294-42a9-8f1e-8932bae6e4ba\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29328375-qkxnn" Oct 05 22:15:00 crc kubenswrapper[4754]: I1005 22:15:00.387771 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qbtvf\" (UniqueName: \"kubernetes.io/projected/9456a56b-1294-42a9-8f1e-8932bae6e4ba-kube-api-access-qbtvf\") pod \"collect-profiles-29328375-qkxnn\" (UID: \"9456a56b-1294-42a9-8f1e-8932bae6e4ba\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29328375-qkxnn" Oct 05 22:15:00 crc kubenswrapper[4754]: I1005 22:15:00.387815 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9456a56b-1294-42a9-8f1e-8932bae6e4ba-config-volume\") pod \"collect-profiles-29328375-qkxnn\" (UID: \"9456a56b-1294-42a9-8f1e-8932bae6e4ba\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29328375-qkxnn" Oct 05 22:15:00 crc kubenswrapper[4754]: I1005 22:15:00.388115 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9456a56b-1294-42a9-8f1e-8932bae6e4ba-secret-volume\") pod \"collect-profiles-29328375-qkxnn\" (UID: \"9456a56b-1294-42a9-8f1e-8932bae6e4ba\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29328375-qkxnn" Oct 05 22:15:00 crc kubenswrapper[4754]: I1005 22:15:00.390109 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9456a56b-1294-42a9-8f1e-8932bae6e4ba-config-volume\") pod \"collect-profiles-29328375-qkxnn\" (UID: \"9456a56b-1294-42a9-8f1e-8932bae6e4ba\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29328375-qkxnn" Oct 05 22:15:00 crc kubenswrapper[4754]: I1005 22:15:00.413169 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9456a56b-1294-42a9-8f1e-8932bae6e4ba-secret-volume\") pod \"collect-profiles-29328375-qkxnn\" (UID: \"9456a56b-1294-42a9-8f1e-8932bae6e4ba\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29328375-qkxnn" Oct 05 22:15:00 crc kubenswrapper[4754]: I1005 22:15:00.414308 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qbtvf\" (UniqueName: \"kubernetes.io/projected/9456a56b-1294-42a9-8f1e-8932bae6e4ba-kube-api-access-qbtvf\") pod \"collect-profiles-29328375-qkxnn\" (UID: \"9456a56b-1294-42a9-8f1e-8932bae6e4ba\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29328375-qkxnn" Oct 05 22:15:00 crc kubenswrapper[4754]: I1005 22:15:00.485048 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29328375-qkxnn" Oct 05 22:15:01 crc kubenswrapper[4754]: I1005 22:15:01.646638 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29328375-qkxnn"] Oct 05 22:15:02 crc kubenswrapper[4754]: I1005 22:15:02.375856 4754 generic.go:334] "Generic (PLEG): container finished" podID="9456a56b-1294-42a9-8f1e-8932bae6e4ba" containerID="50355f80825ab879c472835eaa87d48dd6e83321e8368e4ff5c5bad58dca2b1a" exitCode=0 Oct 05 22:15:02 crc kubenswrapper[4754]: I1005 22:15:02.375908 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29328375-qkxnn" event={"ID":"9456a56b-1294-42a9-8f1e-8932bae6e4ba","Type":"ContainerDied","Data":"50355f80825ab879c472835eaa87d48dd6e83321e8368e4ff5c5bad58dca2b1a"} Oct 05 22:15:02 crc kubenswrapper[4754]: I1005 22:15:02.376170 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29328375-qkxnn" event={"ID":"9456a56b-1294-42a9-8f1e-8932bae6e4ba","Type":"ContainerStarted","Data":"4f599a2cbbb29897cea13b3343d5c76800dbb48769c97fbc7a72a9f5f4928305"} Oct 05 22:15:03 crc kubenswrapper[4754]: I1005 22:15:03.763518 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29328375-qkxnn" Oct 05 22:15:03 crc kubenswrapper[4754]: I1005 22:15:03.964231 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9456a56b-1294-42a9-8f1e-8932bae6e4ba-secret-volume\") pod \"9456a56b-1294-42a9-8f1e-8932bae6e4ba\" (UID: \"9456a56b-1294-42a9-8f1e-8932bae6e4ba\") " Oct 05 22:15:03 crc kubenswrapper[4754]: I1005 22:15:03.964441 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9456a56b-1294-42a9-8f1e-8932bae6e4ba-config-volume\") pod \"9456a56b-1294-42a9-8f1e-8932bae6e4ba\" (UID: \"9456a56b-1294-42a9-8f1e-8932bae6e4ba\") " Oct 05 22:15:03 crc kubenswrapper[4754]: I1005 22:15:03.964483 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qbtvf\" (UniqueName: \"kubernetes.io/projected/9456a56b-1294-42a9-8f1e-8932bae6e4ba-kube-api-access-qbtvf\") pod \"9456a56b-1294-42a9-8f1e-8932bae6e4ba\" (UID: \"9456a56b-1294-42a9-8f1e-8932bae6e4ba\") " Oct 05 22:15:03 crc kubenswrapper[4754]: I1005 22:15:03.965800 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9456a56b-1294-42a9-8f1e-8932bae6e4ba-config-volume" (OuterVolumeSpecName: "config-volume") pod "9456a56b-1294-42a9-8f1e-8932bae6e4ba" (UID: "9456a56b-1294-42a9-8f1e-8932bae6e4ba"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 22:15:03 crc kubenswrapper[4754]: I1005 22:15:03.969769 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9456a56b-1294-42a9-8f1e-8932bae6e4ba-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "9456a56b-1294-42a9-8f1e-8932bae6e4ba" (UID: "9456a56b-1294-42a9-8f1e-8932bae6e4ba"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 22:15:03 crc kubenswrapper[4754]: I1005 22:15:03.971745 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9456a56b-1294-42a9-8f1e-8932bae6e4ba-kube-api-access-qbtvf" (OuterVolumeSpecName: "kube-api-access-qbtvf") pod "9456a56b-1294-42a9-8f1e-8932bae6e4ba" (UID: "9456a56b-1294-42a9-8f1e-8932bae6e4ba"). InnerVolumeSpecName "kube-api-access-qbtvf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 22:15:04 crc kubenswrapper[4754]: I1005 22:15:04.068158 4754 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9456a56b-1294-42a9-8f1e-8932bae6e4ba-config-volume\") on node \"crc\" DevicePath \"\"" Oct 05 22:15:04 crc kubenswrapper[4754]: I1005 22:15:04.068230 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qbtvf\" (UniqueName: \"kubernetes.io/projected/9456a56b-1294-42a9-8f1e-8932bae6e4ba-kube-api-access-qbtvf\") on node \"crc\" DevicePath \"\"" Oct 05 22:15:04 crc kubenswrapper[4754]: I1005 22:15:04.068250 4754 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9456a56b-1294-42a9-8f1e-8932bae6e4ba-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 05 22:15:04 crc kubenswrapper[4754]: I1005 22:15:04.140192 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-4s4rx"] Oct 05 22:15:04 crc kubenswrapper[4754]: E1005 22:15:04.140809 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9456a56b-1294-42a9-8f1e-8932bae6e4ba" containerName="collect-profiles" Oct 05 22:15:04 crc kubenswrapper[4754]: I1005 22:15:04.141020 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="9456a56b-1294-42a9-8f1e-8932bae6e4ba" containerName="collect-profiles" Oct 05 22:15:04 crc kubenswrapper[4754]: I1005 22:15:04.141392 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="9456a56b-1294-42a9-8f1e-8932bae6e4ba" containerName="collect-profiles" Oct 05 22:15:04 crc kubenswrapper[4754]: I1005 22:15:04.143875 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4s4rx" Oct 05 22:15:04 crc kubenswrapper[4754]: I1005 22:15:04.152941 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-4s4rx"] Oct 05 22:15:04 crc kubenswrapper[4754]: I1005 22:15:04.272250 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91b810ad-510b-41f2-833b-bf7ce9c2b1b6-catalog-content\") pod \"redhat-operators-4s4rx\" (UID: \"91b810ad-510b-41f2-833b-bf7ce9c2b1b6\") " pod="openshift-marketplace/redhat-operators-4s4rx" Oct 05 22:15:04 crc kubenswrapper[4754]: I1005 22:15:04.272323 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xhrwg\" (UniqueName: \"kubernetes.io/projected/91b810ad-510b-41f2-833b-bf7ce9c2b1b6-kube-api-access-xhrwg\") pod \"redhat-operators-4s4rx\" (UID: \"91b810ad-510b-41f2-833b-bf7ce9c2b1b6\") " pod="openshift-marketplace/redhat-operators-4s4rx" Oct 05 22:15:04 crc kubenswrapper[4754]: I1005 22:15:04.272402 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91b810ad-510b-41f2-833b-bf7ce9c2b1b6-utilities\") pod \"redhat-operators-4s4rx\" (UID: \"91b810ad-510b-41f2-833b-bf7ce9c2b1b6\") " pod="openshift-marketplace/redhat-operators-4s4rx" Oct 05 22:15:04 crc kubenswrapper[4754]: I1005 22:15:04.373989 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91b810ad-510b-41f2-833b-bf7ce9c2b1b6-catalog-content\") pod \"redhat-operators-4s4rx\" (UID: \"91b810ad-510b-41f2-833b-bf7ce9c2b1b6\") " pod="openshift-marketplace/redhat-operators-4s4rx" Oct 05 22:15:04 crc kubenswrapper[4754]: I1005 22:15:04.374093 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xhrwg\" (UniqueName: \"kubernetes.io/projected/91b810ad-510b-41f2-833b-bf7ce9c2b1b6-kube-api-access-xhrwg\") pod \"redhat-operators-4s4rx\" (UID: \"91b810ad-510b-41f2-833b-bf7ce9c2b1b6\") " pod="openshift-marketplace/redhat-operators-4s4rx" Oct 05 22:15:04 crc kubenswrapper[4754]: I1005 22:15:04.374201 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91b810ad-510b-41f2-833b-bf7ce9c2b1b6-utilities\") pod \"redhat-operators-4s4rx\" (UID: \"91b810ad-510b-41f2-833b-bf7ce9c2b1b6\") " pod="openshift-marketplace/redhat-operators-4s4rx" Oct 05 22:15:04 crc kubenswrapper[4754]: I1005 22:15:04.374430 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91b810ad-510b-41f2-833b-bf7ce9c2b1b6-catalog-content\") pod \"redhat-operators-4s4rx\" (UID: \"91b810ad-510b-41f2-833b-bf7ce9c2b1b6\") " pod="openshift-marketplace/redhat-operators-4s4rx" Oct 05 22:15:04 crc kubenswrapper[4754]: I1005 22:15:04.374715 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91b810ad-510b-41f2-833b-bf7ce9c2b1b6-utilities\") pod \"redhat-operators-4s4rx\" (UID: \"91b810ad-510b-41f2-833b-bf7ce9c2b1b6\") " pod="openshift-marketplace/redhat-operators-4s4rx" Oct 05 22:15:04 crc kubenswrapper[4754]: I1005 22:15:04.391745 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xhrwg\" (UniqueName: \"kubernetes.io/projected/91b810ad-510b-41f2-833b-bf7ce9c2b1b6-kube-api-access-xhrwg\") pod \"redhat-operators-4s4rx\" (UID: \"91b810ad-510b-41f2-833b-bf7ce9c2b1b6\") " pod="openshift-marketplace/redhat-operators-4s4rx" Oct 05 22:15:04 crc kubenswrapper[4754]: I1005 22:15:04.398133 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29328375-qkxnn" event={"ID":"9456a56b-1294-42a9-8f1e-8932bae6e4ba","Type":"ContainerDied","Data":"4f599a2cbbb29897cea13b3343d5c76800dbb48769c97fbc7a72a9f5f4928305"} Oct 05 22:15:04 crc kubenswrapper[4754]: I1005 22:15:04.398172 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4f599a2cbbb29897cea13b3343d5c76800dbb48769c97fbc7a72a9f5f4928305" Oct 05 22:15:04 crc kubenswrapper[4754]: I1005 22:15:04.398224 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29328375-qkxnn" Oct 05 22:15:04 crc kubenswrapper[4754]: I1005 22:15:04.459925 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4s4rx" Oct 05 22:15:04 crc kubenswrapper[4754]: I1005 22:15:04.846993 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29328330-b7xbb"] Oct 05 22:15:04 crc kubenswrapper[4754]: I1005 22:15:04.849174 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29328330-b7xbb"] Oct 05 22:15:04 crc kubenswrapper[4754]: I1005 22:15:04.915754 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-4s4rx"] Oct 05 22:15:05 crc kubenswrapper[4754]: I1005 22:15:05.244741 4754 patch_prober.go:28] interesting pod/machine-config-daemon-b2h9k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 22:15:05 crc kubenswrapper[4754]: I1005 22:15:05.245009 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 22:15:05 crc kubenswrapper[4754]: I1005 22:15:05.245119 4754 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" Oct 05 22:15:05 crc kubenswrapper[4754]: I1005 22:15:05.245893 4754 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c50ffee964171eed8b7ee0cd69364b26a808f447930df84f79e3a9a2b8b5bab8"} pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 05 22:15:05 crc kubenswrapper[4754]: I1005 22:15:05.246006 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" containerName="machine-config-daemon" containerID="cri-o://c50ffee964171eed8b7ee0cd69364b26a808f447930df84f79e3a9a2b8b5bab8" gracePeriod=600 Oct 05 22:15:05 crc kubenswrapper[4754]: I1005 22:15:05.413800 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4s4rx" event={"ID":"91b810ad-510b-41f2-833b-bf7ce9c2b1b6","Type":"ContainerStarted","Data":"cdf6f3cf228e7ef8556b130537432354dcada6ae9f76576be04ff7134eceec9b"} Oct 05 22:15:06 crc kubenswrapper[4754]: I1005 22:15:06.427673 4754 generic.go:334] "Generic (PLEG): container finished" podID="be95a413-6a59-45b4-84b7-b43cae694a26" containerID="c50ffee964171eed8b7ee0cd69364b26a808f447930df84f79e3a9a2b8b5bab8" exitCode=0 Oct 05 22:15:06 crc kubenswrapper[4754]: I1005 22:15:06.428794 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" event={"ID":"be95a413-6a59-45b4-84b7-b43cae694a26","Type":"ContainerDied","Data":"c50ffee964171eed8b7ee0cd69364b26a808f447930df84f79e3a9a2b8b5bab8"} Oct 05 22:15:06 crc kubenswrapper[4754]: I1005 22:15:06.428892 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" event={"ID":"be95a413-6a59-45b4-84b7-b43cae694a26","Type":"ContainerStarted","Data":"05cd65b5014dfd796e8d71d762a213d6758e74cce340c7705d8831b6a5961a0f"} Oct 05 22:15:06 crc kubenswrapper[4754]: I1005 22:15:06.428959 4754 scope.go:117] "RemoveContainer" containerID="d925cdc080b01050e7b9377dba2f144d30f517d60778d239b53617b071498b40" Oct 05 22:15:06 crc kubenswrapper[4754]: I1005 22:15:06.445323 4754 generic.go:334] "Generic (PLEG): container finished" podID="91b810ad-510b-41f2-833b-bf7ce9c2b1b6" containerID="b6d6ae15eaa18cfde13c8e07bc7214a628150319b083273ae7a4d503acfed333" exitCode=0 Oct 05 22:15:06 crc kubenswrapper[4754]: I1005 22:15:06.445470 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4s4rx" event={"ID":"91b810ad-510b-41f2-833b-bf7ce9c2b1b6","Type":"ContainerDied","Data":"b6d6ae15eaa18cfde13c8e07bc7214a628150319b083273ae7a4d503acfed333"} Oct 05 22:15:06 crc kubenswrapper[4754]: I1005 22:15:06.447902 4754 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 05 22:15:06 crc kubenswrapper[4754]: I1005 22:15:06.858007 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8d619d93-19a1-4ecb-8008-64b1d4dc931a" path="/var/lib/kubelet/pods/8d619d93-19a1-4ecb-8008-64b1d4dc931a/volumes" Oct 05 22:15:08 crc kubenswrapper[4754]: I1005 22:15:08.465781 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4s4rx" event={"ID":"91b810ad-510b-41f2-833b-bf7ce9c2b1b6","Type":"ContainerStarted","Data":"0b7ac95b5534933594bc8884ce99da1f8ed930c5a49731bc4814d0824dfe52fc"} Oct 05 22:15:11 crc kubenswrapper[4754]: I1005 22:15:11.497792 4754 generic.go:334] "Generic (PLEG): container finished" podID="91b810ad-510b-41f2-833b-bf7ce9c2b1b6" containerID="0b7ac95b5534933594bc8884ce99da1f8ed930c5a49731bc4814d0824dfe52fc" exitCode=0 Oct 05 22:15:11 crc kubenswrapper[4754]: I1005 22:15:11.498387 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4s4rx" event={"ID":"91b810ad-510b-41f2-833b-bf7ce9c2b1b6","Type":"ContainerDied","Data":"0b7ac95b5534933594bc8884ce99da1f8ed930c5a49731bc4814d0824dfe52fc"} Oct 05 22:15:12 crc kubenswrapper[4754]: I1005 22:15:12.513374 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4s4rx" event={"ID":"91b810ad-510b-41f2-833b-bf7ce9c2b1b6","Type":"ContainerStarted","Data":"f9187dc3887040017ea0ec1f51885f64a747fefdbfbe95e3059caf5ad218fd1a"} Oct 05 22:15:12 crc kubenswrapper[4754]: I1005 22:15:12.542074 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-4s4rx" podStartSLOduration=3.063889601 podStartE2EDuration="8.542055041s" podCreationTimestamp="2025-10-05 22:15:04 +0000 UTC" firstStartedPulling="2025-10-05 22:15:06.447628087 +0000 UTC m=+4830.351746797" lastFinishedPulling="2025-10-05 22:15:11.925793517 +0000 UTC m=+4835.829912237" observedRunningTime="2025-10-05 22:15:12.53624668 +0000 UTC m=+4836.440365400" watchObservedRunningTime="2025-10-05 22:15:12.542055041 +0000 UTC m=+4836.446173761" Oct 05 22:15:14 crc kubenswrapper[4754]: I1005 22:15:14.461004 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-4s4rx" Oct 05 22:15:14 crc kubenswrapper[4754]: I1005 22:15:14.462685 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-4s4rx" Oct 05 22:15:15 crc kubenswrapper[4754]: I1005 22:15:15.537736 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-4s4rx" podUID="91b810ad-510b-41f2-833b-bf7ce9c2b1b6" containerName="registry-server" probeResult="failure" output=< Oct 05 22:15:15 crc kubenswrapper[4754]: timeout: failed to connect service ":50051" within 1s Oct 05 22:15:15 crc kubenswrapper[4754]: > Oct 05 22:15:25 crc kubenswrapper[4754]: I1005 22:15:25.543881 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-4s4rx" podUID="91b810ad-510b-41f2-833b-bf7ce9c2b1b6" containerName="registry-server" probeResult="failure" output=< Oct 05 22:15:25 crc kubenswrapper[4754]: timeout: failed to connect service ":50051" within 1s Oct 05 22:15:25 crc kubenswrapper[4754]: > Oct 05 22:15:30 crc kubenswrapper[4754]: I1005 22:15:30.438735 4754 scope.go:117] "RemoveContainer" containerID="b85dbda630293941d96625b3609eacbc989e66ec844fb6f26d28401d7f031a57" Oct 05 22:15:34 crc kubenswrapper[4754]: I1005 22:15:34.562540 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-4s4rx" Oct 05 22:15:34 crc kubenswrapper[4754]: I1005 22:15:34.663581 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-4s4rx" Oct 05 22:15:35 crc kubenswrapper[4754]: I1005 22:15:35.370739 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-4s4rx"] Oct 05 22:15:35 crc kubenswrapper[4754]: I1005 22:15:35.784181 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-4s4rx" podUID="91b810ad-510b-41f2-833b-bf7ce9c2b1b6" containerName="registry-server" containerID="cri-o://f9187dc3887040017ea0ec1f51885f64a747fefdbfbe95e3059caf5ad218fd1a" gracePeriod=2 Oct 05 22:15:36 crc kubenswrapper[4754]: I1005 22:15:36.251014 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4s4rx" Oct 05 22:15:36 crc kubenswrapper[4754]: I1005 22:15:36.372935 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91b810ad-510b-41f2-833b-bf7ce9c2b1b6-catalog-content\") pod \"91b810ad-510b-41f2-833b-bf7ce9c2b1b6\" (UID: \"91b810ad-510b-41f2-833b-bf7ce9c2b1b6\") " Oct 05 22:15:36 crc kubenswrapper[4754]: I1005 22:15:36.373140 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91b810ad-510b-41f2-833b-bf7ce9c2b1b6-utilities\") pod \"91b810ad-510b-41f2-833b-bf7ce9c2b1b6\" (UID: \"91b810ad-510b-41f2-833b-bf7ce9c2b1b6\") " Oct 05 22:15:36 crc kubenswrapper[4754]: I1005 22:15:36.373199 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xhrwg\" (UniqueName: \"kubernetes.io/projected/91b810ad-510b-41f2-833b-bf7ce9c2b1b6-kube-api-access-xhrwg\") pod \"91b810ad-510b-41f2-833b-bf7ce9c2b1b6\" (UID: \"91b810ad-510b-41f2-833b-bf7ce9c2b1b6\") " Oct 05 22:15:36 crc kubenswrapper[4754]: I1005 22:15:36.373982 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/91b810ad-510b-41f2-833b-bf7ce9c2b1b6-utilities" (OuterVolumeSpecName: "utilities") pod "91b810ad-510b-41f2-833b-bf7ce9c2b1b6" (UID: "91b810ad-510b-41f2-833b-bf7ce9c2b1b6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 22:15:36 crc kubenswrapper[4754]: I1005 22:15:36.384518 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/91b810ad-510b-41f2-833b-bf7ce9c2b1b6-kube-api-access-xhrwg" (OuterVolumeSpecName: "kube-api-access-xhrwg") pod "91b810ad-510b-41f2-833b-bf7ce9c2b1b6" (UID: "91b810ad-510b-41f2-833b-bf7ce9c2b1b6"). InnerVolumeSpecName "kube-api-access-xhrwg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 22:15:36 crc kubenswrapper[4754]: I1005 22:15:36.444279 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/91b810ad-510b-41f2-833b-bf7ce9c2b1b6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "91b810ad-510b-41f2-833b-bf7ce9c2b1b6" (UID: "91b810ad-510b-41f2-833b-bf7ce9c2b1b6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 22:15:36 crc kubenswrapper[4754]: I1005 22:15:36.475774 4754 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91b810ad-510b-41f2-833b-bf7ce9c2b1b6-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 22:15:36 crc kubenswrapper[4754]: I1005 22:15:36.475811 4754 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91b810ad-510b-41f2-833b-bf7ce9c2b1b6-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 22:15:36 crc kubenswrapper[4754]: I1005 22:15:36.475825 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xhrwg\" (UniqueName: \"kubernetes.io/projected/91b810ad-510b-41f2-833b-bf7ce9c2b1b6-kube-api-access-xhrwg\") on node \"crc\" DevicePath \"\"" Oct 05 22:15:36 crc kubenswrapper[4754]: I1005 22:15:36.799590 4754 generic.go:334] "Generic (PLEG): container finished" podID="91b810ad-510b-41f2-833b-bf7ce9c2b1b6" containerID="f9187dc3887040017ea0ec1f51885f64a747fefdbfbe95e3059caf5ad218fd1a" exitCode=0 Oct 05 22:15:36 crc kubenswrapper[4754]: I1005 22:15:36.799647 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4s4rx" event={"ID":"91b810ad-510b-41f2-833b-bf7ce9c2b1b6","Type":"ContainerDied","Data":"f9187dc3887040017ea0ec1f51885f64a747fefdbfbe95e3059caf5ad218fd1a"} Oct 05 22:15:36 crc kubenswrapper[4754]: I1005 22:15:36.799656 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4s4rx" Oct 05 22:15:36 crc kubenswrapper[4754]: I1005 22:15:36.799679 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4s4rx" event={"ID":"91b810ad-510b-41f2-833b-bf7ce9c2b1b6","Type":"ContainerDied","Data":"cdf6f3cf228e7ef8556b130537432354dcada6ae9f76576be04ff7134eceec9b"} Oct 05 22:15:36 crc kubenswrapper[4754]: I1005 22:15:36.799702 4754 scope.go:117] "RemoveContainer" containerID="f9187dc3887040017ea0ec1f51885f64a747fefdbfbe95e3059caf5ad218fd1a" Oct 05 22:15:36 crc kubenswrapper[4754]: I1005 22:15:36.844816 4754 scope.go:117] "RemoveContainer" containerID="0b7ac95b5534933594bc8884ce99da1f8ed930c5a49731bc4814d0824dfe52fc" Oct 05 22:15:36 crc kubenswrapper[4754]: I1005 22:15:36.854992 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-4s4rx"] Oct 05 22:15:36 crc kubenswrapper[4754]: I1005 22:15:36.855038 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-4s4rx"] Oct 05 22:15:36 crc kubenswrapper[4754]: I1005 22:15:36.873752 4754 scope.go:117] "RemoveContainer" containerID="b6d6ae15eaa18cfde13c8e07bc7214a628150319b083273ae7a4d503acfed333" Oct 05 22:15:36 crc kubenswrapper[4754]: I1005 22:15:36.905456 4754 scope.go:117] "RemoveContainer" containerID="f9187dc3887040017ea0ec1f51885f64a747fefdbfbe95e3059caf5ad218fd1a" Oct 05 22:15:36 crc kubenswrapper[4754]: E1005 22:15:36.906109 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f9187dc3887040017ea0ec1f51885f64a747fefdbfbe95e3059caf5ad218fd1a\": container with ID starting with f9187dc3887040017ea0ec1f51885f64a747fefdbfbe95e3059caf5ad218fd1a not found: ID does not exist" containerID="f9187dc3887040017ea0ec1f51885f64a747fefdbfbe95e3059caf5ad218fd1a" Oct 05 22:15:36 crc kubenswrapper[4754]: I1005 22:15:36.906166 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f9187dc3887040017ea0ec1f51885f64a747fefdbfbe95e3059caf5ad218fd1a"} err="failed to get container status \"f9187dc3887040017ea0ec1f51885f64a747fefdbfbe95e3059caf5ad218fd1a\": rpc error: code = NotFound desc = could not find container \"f9187dc3887040017ea0ec1f51885f64a747fefdbfbe95e3059caf5ad218fd1a\": container with ID starting with f9187dc3887040017ea0ec1f51885f64a747fefdbfbe95e3059caf5ad218fd1a not found: ID does not exist" Oct 05 22:15:36 crc kubenswrapper[4754]: I1005 22:15:36.906200 4754 scope.go:117] "RemoveContainer" containerID="0b7ac95b5534933594bc8884ce99da1f8ed930c5a49731bc4814d0824dfe52fc" Oct 05 22:15:36 crc kubenswrapper[4754]: E1005 22:15:36.906987 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0b7ac95b5534933594bc8884ce99da1f8ed930c5a49731bc4814d0824dfe52fc\": container with ID starting with 0b7ac95b5534933594bc8884ce99da1f8ed930c5a49731bc4814d0824dfe52fc not found: ID does not exist" containerID="0b7ac95b5534933594bc8884ce99da1f8ed930c5a49731bc4814d0824dfe52fc" Oct 05 22:15:36 crc kubenswrapper[4754]: I1005 22:15:36.907029 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0b7ac95b5534933594bc8884ce99da1f8ed930c5a49731bc4814d0824dfe52fc"} err="failed to get container status \"0b7ac95b5534933594bc8884ce99da1f8ed930c5a49731bc4814d0824dfe52fc\": rpc error: code = NotFound desc = could not find container \"0b7ac95b5534933594bc8884ce99da1f8ed930c5a49731bc4814d0824dfe52fc\": container with ID starting with 0b7ac95b5534933594bc8884ce99da1f8ed930c5a49731bc4814d0824dfe52fc not found: ID does not exist" Oct 05 22:15:36 crc kubenswrapper[4754]: I1005 22:15:36.907056 4754 scope.go:117] "RemoveContainer" containerID="b6d6ae15eaa18cfde13c8e07bc7214a628150319b083273ae7a4d503acfed333" Oct 05 22:15:36 crc kubenswrapper[4754]: E1005 22:15:36.907454 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b6d6ae15eaa18cfde13c8e07bc7214a628150319b083273ae7a4d503acfed333\": container with ID starting with b6d6ae15eaa18cfde13c8e07bc7214a628150319b083273ae7a4d503acfed333 not found: ID does not exist" containerID="b6d6ae15eaa18cfde13c8e07bc7214a628150319b083273ae7a4d503acfed333" Oct 05 22:15:36 crc kubenswrapper[4754]: I1005 22:15:36.907527 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b6d6ae15eaa18cfde13c8e07bc7214a628150319b083273ae7a4d503acfed333"} err="failed to get container status \"b6d6ae15eaa18cfde13c8e07bc7214a628150319b083273ae7a4d503acfed333\": rpc error: code = NotFound desc = could not find container \"b6d6ae15eaa18cfde13c8e07bc7214a628150319b083273ae7a4d503acfed333\": container with ID starting with b6d6ae15eaa18cfde13c8e07bc7214a628150319b083273ae7a4d503acfed333 not found: ID does not exist" Oct 05 22:15:38 crc kubenswrapper[4754]: I1005 22:15:38.854287 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="91b810ad-510b-41f2-833b-bf7ce9c2b1b6" path="/var/lib/kubelet/pods/91b810ad-510b-41f2-833b-bf7ce9c2b1b6/volumes" Oct 05 22:15:38 crc kubenswrapper[4754]: E1005 22:15:38.940421 4754 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod91b810ad_510b_41f2_833b_bf7ce9c2b1b6.slice\": RecentStats: unable to find data in memory cache]" Oct 05 22:15:49 crc kubenswrapper[4754]: E1005 22:15:49.180600 4754 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod91b810ad_510b_41f2_833b_bf7ce9c2b1b6.slice\": RecentStats: unable to find data in memory cache]" Oct 05 22:15:59 crc kubenswrapper[4754]: E1005 22:15:59.427076 4754 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod91b810ad_510b_41f2_833b_bf7ce9c2b1b6.slice\": RecentStats: unable to find data in memory cache]" Oct 05 22:16:09 crc kubenswrapper[4754]: E1005 22:16:09.704460 4754 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod91b810ad_510b_41f2_833b_bf7ce9c2b1b6.slice\": RecentStats: unable to find data in memory cache]" Oct 05 22:16:19 crc kubenswrapper[4754]: E1005 22:16:19.934009 4754 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod91b810ad_510b_41f2_833b_bf7ce9c2b1b6.slice\": RecentStats: unable to find data in memory cache]" Oct 05 22:16:28 crc kubenswrapper[4754]: I1005 22:16:28.327030 4754 generic.go:334] "Generic (PLEG): container finished" podID="5d9dc270-0dc5-4682-a4cf-1b61e0e3390b" containerID="79c7a645c7b390dae565f4af0e443592f8f1620a90d1c32ab67f31639e2c89f3" exitCode=0 Oct 05 22:16:28 crc kubenswrapper[4754]: I1005 22:16:28.327089 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-44bn6/must-gather-l24vb" event={"ID":"5d9dc270-0dc5-4682-a4cf-1b61e0e3390b","Type":"ContainerDied","Data":"79c7a645c7b390dae565f4af0e443592f8f1620a90d1c32ab67f31639e2c89f3"} Oct 05 22:16:28 crc kubenswrapper[4754]: I1005 22:16:28.336036 4754 scope.go:117] "RemoveContainer" containerID="79c7a645c7b390dae565f4af0e443592f8f1620a90d1c32ab67f31639e2c89f3" Oct 05 22:16:28 crc kubenswrapper[4754]: I1005 22:16:28.827616 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-44bn6_must-gather-l24vb_5d9dc270-0dc5-4682-a4cf-1b61e0e3390b/gather/0.log" Oct 05 22:16:30 crc kubenswrapper[4754]: E1005 22:16:30.198654 4754 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod91b810ad_510b_41f2_833b_bf7ce9c2b1b6.slice\": RecentStats: unable to find data in memory cache]" Oct 05 22:16:30 crc kubenswrapper[4754]: I1005 22:16:30.530410 4754 scope.go:117] "RemoveContainer" containerID="af5377fba7a759eff76ab1fcecce4f748f0fa0b79bd410cb77230eef9b5204e5" Oct 05 22:16:38 crc kubenswrapper[4754]: I1005 22:16:38.332890 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-44bn6/must-gather-l24vb"] Oct 05 22:16:38 crc kubenswrapper[4754]: I1005 22:16:38.333649 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-44bn6/must-gather-l24vb" podUID="5d9dc270-0dc5-4682-a4cf-1b61e0e3390b" containerName="copy" containerID="cri-o://1eae1877e73e7e6bc401df6113094f06180449c379bbdd6e8951cce4dfa749a3" gracePeriod=2 Oct 05 22:16:38 crc kubenswrapper[4754]: I1005 22:16:38.342696 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-44bn6/must-gather-l24vb"] Oct 05 22:16:38 crc kubenswrapper[4754]: I1005 22:16:38.735545 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-44bn6_must-gather-l24vb_5d9dc270-0dc5-4682-a4cf-1b61e0e3390b/copy/0.log" Oct 05 22:16:38 crc kubenswrapper[4754]: I1005 22:16:38.736559 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-44bn6/must-gather-l24vb" Oct 05 22:16:38 crc kubenswrapper[4754]: I1005 22:16:38.809814 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wrxg2\" (UniqueName: \"kubernetes.io/projected/5d9dc270-0dc5-4682-a4cf-1b61e0e3390b-kube-api-access-wrxg2\") pod \"5d9dc270-0dc5-4682-a4cf-1b61e0e3390b\" (UID: \"5d9dc270-0dc5-4682-a4cf-1b61e0e3390b\") " Oct 05 22:16:38 crc kubenswrapper[4754]: I1005 22:16:38.809860 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/5d9dc270-0dc5-4682-a4cf-1b61e0e3390b-must-gather-output\") pod \"5d9dc270-0dc5-4682-a4cf-1b61e0e3390b\" (UID: \"5d9dc270-0dc5-4682-a4cf-1b61e0e3390b\") " Oct 05 22:16:38 crc kubenswrapper[4754]: I1005 22:16:38.815229 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5d9dc270-0dc5-4682-a4cf-1b61e0e3390b-kube-api-access-wrxg2" (OuterVolumeSpecName: "kube-api-access-wrxg2") pod "5d9dc270-0dc5-4682-a4cf-1b61e0e3390b" (UID: "5d9dc270-0dc5-4682-a4cf-1b61e0e3390b"). InnerVolumeSpecName "kube-api-access-wrxg2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 22:16:38 crc kubenswrapper[4754]: I1005 22:16:38.911619 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wrxg2\" (UniqueName: \"kubernetes.io/projected/5d9dc270-0dc5-4682-a4cf-1b61e0e3390b-kube-api-access-wrxg2\") on node \"crc\" DevicePath \"\"" Oct 05 22:16:38 crc kubenswrapper[4754]: I1005 22:16:38.968319 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5d9dc270-0dc5-4682-a4cf-1b61e0e3390b-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "5d9dc270-0dc5-4682-a4cf-1b61e0e3390b" (UID: "5d9dc270-0dc5-4682-a4cf-1b61e0e3390b"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 22:16:39 crc kubenswrapper[4754]: I1005 22:16:39.013318 4754 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/5d9dc270-0dc5-4682-a4cf-1b61e0e3390b-must-gather-output\") on node \"crc\" DevicePath \"\"" Oct 05 22:16:39 crc kubenswrapper[4754]: I1005 22:16:39.459339 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-44bn6_must-gather-l24vb_5d9dc270-0dc5-4682-a4cf-1b61e0e3390b/copy/0.log" Oct 05 22:16:39 crc kubenswrapper[4754]: I1005 22:16:39.460083 4754 generic.go:334] "Generic (PLEG): container finished" podID="5d9dc270-0dc5-4682-a4cf-1b61e0e3390b" containerID="1eae1877e73e7e6bc401df6113094f06180449c379bbdd6e8951cce4dfa749a3" exitCode=143 Oct 05 22:16:39 crc kubenswrapper[4754]: I1005 22:16:39.460136 4754 scope.go:117] "RemoveContainer" containerID="1eae1877e73e7e6bc401df6113094f06180449c379bbdd6e8951cce4dfa749a3" Oct 05 22:16:39 crc kubenswrapper[4754]: I1005 22:16:39.460182 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-44bn6/must-gather-l24vb" Oct 05 22:16:39 crc kubenswrapper[4754]: I1005 22:16:39.504528 4754 scope.go:117] "RemoveContainer" containerID="79c7a645c7b390dae565f4af0e443592f8f1620a90d1c32ab67f31639e2c89f3" Oct 05 22:16:39 crc kubenswrapper[4754]: I1005 22:16:39.563235 4754 scope.go:117] "RemoveContainer" containerID="1eae1877e73e7e6bc401df6113094f06180449c379bbdd6e8951cce4dfa749a3" Oct 05 22:16:39 crc kubenswrapper[4754]: E1005 22:16:39.563744 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1eae1877e73e7e6bc401df6113094f06180449c379bbdd6e8951cce4dfa749a3\": container with ID starting with 1eae1877e73e7e6bc401df6113094f06180449c379bbdd6e8951cce4dfa749a3 not found: ID does not exist" containerID="1eae1877e73e7e6bc401df6113094f06180449c379bbdd6e8951cce4dfa749a3" Oct 05 22:16:39 crc kubenswrapper[4754]: I1005 22:16:39.563802 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1eae1877e73e7e6bc401df6113094f06180449c379bbdd6e8951cce4dfa749a3"} err="failed to get container status \"1eae1877e73e7e6bc401df6113094f06180449c379bbdd6e8951cce4dfa749a3\": rpc error: code = NotFound desc = could not find container \"1eae1877e73e7e6bc401df6113094f06180449c379bbdd6e8951cce4dfa749a3\": container with ID starting with 1eae1877e73e7e6bc401df6113094f06180449c379bbdd6e8951cce4dfa749a3 not found: ID does not exist" Oct 05 22:16:39 crc kubenswrapper[4754]: I1005 22:16:39.563831 4754 scope.go:117] "RemoveContainer" containerID="79c7a645c7b390dae565f4af0e443592f8f1620a90d1c32ab67f31639e2c89f3" Oct 05 22:16:39 crc kubenswrapper[4754]: E1005 22:16:39.564412 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"79c7a645c7b390dae565f4af0e443592f8f1620a90d1c32ab67f31639e2c89f3\": container with ID starting with 79c7a645c7b390dae565f4af0e443592f8f1620a90d1c32ab67f31639e2c89f3 not found: ID does not exist" containerID="79c7a645c7b390dae565f4af0e443592f8f1620a90d1c32ab67f31639e2c89f3" Oct 05 22:16:39 crc kubenswrapper[4754]: I1005 22:16:39.564454 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"79c7a645c7b390dae565f4af0e443592f8f1620a90d1c32ab67f31639e2c89f3"} err="failed to get container status \"79c7a645c7b390dae565f4af0e443592f8f1620a90d1c32ab67f31639e2c89f3\": rpc error: code = NotFound desc = could not find container \"79c7a645c7b390dae565f4af0e443592f8f1620a90d1c32ab67f31639e2c89f3\": container with ID starting with 79c7a645c7b390dae565f4af0e443592f8f1620a90d1c32ab67f31639e2c89f3 not found: ID does not exist" Oct 05 22:16:40 crc kubenswrapper[4754]: I1005 22:16:40.851121 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5d9dc270-0dc5-4682-a4cf-1b61e0e3390b" path="/var/lib/kubelet/pods/5d9dc270-0dc5-4682-a4cf-1b61e0e3390b/volumes" Oct 05 22:17:05 crc kubenswrapper[4754]: I1005 22:17:05.245182 4754 patch_prober.go:28] interesting pod/machine-config-daemon-b2h9k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 22:17:05 crc kubenswrapper[4754]: I1005 22:17:05.245972 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 22:17:17 crc kubenswrapper[4754]: I1005 22:17:17.764786 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-tv57x/must-gather-jqmds"] Oct 05 22:17:17 crc kubenswrapper[4754]: E1005 22:17:17.765678 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91b810ad-510b-41f2-833b-bf7ce9c2b1b6" containerName="extract-content" Oct 05 22:17:17 crc kubenswrapper[4754]: I1005 22:17:17.765692 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="91b810ad-510b-41f2-833b-bf7ce9c2b1b6" containerName="extract-content" Oct 05 22:17:17 crc kubenswrapper[4754]: E1005 22:17:17.765714 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d9dc270-0dc5-4682-a4cf-1b61e0e3390b" containerName="gather" Oct 05 22:17:17 crc kubenswrapper[4754]: I1005 22:17:17.765722 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d9dc270-0dc5-4682-a4cf-1b61e0e3390b" containerName="gather" Oct 05 22:17:17 crc kubenswrapper[4754]: E1005 22:17:17.765745 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91b810ad-510b-41f2-833b-bf7ce9c2b1b6" containerName="extract-utilities" Oct 05 22:17:17 crc kubenswrapper[4754]: I1005 22:17:17.765753 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="91b810ad-510b-41f2-833b-bf7ce9c2b1b6" containerName="extract-utilities" Oct 05 22:17:17 crc kubenswrapper[4754]: E1005 22:17:17.765771 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d9dc270-0dc5-4682-a4cf-1b61e0e3390b" containerName="copy" Oct 05 22:17:17 crc kubenswrapper[4754]: I1005 22:17:17.765780 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d9dc270-0dc5-4682-a4cf-1b61e0e3390b" containerName="copy" Oct 05 22:17:17 crc kubenswrapper[4754]: E1005 22:17:17.765799 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91b810ad-510b-41f2-833b-bf7ce9c2b1b6" containerName="registry-server" Oct 05 22:17:17 crc kubenswrapper[4754]: I1005 22:17:17.765807 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="91b810ad-510b-41f2-833b-bf7ce9c2b1b6" containerName="registry-server" Oct 05 22:17:17 crc kubenswrapper[4754]: I1005 22:17:17.766022 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="91b810ad-510b-41f2-833b-bf7ce9c2b1b6" containerName="registry-server" Oct 05 22:17:17 crc kubenswrapper[4754]: I1005 22:17:17.766040 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="5d9dc270-0dc5-4682-a4cf-1b61e0e3390b" containerName="copy" Oct 05 22:17:17 crc kubenswrapper[4754]: I1005 22:17:17.766052 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="5d9dc270-0dc5-4682-a4cf-1b61e0e3390b" containerName="gather" Oct 05 22:17:17 crc kubenswrapper[4754]: I1005 22:17:17.767156 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-tv57x/must-gather-jqmds" Oct 05 22:17:17 crc kubenswrapper[4754]: I1005 22:17:17.768846 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-tv57x"/"default-dockercfg-hksbp" Oct 05 22:17:17 crc kubenswrapper[4754]: I1005 22:17:17.769335 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-tv57x"/"openshift-service-ca.crt" Oct 05 22:17:17 crc kubenswrapper[4754]: I1005 22:17:17.771103 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-tv57x"/"kube-root-ca.crt" Oct 05 22:17:17 crc kubenswrapper[4754]: I1005 22:17:17.790568 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-tv57x/must-gather-jqmds"] Oct 05 22:17:17 crc kubenswrapper[4754]: I1005 22:17:17.888515 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/3706afbe-bdd8-4410-abe2-889a0c63d1ce-must-gather-output\") pod \"must-gather-jqmds\" (UID: \"3706afbe-bdd8-4410-abe2-889a0c63d1ce\") " pod="openshift-must-gather-tv57x/must-gather-jqmds" Oct 05 22:17:17 crc kubenswrapper[4754]: I1005 22:17:17.888564 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5xjmp\" (UniqueName: \"kubernetes.io/projected/3706afbe-bdd8-4410-abe2-889a0c63d1ce-kube-api-access-5xjmp\") pod \"must-gather-jqmds\" (UID: \"3706afbe-bdd8-4410-abe2-889a0c63d1ce\") " pod="openshift-must-gather-tv57x/must-gather-jqmds" Oct 05 22:17:17 crc kubenswrapper[4754]: I1005 22:17:17.991061 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/3706afbe-bdd8-4410-abe2-889a0c63d1ce-must-gather-output\") pod \"must-gather-jqmds\" (UID: \"3706afbe-bdd8-4410-abe2-889a0c63d1ce\") " pod="openshift-must-gather-tv57x/must-gather-jqmds" Oct 05 22:17:17 crc kubenswrapper[4754]: I1005 22:17:17.991105 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5xjmp\" (UniqueName: \"kubernetes.io/projected/3706afbe-bdd8-4410-abe2-889a0c63d1ce-kube-api-access-5xjmp\") pod \"must-gather-jqmds\" (UID: \"3706afbe-bdd8-4410-abe2-889a0c63d1ce\") " pod="openshift-must-gather-tv57x/must-gather-jqmds" Oct 05 22:17:17 crc kubenswrapper[4754]: I1005 22:17:17.991508 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/3706afbe-bdd8-4410-abe2-889a0c63d1ce-must-gather-output\") pod \"must-gather-jqmds\" (UID: \"3706afbe-bdd8-4410-abe2-889a0c63d1ce\") " pod="openshift-must-gather-tv57x/must-gather-jqmds" Oct 05 22:17:18 crc kubenswrapper[4754]: I1005 22:17:18.007650 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5xjmp\" (UniqueName: \"kubernetes.io/projected/3706afbe-bdd8-4410-abe2-889a0c63d1ce-kube-api-access-5xjmp\") pod \"must-gather-jqmds\" (UID: \"3706afbe-bdd8-4410-abe2-889a0c63d1ce\") " pod="openshift-must-gather-tv57x/must-gather-jqmds" Oct 05 22:17:18 crc kubenswrapper[4754]: I1005 22:17:18.088331 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-tv57x/must-gather-jqmds" Oct 05 22:17:18 crc kubenswrapper[4754]: I1005 22:17:18.570582 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-tv57x/must-gather-jqmds"] Oct 05 22:17:18 crc kubenswrapper[4754]: I1005 22:17:18.864191 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-tv57x/must-gather-jqmds" event={"ID":"3706afbe-bdd8-4410-abe2-889a0c63d1ce","Type":"ContainerStarted","Data":"759d2f194ecac2116f02c4df51a199c4441bc8c62683bb9aec864812932aea09"} Oct 05 22:17:19 crc kubenswrapper[4754]: I1005 22:17:19.878107 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-tv57x/must-gather-jqmds" event={"ID":"3706afbe-bdd8-4410-abe2-889a0c63d1ce","Type":"ContainerStarted","Data":"9cc22ef436639700c0f5487d98a4ffe30c7bf9325abb0d4e0a38b44835327e6c"} Oct 05 22:17:19 crc kubenswrapper[4754]: I1005 22:17:19.878920 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-tv57x/must-gather-jqmds" event={"ID":"3706afbe-bdd8-4410-abe2-889a0c63d1ce","Type":"ContainerStarted","Data":"6ea400c438366f1dc5a9bc9b46ac924bc362770d0ee334f3401c13a6244d0496"} Oct 05 22:17:19 crc kubenswrapper[4754]: I1005 22:17:19.899511 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-tv57x/must-gather-jqmds" podStartSLOduration=2.899482673 podStartE2EDuration="2.899482673s" podCreationTimestamp="2025-10-05 22:17:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 22:17:19.89163911 +0000 UTC m=+4963.795757820" watchObservedRunningTime="2025-10-05 22:17:19.899482673 +0000 UTC m=+4963.803601383" Oct 05 22:17:22 crc kubenswrapper[4754]: I1005 22:17:22.720488 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-tv57x/crc-debug-sxp85"] Oct 05 22:17:22 crc kubenswrapper[4754]: I1005 22:17:22.722057 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-tv57x/crc-debug-sxp85" Oct 05 22:17:22 crc kubenswrapper[4754]: I1005 22:17:22.831365 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8rjcn\" (UniqueName: \"kubernetes.io/projected/da7c6e68-184a-4ef7-83d6-7fb4a2f52638-kube-api-access-8rjcn\") pod \"crc-debug-sxp85\" (UID: \"da7c6e68-184a-4ef7-83d6-7fb4a2f52638\") " pod="openshift-must-gather-tv57x/crc-debug-sxp85" Oct 05 22:17:22 crc kubenswrapper[4754]: I1005 22:17:22.832520 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/da7c6e68-184a-4ef7-83d6-7fb4a2f52638-host\") pod \"crc-debug-sxp85\" (UID: \"da7c6e68-184a-4ef7-83d6-7fb4a2f52638\") " pod="openshift-must-gather-tv57x/crc-debug-sxp85" Oct 05 22:17:22 crc kubenswrapper[4754]: I1005 22:17:22.934435 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8rjcn\" (UniqueName: \"kubernetes.io/projected/da7c6e68-184a-4ef7-83d6-7fb4a2f52638-kube-api-access-8rjcn\") pod \"crc-debug-sxp85\" (UID: \"da7c6e68-184a-4ef7-83d6-7fb4a2f52638\") " pod="openshift-must-gather-tv57x/crc-debug-sxp85" Oct 05 22:17:22 crc kubenswrapper[4754]: I1005 22:17:22.935389 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/da7c6e68-184a-4ef7-83d6-7fb4a2f52638-host\") pod \"crc-debug-sxp85\" (UID: \"da7c6e68-184a-4ef7-83d6-7fb4a2f52638\") " pod="openshift-must-gather-tv57x/crc-debug-sxp85" Oct 05 22:17:22 crc kubenswrapper[4754]: I1005 22:17:22.935512 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/da7c6e68-184a-4ef7-83d6-7fb4a2f52638-host\") pod \"crc-debug-sxp85\" (UID: \"da7c6e68-184a-4ef7-83d6-7fb4a2f52638\") " pod="openshift-must-gather-tv57x/crc-debug-sxp85" Oct 05 22:17:22 crc kubenswrapper[4754]: I1005 22:17:22.957731 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8rjcn\" (UniqueName: \"kubernetes.io/projected/da7c6e68-184a-4ef7-83d6-7fb4a2f52638-kube-api-access-8rjcn\") pod \"crc-debug-sxp85\" (UID: \"da7c6e68-184a-4ef7-83d6-7fb4a2f52638\") " pod="openshift-must-gather-tv57x/crc-debug-sxp85" Oct 05 22:17:23 crc kubenswrapper[4754]: I1005 22:17:23.036935 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-tv57x/crc-debug-sxp85" Oct 05 22:17:23 crc kubenswrapper[4754]: W1005 22:17:23.074825 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podda7c6e68_184a_4ef7_83d6_7fb4a2f52638.slice/crio-4e8d5b90abd6fbb0af4f44b34eb94eb4381e01565afb31c60ddd9674d7925e90 WatchSource:0}: Error finding container 4e8d5b90abd6fbb0af4f44b34eb94eb4381e01565afb31c60ddd9674d7925e90: Status 404 returned error can't find the container with id 4e8d5b90abd6fbb0af4f44b34eb94eb4381e01565afb31c60ddd9674d7925e90 Oct 05 22:17:24 crc kubenswrapper[4754]: I1005 22:17:24.015920 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-tv57x/crc-debug-sxp85" event={"ID":"da7c6e68-184a-4ef7-83d6-7fb4a2f52638","Type":"ContainerStarted","Data":"4f27cae5dddc7591d87022a2de708593963ba14a366abd1e44e5b438ab79efb4"} Oct 05 22:17:24 crc kubenswrapper[4754]: I1005 22:17:24.016619 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-tv57x/crc-debug-sxp85" event={"ID":"da7c6e68-184a-4ef7-83d6-7fb4a2f52638","Type":"ContainerStarted","Data":"4e8d5b90abd6fbb0af4f44b34eb94eb4381e01565afb31c60ddd9674d7925e90"} Oct 05 22:17:24 crc kubenswrapper[4754]: I1005 22:17:24.036923 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-tv57x/crc-debug-sxp85" podStartSLOduration=2.036906585 podStartE2EDuration="2.036906585s" podCreationTimestamp="2025-10-05 22:17:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 22:17:24.031137545 +0000 UTC m=+4967.935256255" watchObservedRunningTime="2025-10-05 22:17:24.036906585 +0000 UTC m=+4967.941025295" Oct 05 22:17:35 crc kubenswrapper[4754]: I1005 22:17:35.245904 4754 patch_prober.go:28] interesting pod/machine-config-daemon-b2h9k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 22:17:35 crc kubenswrapper[4754]: I1005 22:17:35.246326 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 22:17:49 crc kubenswrapper[4754]: I1005 22:17:49.900310 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-fmpcl"] Oct 05 22:17:49 crc kubenswrapper[4754]: I1005 22:17:49.903889 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fmpcl" Oct 05 22:17:49 crc kubenswrapper[4754]: I1005 22:17:49.908547 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-fmpcl"] Oct 05 22:17:50 crc kubenswrapper[4754]: I1005 22:17:50.098858 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9c284318-3b01-4dd5-8aae-18253bc07696-utilities\") pod \"redhat-marketplace-fmpcl\" (UID: \"9c284318-3b01-4dd5-8aae-18253bc07696\") " pod="openshift-marketplace/redhat-marketplace-fmpcl" Oct 05 22:17:50 crc kubenswrapper[4754]: I1005 22:17:50.098921 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9c284318-3b01-4dd5-8aae-18253bc07696-catalog-content\") pod \"redhat-marketplace-fmpcl\" (UID: \"9c284318-3b01-4dd5-8aae-18253bc07696\") " pod="openshift-marketplace/redhat-marketplace-fmpcl" Oct 05 22:17:50 crc kubenswrapper[4754]: I1005 22:17:50.098966 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-64x9s\" (UniqueName: \"kubernetes.io/projected/9c284318-3b01-4dd5-8aae-18253bc07696-kube-api-access-64x9s\") pod \"redhat-marketplace-fmpcl\" (UID: \"9c284318-3b01-4dd5-8aae-18253bc07696\") " pod="openshift-marketplace/redhat-marketplace-fmpcl" Oct 05 22:17:50 crc kubenswrapper[4754]: I1005 22:17:50.203437 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9c284318-3b01-4dd5-8aae-18253bc07696-utilities\") pod \"redhat-marketplace-fmpcl\" (UID: \"9c284318-3b01-4dd5-8aae-18253bc07696\") " pod="openshift-marketplace/redhat-marketplace-fmpcl" Oct 05 22:17:50 crc kubenswrapper[4754]: I1005 22:17:50.204559 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9c284318-3b01-4dd5-8aae-18253bc07696-catalog-content\") pod \"redhat-marketplace-fmpcl\" (UID: \"9c284318-3b01-4dd5-8aae-18253bc07696\") " pod="openshift-marketplace/redhat-marketplace-fmpcl" Oct 05 22:17:50 crc kubenswrapper[4754]: I1005 22:17:50.204703 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-64x9s\" (UniqueName: \"kubernetes.io/projected/9c284318-3b01-4dd5-8aae-18253bc07696-kube-api-access-64x9s\") pod \"redhat-marketplace-fmpcl\" (UID: \"9c284318-3b01-4dd5-8aae-18253bc07696\") " pod="openshift-marketplace/redhat-marketplace-fmpcl" Oct 05 22:17:50 crc kubenswrapper[4754]: I1005 22:17:50.204486 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9c284318-3b01-4dd5-8aae-18253bc07696-utilities\") pod \"redhat-marketplace-fmpcl\" (UID: \"9c284318-3b01-4dd5-8aae-18253bc07696\") " pod="openshift-marketplace/redhat-marketplace-fmpcl" Oct 05 22:17:50 crc kubenswrapper[4754]: I1005 22:17:50.206979 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9c284318-3b01-4dd5-8aae-18253bc07696-catalog-content\") pod \"redhat-marketplace-fmpcl\" (UID: \"9c284318-3b01-4dd5-8aae-18253bc07696\") " pod="openshift-marketplace/redhat-marketplace-fmpcl" Oct 05 22:17:50 crc kubenswrapper[4754]: I1005 22:17:50.249009 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-64x9s\" (UniqueName: \"kubernetes.io/projected/9c284318-3b01-4dd5-8aae-18253bc07696-kube-api-access-64x9s\") pod \"redhat-marketplace-fmpcl\" (UID: \"9c284318-3b01-4dd5-8aae-18253bc07696\") " pod="openshift-marketplace/redhat-marketplace-fmpcl" Oct 05 22:17:50 crc kubenswrapper[4754]: I1005 22:17:50.268031 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fmpcl" Oct 05 22:17:50 crc kubenswrapper[4754]: I1005 22:17:50.869354 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-fmpcl"] Oct 05 22:17:51 crc kubenswrapper[4754]: I1005 22:17:51.260370 4754 generic.go:334] "Generic (PLEG): container finished" podID="9c284318-3b01-4dd5-8aae-18253bc07696" containerID="bd9c4e885e807d959094fad9e1dc8a8db3a778431b2ae6270dde5281f9a2a3ba" exitCode=0 Oct 05 22:17:51 crc kubenswrapper[4754]: I1005 22:17:51.260467 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fmpcl" event={"ID":"9c284318-3b01-4dd5-8aae-18253bc07696","Type":"ContainerDied","Data":"bd9c4e885e807d959094fad9e1dc8a8db3a778431b2ae6270dde5281f9a2a3ba"} Oct 05 22:17:51 crc kubenswrapper[4754]: I1005 22:17:51.260695 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fmpcl" event={"ID":"9c284318-3b01-4dd5-8aae-18253bc07696","Type":"ContainerStarted","Data":"aa8ec181b4f57faa1c12707cf50cf93f603444c4ca7edc32bb7139e934bb84a5"} Oct 05 22:17:52 crc kubenswrapper[4754]: I1005 22:17:52.281630 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fmpcl" event={"ID":"9c284318-3b01-4dd5-8aae-18253bc07696","Type":"ContainerStarted","Data":"bcfb48096f3ce8cd5b67033dc9bc624a438998623d150dd6f1ae88702aee3b73"} Oct 05 22:17:53 crc kubenswrapper[4754]: I1005 22:17:53.291974 4754 generic.go:334] "Generic (PLEG): container finished" podID="9c284318-3b01-4dd5-8aae-18253bc07696" containerID="bcfb48096f3ce8cd5b67033dc9bc624a438998623d150dd6f1ae88702aee3b73" exitCode=0 Oct 05 22:17:53 crc kubenswrapper[4754]: I1005 22:17:53.293079 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fmpcl" event={"ID":"9c284318-3b01-4dd5-8aae-18253bc07696","Type":"ContainerDied","Data":"bcfb48096f3ce8cd5b67033dc9bc624a438998623d150dd6f1ae88702aee3b73"} Oct 05 22:17:54 crc kubenswrapper[4754]: I1005 22:17:54.305610 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fmpcl" event={"ID":"9c284318-3b01-4dd5-8aae-18253bc07696","Type":"ContainerStarted","Data":"8e3230261c44141824aba256bb53b687e5bc8ecf68ac69d020e161e59a67f80d"} Oct 05 22:17:54 crc kubenswrapper[4754]: I1005 22:17:54.328321 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-fmpcl" podStartSLOduration=2.8989670480000003 podStartE2EDuration="5.328306348s" podCreationTimestamp="2025-10-05 22:17:49 +0000 UTC" firstStartedPulling="2025-10-05 22:17:51.264499852 +0000 UTC m=+4995.168618552" lastFinishedPulling="2025-10-05 22:17:53.693839152 +0000 UTC m=+4997.597957852" observedRunningTime="2025-10-05 22:17:54.321106701 +0000 UTC m=+4998.225225411" watchObservedRunningTime="2025-10-05 22:17:54.328306348 +0000 UTC m=+4998.232425058" Oct 05 22:18:00 crc kubenswrapper[4754]: I1005 22:18:00.268860 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-fmpcl" Oct 05 22:18:00 crc kubenswrapper[4754]: I1005 22:18:00.270084 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-fmpcl" Oct 05 22:18:00 crc kubenswrapper[4754]: I1005 22:18:00.324078 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-fmpcl" Oct 05 22:18:00 crc kubenswrapper[4754]: I1005 22:18:00.399009 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-fmpcl" Oct 05 22:18:00 crc kubenswrapper[4754]: I1005 22:18:00.565036 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-fmpcl"] Oct 05 22:18:02 crc kubenswrapper[4754]: I1005 22:18:02.383677 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-fmpcl" podUID="9c284318-3b01-4dd5-8aae-18253bc07696" containerName="registry-server" containerID="cri-o://8e3230261c44141824aba256bb53b687e5bc8ecf68ac69d020e161e59a67f80d" gracePeriod=2 Oct 05 22:18:03 crc kubenswrapper[4754]: I1005 22:18:03.419983 4754 generic.go:334] "Generic (PLEG): container finished" podID="9c284318-3b01-4dd5-8aae-18253bc07696" containerID="8e3230261c44141824aba256bb53b687e5bc8ecf68ac69d020e161e59a67f80d" exitCode=0 Oct 05 22:18:03 crc kubenswrapper[4754]: I1005 22:18:03.420447 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fmpcl" event={"ID":"9c284318-3b01-4dd5-8aae-18253bc07696","Type":"ContainerDied","Data":"8e3230261c44141824aba256bb53b687e5bc8ecf68ac69d020e161e59a67f80d"} Oct 05 22:18:03 crc kubenswrapper[4754]: I1005 22:18:03.503671 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fmpcl" Oct 05 22:18:03 crc kubenswrapper[4754]: I1005 22:18:03.608087 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-64x9s\" (UniqueName: \"kubernetes.io/projected/9c284318-3b01-4dd5-8aae-18253bc07696-kube-api-access-64x9s\") pod \"9c284318-3b01-4dd5-8aae-18253bc07696\" (UID: \"9c284318-3b01-4dd5-8aae-18253bc07696\") " Oct 05 22:18:03 crc kubenswrapper[4754]: I1005 22:18:03.608206 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9c284318-3b01-4dd5-8aae-18253bc07696-utilities\") pod \"9c284318-3b01-4dd5-8aae-18253bc07696\" (UID: \"9c284318-3b01-4dd5-8aae-18253bc07696\") " Oct 05 22:18:03 crc kubenswrapper[4754]: I1005 22:18:03.608311 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9c284318-3b01-4dd5-8aae-18253bc07696-catalog-content\") pod \"9c284318-3b01-4dd5-8aae-18253bc07696\" (UID: \"9c284318-3b01-4dd5-8aae-18253bc07696\") " Oct 05 22:18:03 crc kubenswrapper[4754]: I1005 22:18:03.609452 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9c284318-3b01-4dd5-8aae-18253bc07696-utilities" (OuterVolumeSpecName: "utilities") pod "9c284318-3b01-4dd5-8aae-18253bc07696" (UID: "9c284318-3b01-4dd5-8aae-18253bc07696"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 22:18:03 crc kubenswrapper[4754]: I1005 22:18:03.623262 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9c284318-3b01-4dd5-8aae-18253bc07696-kube-api-access-64x9s" (OuterVolumeSpecName: "kube-api-access-64x9s") pod "9c284318-3b01-4dd5-8aae-18253bc07696" (UID: "9c284318-3b01-4dd5-8aae-18253bc07696"). InnerVolumeSpecName "kube-api-access-64x9s". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 22:18:03 crc kubenswrapper[4754]: I1005 22:18:03.626227 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9c284318-3b01-4dd5-8aae-18253bc07696-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9c284318-3b01-4dd5-8aae-18253bc07696" (UID: "9c284318-3b01-4dd5-8aae-18253bc07696"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 22:18:03 crc kubenswrapper[4754]: I1005 22:18:03.710886 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-64x9s\" (UniqueName: \"kubernetes.io/projected/9c284318-3b01-4dd5-8aae-18253bc07696-kube-api-access-64x9s\") on node \"crc\" DevicePath \"\"" Oct 05 22:18:03 crc kubenswrapper[4754]: I1005 22:18:03.710919 4754 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9c284318-3b01-4dd5-8aae-18253bc07696-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 22:18:03 crc kubenswrapper[4754]: I1005 22:18:03.710928 4754 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9c284318-3b01-4dd5-8aae-18253bc07696-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 22:18:04 crc kubenswrapper[4754]: I1005 22:18:04.431196 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fmpcl" event={"ID":"9c284318-3b01-4dd5-8aae-18253bc07696","Type":"ContainerDied","Data":"aa8ec181b4f57faa1c12707cf50cf93f603444c4ca7edc32bb7139e934bb84a5"} Oct 05 22:18:04 crc kubenswrapper[4754]: I1005 22:18:04.432353 4754 scope.go:117] "RemoveContainer" containerID="8e3230261c44141824aba256bb53b687e5bc8ecf68ac69d020e161e59a67f80d" Oct 05 22:18:04 crc kubenswrapper[4754]: I1005 22:18:04.431255 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fmpcl" Oct 05 22:18:04 crc kubenswrapper[4754]: I1005 22:18:04.460368 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-fmpcl"] Oct 05 22:18:04 crc kubenswrapper[4754]: I1005 22:18:04.460569 4754 scope.go:117] "RemoveContainer" containerID="bcfb48096f3ce8cd5b67033dc9bc624a438998623d150dd6f1ae88702aee3b73" Oct 05 22:18:04 crc kubenswrapper[4754]: I1005 22:18:04.472428 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-fmpcl"] Oct 05 22:18:04 crc kubenswrapper[4754]: I1005 22:18:04.487719 4754 scope.go:117] "RemoveContainer" containerID="bd9c4e885e807d959094fad9e1dc8a8db3a778431b2ae6270dde5281f9a2a3ba" Oct 05 22:18:04 crc kubenswrapper[4754]: I1005 22:18:04.852574 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9c284318-3b01-4dd5-8aae-18253bc07696" path="/var/lib/kubelet/pods/9c284318-3b01-4dd5-8aae-18253bc07696/volumes" Oct 05 22:18:05 crc kubenswrapper[4754]: I1005 22:18:05.244778 4754 patch_prober.go:28] interesting pod/machine-config-daemon-b2h9k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 22:18:05 crc kubenswrapper[4754]: I1005 22:18:05.244823 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 22:18:05 crc kubenswrapper[4754]: I1005 22:18:05.244861 4754 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" Oct 05 22:18:05 crc kubenswrapper[4754]: I1005 22:18:05.245567 4754 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"05cd65b5014dfd796e8d71d762a213d6758e74cce340c7705d8831b6a5961a0f"} pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 05 22:18:05 crc kubenswrapper[4754]: I1005 22:18:05.245622 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" containerName="machine-config-daemon" containerID="cri-o://05cd65b5014dfd796e8d71d762a213d6758e74cce340c7705d8831b6a5961a0f" gracePeriod=600 Oct 05 22:18:05 crc kubenswrapper[4754]: E1005 22:18:05.367590 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 22:18:05 crc kubenswrapper[4754]: I1005 22:18:05.441281 4754 generic.go:334] "Generic (PLEG): container finished" podID="be95a413-6a59-45b4-84b7-b43cae694a26" containerID="05cd65b5014dfd796e8d71d762a213d6758e74cce340c7705d8831b6a5961a0f" exitCode=0 Oct 05 22:18:05 crc kubenswrapper[4754]: I1005 22:18:05.441343 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" event={"ID":"be95a413-6a59-45b4-84b7-b43cae694a26","Type":"ContainerDied","Data":"05cd65b5014dfd796e8d71d762a213d6758e74cce340c7705d8831b6a5961a0f"} Oct 05 22:18:05 crc kubenswrapper[4754]: I1005 22:18:05.441388 4754 scope.go:117] "RemoveContainer" containerID="c50ffee964171eed8b7ee0cd69364b26a808f447930df84f79e3a9a2b8b5bab8" Oct 05 22:18:05 crc kubenswrapper[4754]: I1005 22:18:05.442030 4754 scope.go:117] "RemoveContainer" containerID="05cd65b5014dfd796e8d71d762a213d6758e74cce340c7705d8831b6a5961a0f" Oct 05 22:18:05 crc kubenswrapper[4754]: E1005 22:18:05.442286 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 22:18:08 crc kubenswrapper[4754]: I1005 22:18:08.646145 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-hcx7v"] Oct 05 22:18:08 crc kubenswrapper[4754]: E1005 22:18:08.647190 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c284318-3b01-4dd5-8aae-18253bc07696" containerName="extract-content" Oct 05 22:18:08 crc kubenswrapper[4754]: I1005 22:18:08.647204 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c284318-3b01-4dd5-8aae-18253bc07696" containerName="extract-content" Oct 05 22:18:08 crc kubenswrapper[4754]: E1005 22:18:08.647226 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c284318-3b01-4dd5-8aae-18253bc07696" containerName="extract-utilities" Oct 05 22:18:08 crc kubenswrapper[4754]: I1005 22:18:08.647232 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c284318-3b01-4dd5-8aae-18253bc07696" containerName="extract-utilities" Oct 05 22:18:08 crc kubenswrapper[4754]: E1005 22:18:08.647243 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c284318-3b01-4dd5-8aae-18253bc07696" containerName="registry-server" Oct 05 22:18:08 crc kubenswrapper[4754]: I1005 22:18:08.647249 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c284318-3b01-4dd5-8aae-18253bc07696" containerName="registry-server" Oct 05 22:18:08 crc kubenswrapper[4754]: I1005 22:18:08.647441 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c284318-3b01-4dd5-8aae-18253bc07696" containerName="registry-server" Oct 05 22:18:08 crc kubenswrapper[4754]: I1005 22:18:08.649270 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hcx7v" Oct 05 22:18:08 crc kubenswrapper[4754]: I1005 22:18:08.656512 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hcx7v"] Oct 05 22:18:08 crc kubenswrapper[4754]: I1005 22:18:08.809428 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7x57c\" (UniqueName: \"kubernetes.io/projected/4ac95b66-86a3-4087-b752-e487c0a1adf6-kube-api-access-7x57c\") pod \"community-operators-hcx7v\" (UID: \"4ac95b66-86a3-4087-b752-e487c0a1adf6\") " pod="openshift-marketplace/community-operators-hcx7v" Oct 05 22:18:08 crc kubenswrapper[4754]: I1005 22:18:08.809507 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4ac95b66-86a3-4087-b752-e487c0a1adf6-catalog-content\") pod \"community-operators-hcx7v\" (UID: \"4ac95b66-86a3-4087-b752-e487c0a1adf6\") " pod="openshift-marketplace/community-operators-hcx7v" Oct 05 22:18:08 crc kubenswrapper[4754]: I1005 22:18:08.809559 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4ac95b66-86a3-4087-b752-e487c0a1adf6-utilities\") pod \"community-operators-hcx7v\" (UID: \"4ac95b66-86a3-4087-b752-e487c0a1adf6\") " pod="openshift-marketplace/community-operators-hcx7v" Oct 05 22:18:08 crc kubenswrapper[4754]: I1005 22:18:08.910633 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7x57c\" (UniqueName: \"kubernetes.io/projected/4ac95b66-86a3-4087-b752-e487c0a1adf6-kube-api-access-7x57c\") pod \"community-operators-hcx7v\" (UID: \"4ac95b66-86a3-4087-b752-e487c0a1adf6\") " pod="openshift-marketplace/community-operators-hcx7v" Oct 05 22:18:08 crc kubenswrapper[4754]: I1005 22:18:08.910708 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4ac95b66-86a3-4087-b752-e487c0a1adf6-catalog-content\") pod \"community-operators-hcx7v\" (UID: \"4ac95b66-86a3-4087-b752-e487c0a1adf6\") " pod="openshift-marketplace/community-operators-hcx7v" Oct 05 22:18:08 crc kubenswrapper[4754]: I1005 22:18:08.910759 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4ac95b66-86a3-4087-b752-e487c0a1adf6-utilities\") pod \"community-operators-hcx7v\" (UID: \"4ac95b66-86a3-4087-b752-e487c0a1adf6\") " pod="openshift-marketplace/community-operators-hcx7v" Oct 05 22:18:08 crc kubenswrapper[4754]: I1005 22:18:08.911182 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4ac95b66-86a3-4087-b752-e487c0a1adf6-utilities\") pod \"community-operators-hcx7v\" (UID: \"4ac95b66-86a3-4087-b752-e487c0a1adf6\") " pod="openshift-marketplace/community-operators-hcx7v" Oct 05 22:18:08 crc kubenswrapper[4754]: I1005 22:18:08.911197 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4ac95b66-86a3-4087-b752-e487c0a1adf6-catalog-content\") pod \"community-operators-hcx7v\" (UID: \"4ac95b66-86a3-4087-b752-e487c0a1adf6\") " pod="openshift-marketplace/community-operators-hcx7v" Oct 05 22:18:08 crc kubenswrapper[4754]: I1005 22:18:08.930329 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7x57c\" (UniqueName: \"kubernetes.io/projected/4ac95b66-86a3-4087-b752-e487c0a1adf6-kube-api-access-7x57c\") pod \"community-operators-hcx7v\" (UID: \"4ac95b66-86a3-4087-b752-e487c0a1adf6\") " pod="openshift-marketplace/community-operators-hcx7v" Oct 05 22:18:08 crc kubenswrapper[4754]: I1005 22:18:08.976127 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hcx7v" Oct 05 22:18:09 crc kubenswrapper[4754]: I1005 22:18:09.588855 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hcx7v"] Oct 05 22:18:10 crc kubenswrapper[4754]: I1005 22:18:10.480456 4754 generic.go:334] "Generic (PLEG): container finished" podID="4ac95b66-86a3-4087-b752-e487c0a1adf6" containerID="c8035f77b5bbd333b9475ce0ba0b58bb897290257ff09eb925588420331e20db" exitCode=0 Oct 05 22:18:10 crc kubenswrapper[4754]: I1005 22:18:10.480529 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hcx7v" event={"ID":"4ac95b66-86a3-4087-b752-e487c0a1adf6","Type":"ContainerDied","Data":"c8035f77b5bbd333b9475ce0ba0b58bb897290257ff09eb925588420331e20db"} Oct 05 22:18:10 crc kubenswrapper[4754]: I1005 22:18:10.481477 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hcx7v" event={"ID":"4ac95b66-86a3-4087-b752-e487c0a1adf6","Type":"ContainerStarted","Data":"92417658e80c6239be57a7461562f5f0de87869fe1590e5a6fd2b335f36009e8"} Oct 05 22:18:11 crc kubenswrapper[4754]: I1005 22:18:11.494348 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hcx7v" event={"ID":"4ac95b66-86a3-4087-b752-e487c0a1adf6","Type":"ContainerStarted","Data":"0faf60fb217b4993dd25cd50d21628f6f45a9a639ce2dd529e023ac99a2103c4"} Oct 05 22:18:11 crc kubenswrapper[4754]: I1005 22:18:11.639164 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-mgzk9"] Oct 05 22:18:11 crc kubenswrapper[4754]: I1005 22:18:11.686567 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mgzk9"] Oct 05 22:18:11 crc kubenswrapper[4754]: I1005 22:18:11.687007 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mgzk9" Oct 05 22:18:11 crc kubenswrapper[4754]: I1005 22:18:11.778778 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c5404c28-6a2f-4605-a4b8-c380747a1c07-catalog-content\") pod \"certified-operators-mgzk9\" (UID: \"c5404c28-6a2f-4605-a4b8-c380747a1c07\") " pod="openshift-marketplace/certified-operators-mgzk9" Oct 05 22:18:11 crc kubenswrapper[4754]: I1005 22:18:11.778845 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c5404c28-6a2f-4605-a4b8-c380747a1c07-utilities\") pod \"certified-operators-mgzk9\" (UID: \"c5404c28-6a2f-4605-a4b8-c380747a1c07\") " pod="openshift-marketplace/certified-operators-mgzk9" Oct 05 22:18:11 crc kubenswrapper[4754]: I1005 22:18:11.778946 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h95l2\" (UniqueName: \"kubernetes.io/projected/c5404c28-6a2f-4605-a4b8-c380747a1c07-kube-api-access-h95l2\") pod \"certified-operators-mgzk9\" (UID: \"c5404c28-6a2f-4605-a4b8-c380747a1c07\") " pod="openshift-marketplace/certified-operators-mgzk9" Oct 05 22:18:11 crc kubenswrapper[4754]: I1005 22:18:11.880218 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h95l2\" (UniqueName: \"kubernetes.io/projected/c5404c28-6a2f-4605-a4b8-c380747a1c07-kube-api-access-h95l2\") pod \"certified-operators-mgzk9\" (UID: \"c5404c28-6a2f-4605-a4b8-c380747a1c07\") " pod="openshift-marketplace/certified-operators-mgzk9" Oct 05 22:18:11 crc kubenswrapper[4754]: I1005 22:18:11.880314 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c5404c28-6a2f-4605-a4b8-c380747a1c07-catalog-content\") pod \"certified-operators-mgzk9\" (UID: \"c5404c28-6a2f-4605-a4b8-c380747a1c07\") " pod="openshift-marketplace/certified-operators-mgzk9" Oct 05 22:18:11 crc kubenswrapper[4754]: I1005 22:18:11.880351 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c5404c28-6a2f-4605-a4b8-c380747a1c07-utilities\") pod \"certified-operators-mgzk9\" (UID: \"c5404c28-6a2f-4605-a4b8-c380747a1c07\") " pod="openshift-marketplace/certified-operators-mgzk9" Oct 05 22:18:11 crc kubenswrapper[4754]: I1005 22:18:11.880825 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c5404c28-6a2f-4605-a4b8-c380747a1c07-catalog-content\") pod \"certified-operators-mgzk9\" (UID: \"c5404c28-6a2f-4605-a4b8-c380747a1c07\") " pod="openshift-marketplace/certified-operators-mgzk9" Oct 05 22:18:11 crc kubenswrapper[4754]: I1005 22:18:11.880880 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c5404c28-6a2f-4605-a4b8-c380747a1c07-utilities\") pod \"certified-operators-mgzk9\" (UID: \"c5404c28-6a2f-4605-a4b8-c380747a1c07\") " pod="openshift-marketplace/certified-operators-mgzk9" Oct 05 22:18:11 crc kubenswrapper[4754]: I1005 22:18:11.906464 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h95l2\" (UniqueName: \"kubernetes.io/projected/c5404c28-6a2f-4605-a4b8-c380747a1c07-kube-api-access-h95l2\") pod \"certified-operators-mgzk9\" (UID: \"c5404c28-6a2f-4605-a4b8-c380747a1c07\") " pod="openshift-marketplace/certified-operators-mgzk9" Oct 05 22:18:12 crc kubenswrapper[4754]: I1005 22:18:12.039535 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mgzk9" Oct 05 22:18:12 crc kubenswrapper[4754]: I1005 22:18:12.516798 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mgzk9"] Oct 05 22:18:13 crc kubenswrapper[4754]: I1005 22:18:13.522919 4754 generic.go:334] "Generic (PLEG): container finished" podID="c5404c28-6a2f-4605-a4b8-c380747a1c07" containerID="0f1a7b8608758f4229ca8aa4a650a0ef19c1b5c7257c858162ff497b5ae89765" exitCode=0 Oct 05 22:18:13 crc kubenswrapper[4754]: I1005 22:18:13.522973 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mgzk9" event={"ID":"c5404c28-6a2f-4605-a4b8-c380747a1c07","Type":"ContainerDied","Data":"0f1a7b8608758f4229ca8aa4a650a0ef19c1b5c7257c858162ff497b5ae89765"} Oct 05 22:18:13 crc kubenswrapper[4754]: I1005 22:18:13.523404 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mgzk9" event={"ID":"c5404c28-6a2f-4605-a4b8-c380747a1c07","Type":"ContainerStarted","Data":"afdaa743d81b995b3130ad99253c063c4bd7fc5b3ac79f7b4ca80465ad11a6b8"} Oct 05 22:18:13 crc kubenswrapper[4754]: I1005 22:18:13.526883 4754 generic.go:334] "Generic (PLEG): container finished" podID="4ac95b66-86a3-4087-b752-e487c0a1adf6" containerID="0faf60fb217b4993dd25cd50d21628f6f45a9a639ce2dd529e023ac99a2103c4" exitCode=0 Oct 05 22:18:13 crc kubenswrapper[4754]: I1005 22:18:13.526920 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hcx7v" event={"ID":"4ac95b66-86a3-4087-b752-e487c0a1adf6","Type":"ContainerDied","Data":"0faf60fb217b4993dd25cd50d21628f6f45a9a639ce2dd529e023ac99a2103c4"} Oct 05 22:18:14 crc kubenswrapper[4754]: I1005 22:18:14.551002 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hcx7v" event={"ID":"4ac95b66-86a3-4087-b752-e487c0a1adf6","Type":"ContainerStarted","Data":"2c3b833bba4aa41aadaabc2b56d796b743890f398c31cd481a65bdcb50627dff"} Oct 05 22:18:14 crc kubenswrapper[4754]: I1005 22:18:14.560391 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mgzk9" event={"ID":"c5404c28-6a2f-4605-a4b8-c380747a1c07","Type":"ContainerStarted","Data":"dc0f7ec57532b93e585b5731a1184a5a661840486c94695f7a1890689c119e26"} Oct 05 22:18:14 crc kubenswrapper[4754]: I1005 22:18:14.576790 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-hcx7v" podStartSLOduration=2.98297897 podStartE2EDuration="6.576776402s" podCreationTimestamp="2025-10-05 22:18:08 +0000 UTC" firstStartedPulling="2025-10-05 22:18:10.481847073 +0000 UTC m=+5014.385965773" lastFinishedPulling="2025-10-05 22:18:14.075644495 +0000 UTC m=+5017.979763205" observedRunningTime="2025-10-05 22:18:14.57399252 +0000 UTC m=+5018.478111230" watchObservedRunningTime="2025-10-05 22:18:14.576776402 +0000 UTC m=+5018.480895112" Oct 05 22:18:16 crc kubenswrapper[4754]: I1005 22:18:16.581310 4754 generic.go:334] "Generic (PLEG): container finished" podID="c5404c28-6a2f-4605-a4b8-c380747a1c07" containerID="dc0f7ec57532b93e585b5731a1184a5a661840486c94695f7a1890689c119e26" exitCode=0 Oct 05 22:18:16 crc kubenswrapper[4754]: I1005 22:18:16.581381 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mgzk9" event={"ID":"c5404c28-6a2f-4605-a4b8-c380747a1c07","Type":"ContainerDied","Data":"dc0f7ec57532b93e585b5731a1184a5a661840486c94695f7a1890689c119e26"} Oct 05 22:18:17 crc kubenswrapper[4754]: I1005 22:18:17.837370 4754 scope.go:117] "RemoveContainer" containerID="05cd65b5014dfd796e8d71d762a213d6758e74cce340c7705d8831b6a5961a0f" Oct 05 22:18:17 crc kubenswrapper[4754]: E1005 22:18:17.837821 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 22:18:18 crc kubenswrapper[4754]: I1005 22:18:18.606148 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mgzk9" event={"ID":"c5404c28-6a2f-4605-a4b8-c380747a1c07","Type":"ContainerStarted","Data":"a9c0b2b3965ac00a90416e670c235449fa8cdec66b5463d2e1f165a503572906"} Oct 05 22:18:18 crc kubenswrapper[4754]: I1005 22:18:18.630518 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-mgzk9" podStartSLOduration=4.176092357 podStartE2EDuration="7.630480751s" podCreationTimestamp="2025-10-05 22:18:11 +0000 UTC" firstStartedPulling="2025-10-05 22:18:13.525146118 +0000 UTC m=+5017.429264828" lastFinishedPulling="2025-10-05 22:18:16.979534512 +0000 UTC m=+5020.883653222" observedRunningTime="2025-10-05 22:18:18.621765035 +0000 UTC m=+5022.525883755" watchObservedRunningTime="2025-10-05 22:18:18.630480751 +0000 UTC m=+5022.534599481" Oct 05 22:18:18 crc kubenswrapper[4754]: I1005 22:18:18.977349 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-hcx7v" Oct 05 22:18:18 crc kubenswrapper[4754]: I1005 22:18:18.977417 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-hcx7v" Oct 05 22:18:20 crc kubenswrapper[4754]: I1005 22:18:20.021916 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-hcx7v" podUID="4ac95b66-86a3-4087-b752-e487c0a1adf6" containerName="registry-server" probeResult="failure" output=< Oct 05 22:18:20 crc kubenswrapper[4754]: timeout: failed to connect service ":50051" within 1s Oct 05 22:18:20 crc kubenswrapper[4754]: > Oct 05 22:18:22 crc kubenswrapper[4754]: I1005 22:18:22.040476 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-mgzk9" Oct 05 22:18:22 crc kubenswrapper[4754]: I1005 22:18:22.040949 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-mgzk9" Oct 05 22:18:22 crc kubenswrapper[4754]: I1005 22:18:22.091336 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-mgzk9" Oct 05 22:18:22 crc kubenswrapper[4754]: I1005 22:18:22.678173 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-mgzk9" Oct 05 22:18:22 crc kubenswrapper[4754]: I1005 22:18:22.723159 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-mgzk9"] Oct 05 22:18:24 crc kubenswrapper[4754]: I1005 22:18:24.647537 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-mgzk9" podUID="c5404c28-6a2f-4605-a4b8-c380747a1c07" containerName="registry-server" containerID="cri-o://a9c0b2b3965ac00a90416e670c235449fa8cdec66b5463d2e1f165a503572906" gracePeriod=2 Oct 05 22:18:25 crc kubenswrapper[4754]: I1005 22:18:25.487221 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mgzk9" Oct 05 22:18:25 crc kubenswrapper[4754]: I1005 22:18:25.557585 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h95l2\" (UniqueName: \"kubernetes.io/projected/c5404c28-6a2f-4605-a4b8-c380747a1c07-kube-api-access-h95l2\") pod \"c5404c28-6a2f-4605-a4b8-c380747a1c07\" (UID: \"c5404c28-6a2f-4605-a4b8-c380747a1c07\") " Oct 05 22:18:25 crc kubenswrapper[4754]: I1005 22:18:25.557652 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c5404c28-6a2f-4605-a4b8-c380747a1c07-catalog-content\") pod \"c5404c28-6a2f-4605-a4b8-c380747a1c07\" (UID: \"c5404c28-6a2f-4605-a4b8-c380747a1c07\") " Oct 05 22:18:25 crc kubenswrapper[4754]: I1005 22:18:25.557715 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c5404c28-6a2f-4605-a4b8-c380747a1c07-utilities\") pod \"c5404c28-6a2f-4605-a4b8-c380747a1c07\" (UID: \"c5404c28-6a2f-4605-a4b8-c380747a1c07\") " Oct 05 22:18:25 crc kubenswrapper[4754]: I1005 22:18:25.558755 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c5404c28-6a2f-4605-a4b8-c380747a1c07-utilities" (OuterVolumeSpecName: "utilities") pod "c5404c28-6a2f-4605-a4b8-c380747a1c07" (UID: "c5404c28-6a2f-4605-a4b8-c380747a1c07"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 22:18:25 crc kubenswrapper[4754]: I1005 22:18:25.558886 4754 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c5404c28-6a2f-4605-a4b8-c380747a1c07-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 22:18:25 crc kubenswrapper[4754]: I1005 22:18:25.568676 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c5404c28-6a2f-4605-a4b8-c380747a1c07-kube-api-access-h95l2" (OuterVolumeSpecName: "kube-api-access-h95l2") pod "c5404c28-6a2f-4605-a4b8-c380747a1c07" (UID: "c5404c28-6a2f-4605-a4b8-c380747a1c07"). InnerVolumeSpecName "kube-api-access-h95l2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 22:18:25 crc kubenswrapper[4754]: I1005 22:18:25.611990 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c5404c28-6a2f-4605-a4b8-c380747a1c07-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c5404c28-6a2f-4605-a4b8-c380747a1c07" (UID: "c5404c28-6a2f-4605-a4b8-c380747a1c07"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 22:18:25 crc kubenswrapper[4754]: I1005 22:18:25.658157 4754 generic.go:334] "Generic (PLEG): container finished" podID="c5404c28-6a2f-4605-a4b8-c380747a1c07" containerID="a9c0b2b3965ac00a90416e670c235449fa8cdec66b5463d2e1f165a503572906" exitCode=0 Oct 05 22:18:25 crc kubenswrapper[4754]: I1005 22:18:25.659187 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mgzk9" event={"ID":"c5404c28-6a2f-4605-a4b8-c380747a1c07","Type":"ContainerDied","Data":"a9c0b2b3965ac00a90416e670c235449fa8cdec66b5463d2e1f165a503572906"} Oct 05 22:18:25 crc kubenswrapper[4754]: I1005 22:18:25.660014 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mgzk9" event={"ID":"c5404c28-6a2f-4605-a4b8-c380747a1c07","Type":"ContainerDied","Data":"afdaa743d81b995b3130ad99253c063c4bd7fc5b3ac79f7b4ca80465ad11a6b8"} Oct 05 22:18:25 crc kubenswrapper[4754]: I1005 22:18:25.659893 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h95l2\" (UniqueName: \"kubernetes.io/projected/c5404c28-6a2f-4605-a4b8-c380747a1c07-kube-api-access-h95l2\") on node \"crc\" DevicePath \"\"" Oct 05 22:18:25 crc kubenswrapper[4754]: I1005 22:18:25.660178 4754 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c5404c28-6a2f-4605-a4b8-c380747a1c07-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 22:18:25 crc kubenswrapper[4754]: I1005 22:18:25.660088 4754 scope.go:117] "RemoveContainer" containerID="a9c0b2b3965ac00a90416e670c235449fa8cdec66b5463d2e1f165a503572906" Oct 05 22:18:25 crc kubenswrapper[4754]: I1005 22:18:25.659278 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mgzk9" Oct 05 22:18:25 crc kubenswrapper[4754]: I1005 22:18:25.702091 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-mgzk9"] Oct 05 22:18:25 crc kubenswrapper[4754]: I1005 22:18:25.703860 4754 scope.go:117] "RemoveContainer" containerID="dc0f7ec57532b93e585b5731a1184a5a661840486c94695f7a1890689c119e26" Oct 05 22:18:25 crc kubenswrapper[4754]: I1005 22:18:25.708753 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-mgzk9"] Oct 05 22:18:25 crc kubenswrapper[4754]: I1005 22:18:25.754895 4754 scope.go:117] "RemoveContainer" containerID="0f1a7b8608758f4229ca8aa4a650a0ef19c1b5c7257c858162ff497b5ae89765" Oct 05 22:18:25 crc kubenswrapper[4754]: I1005 22:18:25.779017 4754 scope.go:117] "RemoveContainer" containerID="a9c0b2b3965ac00a90416e670c235449fa8cdec66b5463d2e1f165a503572906" Oct 05 22:18:25 crc kubenswrapper[4754]: E1005 22:18:25.781856 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a9c0b2b3965ac00a90416e670c235449fa8cdec66b5463d2e1f165a503572906\": container with ID starting with a9c0b2b3965ac00a90416e670c235449fa8cdec66b5463d2e1f165a503572906 not found: ID does not exist" containerID="a9c0b2b3965ac00a90416e670c235449fa8cdec66b5463d2e1f165a503572906" Oct 05 22:18:25 crc kubenswrapper[4754]: I1005 22:18:25.781984 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a9c0b2b3965ac00a90416e670c235449fa8cdec66b5463d2e1f165a503572906"} err="failed to get container status \"a9c0b2b3965ac00a90416e670c235449fa8cdec66b5463d2e1f165a503572906\": rpc error: code = NotFound desc = could not find container \"a9c0b2b3965ac00a90416e670c235449fa8cdec66b5463d2e1f165a503572906\": container with ID starting with a9c0b2b3965ac00a90416e670c235449fa8cdec66b5463d2e1f165a503572906 not found: ID does not exist" Oct 05 22:18:25 crc kubenswrapper[4754]: I1005 22:18:25.782060 4754 scope.go:117] "RemoveContainer" containerID="dc0f7ec57532b93e585b5731a1184a5a661840486c94695f7a1890689c119e26" Oct 05 22:18:25 crc kubenswrapper[4754]: E1005 22:18:25.785761 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dc0f7ec57532b93e585b5731a1184a5a661840486c94695f7a1890689c119e26\": container with ID starting with dc0f7ec57532b93e585b5731a1184a5a661840486c94695f7a1890689c119e26 not found: ID does not exist" containerID="dc0f7ec57532b93e585b5731a1184a5a661840486c94695f7a1890689c119e26" Oct 05 22:18:25 crc kubenswrapper[4754]: I1005 22:18:25.785810 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc0f7ec57532b93e585b5731a1184a5a661840486c94695f7a1890689c119e26"} err="failed to get container status \"dc0f7ec57532b93e585b5731a1184a5a661840486c94695f7a1890689c119e26\": rpc error: code = NotFound desc = could not find container \"dc0f7ec57532b93e585b5731a1184a5a661840486c94695f7a1890689c119e26\": container with ID starting with dc0f7ec57532b93e585b5731a1184a5a661840486c94695f7a1890689c119e26 not found: ID does not exist" Oct 05 22:18:25 crc kubenswrapper[4754]: I1005 22:18:25.785836 4754 scope.go:117] "RemoveContainer" containerID="0f1a7b8608758f4229ca8aa4a650a0ef19c1b5c7257c858162ff497b5ae89765" Oct 05 22:18:25 crc kubenswrapper[4754]: E1005 22:18:25.786080 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0f1a7b8608758f4229ca8aa4a650a0ef19c1b5c7257c858162ff497b5ae89765\": container with ID starting with 0f1a7b8608758f4229ca8aa4a650a0ef19c1b5c7257c858162ff497b5ae89765 not found: ID does not exist" containerID="0f1a7b8608758f4229ca8aa4a650a0ef19c1b5c7257c858162ff497b5ae89765" Oct 05 22:18:25 crc kubenswrapper[4754]: I1005 22:18:25.786178 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0f1a7b8608758f4229ca8aa4a650a0ef19c1b5c7257c858162ff497b5ae89765"} err="failed to get container status \"0f1a7b8608758f4229ca8aa4a650a0ef19c1b5c7257c858162ff497b5ae89765\": rpc error: code = NotFound desc = could not find container \"0f1a7b8608758f4229ca8aa4a650a0ef19c1b5c7257c858162ff497b5ae89765\": container with ID starting with 0f1a7b8608758f4229ca8aa4a650a0ef19c1b5c7257c858162ff497b5ae89765 not found: ID does not exist" Oct 05 22:18:26 crc kubenswrapper[4754]: I1005 22:18:26.850534 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c5404c28-6a2f-4605-a4b8-c380747a1c07" path="/var/lib/kubelet/pods/c5404c28-6a2f-4605-a4b8-c380747a1c07/volumes" Oct 05 22:18:29 crc kubenswrapper[4754]: I1005 22:18:29.020365 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-hcx7v" Oct 05 22:18:29 crc kubenswrapper[4754]: I1005 22:18:29.106709 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-hcx7v" Oct 05 22:18:29 crc kubenswrapper[4754]: I1005 22:18:29.257806 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-hcx7v"] Oct 05 22:18:30 crc kubenswrapper[4754]: I1005 22:18:30.686655 4754 scope.go:117] "RemoveContainer" containerID="98c8eb1868eeae834122a92fa4c4fcbd2c64c1e430a55d02f2cfa2d4a3447b4d" Oct 05 22:18:30 crc kubenswrapper[4754]: I1005 22:18:30.700264 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-hcx7v" podUID="4ac95b66-86a3-4087-b752-e487c0a1adf6" containerName="registry-server" containerID="cri-o://2c3b833bba4aa41aadaabc2b56d796b743890f398c31cd481a65bdcb50627dff" gracePeriod=2 Oct 05 22:18:31 crc kubenswrapper[4754]: I1005 22:18:31.201375 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hcx7v" Oct 05 22:18:31 crc kubenswrapper[4754]: I1005 22:18:31.254048 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4ac95b66-86a3-4087-b752-e487c0a1adf6-utilities\") pod \"4ac95b66-86a3-4087-b752-e487c0a1adf6\" (UID: \"4ac95b66-86a3-4087-b752-e487c0a1adf6\") " Oct 05 22:18:31 crc kubenswrapper[4754]: I1005 22:18:31.254140 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7x57c\" (UniqueName: \"kubernetes.io/projected/4ac95b66-86a3-4087-b752-e487c0a1adf6-kube-api-access-7x57c\") pod \"4ac95b66-86a3-4087-b752-e487c0a1adf6\" (UID: \"4ac95b66-86a3-4087-b752-e487c0a1adf6\") " Oct 05 22:18:31 crc kubenswrapper[4754]: I1005 22:18:31.254225 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4ac95b66-86a3-4087-b752-e487c0a1adf6-catalog-content\") pod \"4ac95b66-86a3-4087-b752-e487c0a1adf6\" (UID: \"4ac95b66-86a3-4087-b752-e487c0a1adf6\") " Oct 05 22:18:31 crc kubenswrapper[4754]: I1005 22:18:31.254759 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4ac95b66-86a3-4087-b752-e487c0a1adf6-utilities" (OuterVolumeSpecName: "utilities") pod "4ac95b66-86a3-4087-b752-e487c0a1adf6" (UID: "4ac95b66-86a3-4087-b752-e487c0a1adf6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 22:18:31 crc kubenswrapper[4754]: I1005 22:18:31.259057 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4ac95b66-86a3-4087-b752-e487c0a1adf6-kube-api-access-7x57c" (OuterVolumeSpecName: "kube-api-access-7x57c") pod "4ac95b66-86a3-4087-b752-e487c0a1adf6" (UID: "4ac95b66-86a3-4087-b752-e487c0a1adf6"). InnerVolumeSpecName "kube-api-access-7x57c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 22:18:31 crc kubenswrapper[4754]: I1005 22:18:31.323859 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4ac95b66-86a3-4087-b752-e487c0a1adf6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4ac95b66-86a3-4087-b752-e487c0a1adf6" (UID: "4ac95b66-86a3-4087-b752-e487c0a1adf6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 22:18:31 crc kubenswrapper[4754]: I1005 22:18:31.356235 4754 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4ac95b66-86a3-4087-b752-e487c0a1adf6-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 22:18:31 crc kubenswrapper[4754]: I1005 22:18:31.356267 4754 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4ac95b66-86a3-4087-b752-e487c0a1adf6-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 22:18:31 crc kubenswrapper[4754]: I1005 22:18:31.356279 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7x57c\" (UniqueName: \"kubernetes.io/projected/4ac95b66-86a3-4087-b752-e487c0a1adf6-kube-api-access-7x57c\") on node \"crc\" DevicePath \"\"" Oct 05 22:18:31 crc kubenswrapper[4754]: I1005 22:18:31.710050 4754 generic.go:334] "Generic (PLEG): container finished" podID="4ac95b66-86a3-4087-b752-e487c0a1adf6" containerID="2c3b833bba4aa41aadaabc2b56d796b743890f398c31cd481a65bdcb50627dff" exitCode=0 Oct 05 22:18:31 crc kubenswrapper[4754]: I1005 22:18:31.710097 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hcx7v" event={"ID":"4ac95b66-86a3-4087-b752-e487c0a1adf6","Type":"ContainerDied","Data":"2c3b833bba4aa41aadaabc2b56d796b743890f398c31cd481a65bdcb50627dff"} Oct 05 22:18:31 crc kubenswrapper[4754]: I1005 22:18:31.710125 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hcx7v" event={"ID":"4ac95b66-86a3-4087-b752-e487c0a1adf6","Type":"ContainerDied","Data":"92417658e80c6239be57a7461562f5f0de87869fe1590e5a6fd2b335f36009e8"} Oct 05 22:18:31 crc kubenswrapper[4754]: I1005 22:18:31.710142 4754 scope.go:117] "RemoveContainer" containerID="2c3b833bba4aa41aadaabc2b56d796b743890f398c31cd481a65bdcb50627dff" Oct 05 22:18:31 crc kubenswrapper[4754]: I1005 22:18:31.710236 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hcx7v" Oct 05 22:18:31 crc kubenswrapper[4754]: I1005 22:18:31.729632 4754 scope.go:117] "RemoveContainer" containerID="0faf60fb217b4993dd25cd50d21628f6f45a9a639ce2dd529e023ac99a2103c4" Oct 05 22:18:31 crc kubenswrapper[4754]: I1005 22:18:31.758590 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-hcx7v"] Oct 05 22:18:31 crc kubenswrapper[4754]: I1005 22:18:31.770272 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-hcx7v"] Oct 05 22:18:31 crc kubenswrapper[4754]: I1005 22:18:31.779650 4754 scope.go:117] "RemoveContainer" containerID="c8035f77b5bbd333b9475ce0ba0b58bb897290257ff09eb925588420331e20db" Oct 05 22:18:31 crc kubenswrapper[4754]: I1005 22:18:31.802547 4754 scope.go:117] "RemoveContainer" containerID="2c3b833bba4aa41aadaabc2b56d796b743890f398c31cd481a65bdcb50627dff" Oct 05 22:18:31 crc kubenswrapper[4754]: E1005 22:18:31.803796 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2c3b833bba4aa41aadaabc2b56d796b743890f398c31cd481a65bdcb50627dff\": container with ID starting with 2c3b833bba4aa41aadaabc2b56d796b743890f398c31cd481a65bdcb50627dff not found: ID does not exist" containerID="2c3b833bba4aa41aadaabc2b56d796b743890f398c31cd481a65bdcb50627dff" Oct 05 22:18:31 crc kubenswrapper[4754]: I1005 22:18:31.803841 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2c3b833bba4aa41aadaabc2b56d796b743890f398c31cd481a65bdcb50627dff"} err="failed to get container status \"2c3b833bba4aa41aadaabc2b56d796b743890f398c31cd481a65bdcb50627dff\": rpc error: code = NotFound desc = could not find container \"2c3b833bba4aa41aadaabc2b56d796b743890f398c31cd481a65bdcb50627dff\": container with ID starting with 2c3b833bba4aa41aadaabc2b56d796b743890f398c31cd481a65bdcb50627dff not found: ID does not exist" Oct 05 22:18:31 crc kubenswrapper[4754]: I1005 22:18:31.803868 4754 scope.go:117] "RemoveContainer" containerID="0faf60fb217b4993dd25cd50d21628f6f45a9a639ce2dd529e023ac99a2103c4" Oct 05 22:18:31 crc kubenswrapper[4754]: E1005 22:18:31.805287 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0faf60fb217b4993dd25cd50d21628f6f45a9a639ce2dd529e023ac99a2103c4\": container with ID starting with 0faf60fb217b4993dd25cd50d21628f6f45a9a639ce2dd529e023ac99a2103c4 not found: ID does not exist" containerID="0faf60fb217b4993dd25cd50d21628f6f45a9a639ce2dd529e023ac99a2103c4" Oct 05 22:18:31 crc kubenswrapper[4754]: I1005 22:18:31.805315 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0faf60fb217b4993dd25cd50d21628f6f45a9a639ce2dd529e023ac99a2103c4"} err="failed to get container status \"0faf60fb217b4993dd25cd50d21628f6f45a9a639ce2dd529e023ac99a2103c4\": rpc error: code = NotFound desc = could not find container \"0faf60fb217b4993dd25cd50d21628f6f45a9a639ce2dd529e023ac99a2103c4\": container with ID starting with 0faf60fb217b4993dd25cd50d21628f6f45a9a639ce2dd529e023ac99a2103c4 not found: ID does not exist" Oct 05 22:18:31 crc kubenswrapper[4754]: I1005 22:18:31.805331 4754 scope.go:117] "RemoveContainer" containerID="c8035f77b5bbd333b9475ce0ba0b58bb897290257ff09eb925588420331e20db" Oct 05 22:18:31 crc kubenswrapper[4754]: E1005 22:18:31.805604 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c8035f77b5bbd333b9475ce0ba0b58bb897290257ff09eb925588420331e20db\": container with ID starting with c8035f77b5bbd333b9475ce0ba0b58bb897290257ff09eb925588420331e20db not found: ID does not exist" containerID="c8035f77b5bbd333b9475ce0ba0b58bb897290257ff09eb925588420331e20db" Oct 05 22:18:31 crc kubenswrapper[4754]: I1005 22:18:31.805733 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c8035f77b5bbd333b9475ce0ba0b58bb897290257ff09eb925588420331e20db"} err="failed to get container status \"c8035f77b5bbd333b9475ce0ba0b58bb897290257ff09eb925588420331e20db\": rpc error: code = NotFound desc = could not find container \"c8035f77b5bbd333b9475ce0ba0b58bb897290257ff09eb925588420331e20db\": container with ID starting with c8035f77b5bbd333b9475ce0ba0b58bb897290257ff09eb925588420331e20db not found: ID does not exist" Oct 05 22:18:32 crc kubenswrapper[4754]: I1005 22:18:32.837704 4754 scope.go:117] "RemoveContainer" containerID="05cd65b5014dfd796e8d71d762a213d6758e74cce340c7705d8831b6a5961a0f" Oct 05 22:18:32 crc kubenswrapper[4754]: E1005 22:18:32.838013 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 22:18:32 crc kubenswrapper[4754]: I1005 22:18:32.846941 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4ac95b66-86a3-4087-b752-e487c0a1adf6" path="/var/lib/kubelet/pods/4ac95b66-86a3-4087-b752-e487c0a1adf6/volumes" Oct 05 22:18:44 crc kubenswrapper[4754]: I1005 22:18:44.838385 4754 scope.go:117] "RemoveContainer" containerID="05cd65b5014dfd796e8d71d762a213d6758e74cce340c7705d8831b6a5961a0f" Oct 05 22:18:44 crc kubenswrapper[4754]: E1005 22:18:44.839030 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 22:18:50 crc kubenswrapper[4754]: I1005 22:18:50.695034 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-765c6857c8-q7glx_bd23550d-8815-4211-b8be-10ee8ab95c1f/barbican-api-log/0.log" Oct 05 22:18:50 crc kubenswrapper[4754]: I1005 22:18:50.752429 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-765c6857c8-q7glx_bd23550d-8815-4211-b8be-10ee8ab95c1f/barbican-api/0.log" Oct 05 22:18:51 crc kubenswrapper[4754]: I1005 22:18:51.565600 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-9dc4bfc9d-cfm98_358e33ff-b4ce-46e3-b077-e1df373f2a75/barbican-keystone-listener/0.log" Oct 05 22:18:51 crc kubenswrapper[4754]: I1005 22:18:51.630169 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-9dc4bfc9d-cfm98_358e33ff-b4ce-46e3-b077-e1df373f2a75/barbican-keystone-listener-log/0.log" Oct 05 22:18:51 crc kubenswrapper[4754]: I1005 22:18:51.795704 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-54b7554787-vqrt4_346d8b39-98ea-442d-910e-b40a6e4b85c4/barbican-worker/0.log" Oct 05 22:18:51 crc kubenswrapper[4754]: I1005 22:18:51.864901 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-54b7554787-vqrt4_346d8b39-98ea-442d-910e-b40a6e4b85c4/barbican-worker-log/0.log" Oct 05 22:18:52 crc kubenswrapper[4754]: I1005 22:18:52.011931 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-6cdph_31cbd0a7-7e3d-4bd6-b84c-816431fdc633/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Oct 05 22:18:52 crc kubenswrapper[4754]: I1005 22:18:52.217136 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_b5010900-7b57-48e9-9621-73940c2aefa1/ceilometer-notification-agent/0.log" Oct 05 22:18:52 crc kubenswrapper[4754]: I1005 22:18:52.227792 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_b5010900-7b57-48e9-9621-73940c2aefa1/ceilometer-central-agent/0.log" Oct 05 22:18:52 crc kubenswrapper[4754]: I1005 22:18:52.280827 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_b5010900-7b57-48e9-9621-73940c2aefa1/proxy-httpd/0.log" Oct 05 22:18:52 crc kubenswrapper[4754]: I1005 22:18:52.398872 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_b5010900-7b57-48e9-9621-73940c2aefa1/sg-core/0.log" Oct 05 22:18:52 crc kubenswrapper[4754]: I1005 22:18:52.544784 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_90277df3-c655-423b-b6a6-8f3142ab046e/cinder-api/0.log" Oct 05 22:18:52 crc kubenswrapper[4754]: I1005 22:18:52.983771 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_90277df3-c655-423b-b6a6-8f3142ab046e/cinder-api-log/0.log" Oct 05 22:18:53 crc kubenswrapper[4754]: I1005 22:18:53.077196 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_8da021ab-a4b1-4a42-bf58-2973a0578d8e/cinder-scheduler/0.log" Oct 05 22:18:53 crc kubenswrapper[4754]: I1005 22:18:53.237189 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_8da021ab-a4b1-4a42-bf58-2973a0578d8e/probe/0.log" Oct 05 22:18:53 crc kubenswrapper[4754]: I1005 22:18:53.289575 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-7xfdc_5421db8c-f81e-483e-ac15-c0b725a2c277/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Oct 05 22:18:53 crc kubenswrapper[4754]: I1005 22:18:53.527221 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-k54sd_77f69424-dcfc-4915-b704-3447e3cf31eb/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 05 22:18:53 crc kubenswrapper[4754]: I1005 22:18:53.782000 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-p2gbk_50e8c589-a949-4193-94a1-022fea47e2ba/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 05 22:18:53 crc kubenswrapper[4754]: I1005 22:18:53.789309 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-79dc84bdb7-ngcb6_f031b1f6-304e-4419-a4b2-97a8415a530a/init/0.log" Oct 05 22:18:53 crc kubenswrapper[4754]: I1005 22:18:53.948770 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-79dc84bdb7-ngcb6_f031b1f6-304e-4419-a4b2-97a8415a530a/init/0.log" Oct 05 22:18:54 crc kubenswrapper[4754]: I1005 22:18:54.015960 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-5x9f9_920f98be-147d-4b28-a82c-fa77c86fd8f1/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Oct 05 22:18:54 crc kubenswrapper[4754]: I1005 22:18:54.123738 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-79dc84bdb7-ngcb6_f031b1f6-304e-4419-a4b2-97a8415a530a/dnsmasq-dns/0.log" Oct 05 22:18:54 crc kubenswrapper[4754]: I1005 22:18:54.261195 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_49d68627-98f5-4eec-b544-1de274b6e786/glance-httpd/0.log" Oct 05 22:18:54 crc kubenswrapper[4754]: I1005 22:18:54.314283 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_49d68627-98f5-4eec-b544-1de274b6e786/glance-log/0.log" Oct 05 22:18:54 crc kubenswrapper[4754]: I1005 22:18:54.431074 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_b378ce06-065f-4d9e-80d2-d98d834a2556/glance-httpd/0.log" Oct 05 22:18:54 crc kubenswrapper[4754]: I1005 22:18:54.483276 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_b378ce06-065f-4d9e-80d2-d98d834a2556/glance-log/0.log" Oct 05 22:18:54 crc kubenswrapper[4754]: I1005 22:18:54.627763 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-574454d6cb-vbnk6_39d1aac3-14a1-45bf-9a9d-ba4cdb0afcb1/horizon/3.log" Oct 05 22:18:54 crc kubenswrapper[4754]: I1005 22:18:54.717839 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-574454d6cb-vbnk6_39d1aac3-14a1-45bf-9a9d-ba4cdb0afcb1/horizon/2.log" Oct 05 22:18:54 crc kubenswrapper[4754]: I1005 22:18:54.944937 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-ch8jw_27e32907-804b-43ba-ad6b-741e84eee1ab/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Oct 05 22:18:55 crc kubenswrapper[4754]: I1005 22:18:55.150462 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-csfdg_35d09fe3-0746-4e6c-82b2-cc5bef030483/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 05 22:18:55 crc kubenswrapper[4754]: I1005 22:18:55.225166 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-574454d6cb-vbnk6_39d1aac3-14a1-45bf-9a9d-ba4cdb0afcb1/horizon-log/0.log" Oct 05 22:18:55 crc kubenswrapper[4754]: I1005 22:18:55.512421 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29328361-wb6sz_fc91c348-ca21-48c0-a694-f9850277e59b/keystone-cron/0.log" Oct 05 22:18:55 crc kubenswrapper[4754]: I1005 22:18:55.658303 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_b54ee63c-a198-4855-a04c-fc753939dbc0/kube-state-metrics/0.log" Oct 05 22:18:55 crc kubenswrapper[4754]: I1005 22:18:55.692247 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-55bcb5dcdd-jhgrw_d4c2091b-e2d2-4861-8190-79d548039f18/keystone-api/0.log" Oct 05 22:18:56 crc kubenswrapper[4754]: I1005 22:18:56.137193 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-c7xwk_43f1b88f-48f4-4ee0-9fcd-8f53bee2f461/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Oct 05 22:18:56 crc kubenswrapper[4754]: I1005 22:18:56.834814 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-5d6f8484b5-vstqw_cb41212f-892e-4fe5-ae36-7ba898943277/neutron-httpd/0.log" Oct 05 22:18:56 crc kubenswrapper[4754]: I1005 22:18:56.941359 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-6rhpl_61d4efe0-4a19-4b36-9509-67354a6fd537/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Oct 05 22:18:57 crc kubenswrapper[4754]: I1005 22:18:57.007962 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-5d6f8484b5-vstqw_cb41212f-892e-4fe5-ae36-7ba898943277/neutron-api/0.log" Oct 05 22:18:58 crc kubenswrapper[4754]: I1005 22:18:58.148357 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_b8e2a011-2922-4e21-8e47-b6fcebbc21e9/nova-cell0-conductor-conductor/0.log" Oct 05 22:18:58 crc kubenswrapper[4754]: I1005 22:18:58.461169 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_527adc21-0816-4fba-b7a1-22b3db2e1d73/nova-api-log/0.log" Oct 05 22:18:58 crc kubenswrapper[4754]: I1005 22:18:58.880676 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_9cecb9f9-02d8-42fc-b9ff-b38d3a5aea1b/nova-cell1-conductor-conductor/0.log" Oct 05 22:18:58 crc kubenswrapper[4754]: I1005 22:18:58.941573 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_527adc21-0816-4fba-b7a1-22b3db2e1d73/nova-api-api/0.log" Oct 05 22:18:59 crc kubenswrapper[4754]: I1005 22:18:59.127999 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_eaf71ba6-38b3-4dd1-98e7-e3e5e5d9e428/nova-cell1-novncproxy-novncproxy/0.log" Oct 05 22:18:59 crc kubenswrapper[4754]: I1005 22:18:59.308679 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-fx4cg_542410e6-102a-4e41-b3a5-528f37501180/nova-edpm-deployment-openstack-edpm-ipam/0.log" Oct 05 22:18:59 crc kubenswrapper[4754]: I1005 22:18:59.682285 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_bfc1abc2-68ba-4fc4-9dac-5e15ce2f52cd/nova-metadata-log/0.log" Oct 05 22:18:59 crc kubenswrapper[4754]: I1005 22:18:59.841478 4754 scope.go:117] "RemoveContainer" containerID="05cd65b5014dfd796e8d71d762a213d6758e74cce340c7705d8831b6a5961a0f" Oct 05 22:18:59 crc kubenswrapper[4754]: E1005 22:18:59.841883 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 22:19:00 crc kubenswrapper[4754]: I1005 22:19:00.504855 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_d97f6b34-ef35-4b21-81eb-9e7009945fa6/mysql-bootstrap/0.log" Oct 05 22:19:00 crc kubenswrapper[4754]: I1005 22:19:00.531522 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_3e33c8f1-8161-4919-98e0-971f46b33b05/nova-scheduler-scheduler/0.log" Oct 05 22:19:01 crc kubenswrapper[4754]: I1005 22:19:01.073726 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_d97f6b34-ef35-4b21-81eb-9e7009945fa6/mysql-bootstrap/0.log" Oct 05 22:19:01 crc kubenswrapper[4754]: I1005 22:19:01.164051 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_d97f6b34-ef35-4b21-81eb-9e7009945fa6/galera/0.log" Oct 05 22:19:01 crc kubenswrapper[4754]: I1005 22:19:01.447033 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_d311946e-91c6-4be6-9d48-1b823ee77607/mysql-bootstrap/0.log" Oct 05 22:19:01 crc kubenswrapper[4754]: I1005 22:19:01.552300 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_d311946e-91c6-4be6-9d48-1b823ee77607/mysql-bootstrap/0.log" Oct 05 22:19:01 crc kubenswrapper[4754]: I1005 22:19:01.688817 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_d311946e-91c6-4be6-9d48-1b823ee77607/galera/0.log" Oct 05 22:19:01 crc kubenswrapper[4754]: I1005 22:19:01.711577 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_bfc1abc2-68ba-4fc4-9dac-5e15ce2f52cd/nova-metadata-metadata/0.log" Oct 05 22:19:01 crc kubenswrapper[4754]: I1005 22:19:01.908170 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_95ef109d-c48b-40ea-b491-3e730d4d651b/openstackclient/0.log" Oct 05 22:19:02 crc kubenswrapper[4754]: I1005 22:19:02.062332 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-8xfw8_06242eb6-ddde-49ea-b4a1-c61aad6f6402/openstack-network-exporter/0.log" Oct 05 22:19:02 crc kubenswrapper[4754]: I1005 22:19:02.303943 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-2krf2_1283a11a-7581-4146-87db-b62d6d5d6dc9/ovsdb-server-init/0.log" Oct 05 22:19:02 crc kubenswrapper[4754]: I1005 22:19:02.575339 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-2krf2_1283a11a-7581-4146-87db-b62d6d5d6dc9/ovs-vswitchd/0.log" Oct 05 22:19:02 crc kubenswrapper[4754]: I1005 22:19:02.637906 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-2krf2_1283a11a-7581-4146-87db-b62d6d5d6dc9/ovsdb-server-init/0.log" Oct 05 22:19:02 crc kubenswrapper[4754]: I1005 22:19:02.758075 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-2krf2_1283a11a-7581-4146-87db-b62d6d5d6dc9/ovsdb-server/0.log" Oct 05 22:19:02 crc kubenswrapper[4754]: I1005 22:19:02.936785 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-pp728_e3785031-57bd-44ab-b30e-b609cca6dcaf/ovn-controller/0.log" Oct 05 22:19:03 crc kubenswrapper[4754]: I1005 22:19:03.290237 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_316121e1-b760-4eaf-9b8e-ec5bacc6b117/openstack-network-exporter/0.log" Oct 05 22:19:03 crc kubenswrapper[4754]: I1005 22:19:03.333937 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-cwth8_c26e6366-e15d-4c7b-a370-a601e2cee56f/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Oct 05 22:19:03 crc kubenswrapper[4754]: I1005 22:19:03.584350 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_a5fdaf77-8a4c-4267-a32e-22a1c6be8f97/openstack-network-exporter/0.log" Oct 05 22:19:03 crc kubenswrapper[4754]: I1005 22:19:03.621160 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_316121e1-b760-4eaf-9b8e-ec5bacc6b117/ovn-northd/0.log" Oct 05 22:19:03 crc kubenswrapper[4754]: I1005 22:19:03.896434 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_a5fdaf77-8a4c-4267-a32e-22a1c6be8f97/ovsdbserver-nb/0.log" Oct 05 22:19:03 crc kubenswrapper[4754]: I1005 22:19:03.902050 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_3c38a205-1079-46c2-9db5-c380d19399e3/openstack-network-exporter/0.log" Oct 05 22:19:04 crc kubenswrapper[4754]: I1005 22:19:04.163570 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_3c38a205-1079-46c2-9db5-c380d19399e3/ovsdbserver-sb/0.log" Oct 05 22:19:04 crc kubenswrapper[4754]: I1005 22:19:04.444532 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-66d84555d4-xzsp4_ac596dcc-74ee-4144-bd90-3e0967d0478d/placement-api/0.log" Oct 05 22:19:04 crc kubenswrapper[4754]: I1005 22:19:04.583668 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-66d84555d4-xzsp4_ac596dcc-74ee-4144-bd90-3e0967d0478d/placement-log/0.log" Oct 05 22:19:04 crc kubenswrapper[4754]: I1005 22:19:04.645674 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_7eb0b89b-e87d-4d5e-9939-d3a0f97f94db/setup-container/0.log" Oct 05 22:19:04 crc kubenswrapper[4754]: I1005 22:19:04.864566 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_7eb0b89b-e87d-4d5e-9939-d3a0f97f94db/setup-container/0.log" Oct 05 22:19:04 crc kubenswrapper[4754]: I1005 22:19:04.909863 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_7eb0b89b-e87d-4d5e-9939-d3a0f97f94db/rabbitmq/0.log" Oct 05 22:19:05 crc kubenswrapper[4754]: I1005 22:19:05.172935 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_3a6824aa-cdc6-4598-b296-1db23f99b156/setup-container/0.log" Oct 05 22:19:05 crc kubenswrapper[4754]: I1005 22:19:05.415239 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_3a6824aa-cdc6-4598-b296-1db23f99b156/rabbitmq/0.log" Oct 05 22:19:05 crc kubenswrapper[4754]: I1005 22:19:05.447330 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_3a6824aa-cdc6-4598-b296-1db23f99b156/setup-container/0.log" Oct 05 22:19:05 crc kubenswrapper[4754]: I1005 22:19:05.721275 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-4srnb_66ef090a-4d9c-41f2-b225-a3bce5bb58c1/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 05 22:19:05 crc kubenswrapper[4754]: I1005 22:19:05.866512 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-xbw7s_fb37244c-6380-4683-a2ed-a96dd32525c6/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Oct 05 22:19:06 crc kubenswrapper[4754]: I1005 22:19:06.045722 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-bwlth_be10d33f-8bb1-4157-9a5f-1f5bbee194f0/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Oct 05 22:19:06 crc kubenswrapper[4754]: I1005 22:19:06.443672 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-mx4cz_50b5cacd-e284-4f29-acb5-1908246d11d6/ssh-known-hosts-edpm-deployment/0.log" Oct 05 22:19:06 crc kubenswrapper[4754]: I1005 22:19:06.679938 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-m5p76_b30dea9b-eb5b-46c2-88ed-e5b6df21356a/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 05 22:19:06 crc kubenswrapper[4754]: I1005 22:19:06.872086 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-646d45cbfc-6sjtl_e4c2921a-fff3-4c97-98df-206a94054d24/proxy-server/0.log" Oct 05 22:19:06 crc kubenswrapper[4754]: I1005 22:19:06.937180 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-646d45cbfc-6sjtl_e4c2921a-fff3-4c97-98df-206a94054d24/proxy-httpd/0.log" Oct 05 22:19:07 crc kubenswrapper[4754]: I1005 22:19:07.418033 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-ljsff_89a55f59-6693-4e0d-b8f8-dfb258d7a078/swift-ring-rebalance/0.log" Oct 05 22:19:07 crc kubenswrapper[4754]: I1005 22:19:07.431722 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9af24378-0013-4510-aee6-d2fe2172935c/account-auditor/0.log" Oct 05 22:19:07 crc kubenswrapper[4754]: I1005 22:19:07.637241 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9af24378-0013-4510-aee6-d2fe2172935c/account-reaper/0.log" Oct 05 22:19:07 crc kubenswrapper[4754]: I1005 22:19:07.714297 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9af24378-0013-4510-aee6-d2fe2172935c/account-replicator/0.log" Oct 05 22:19:07 crc kubenswrapper[4754]: I1005 22:19:07.847798 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9af24378-0013-4510-aee6-d2fe2172935c/account-server/0.log" Oct 05 22:19:07 crc kubenswrapper[4754]: I1005 22:19:07.892442 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9af24378-0013-4510-aee6-d2fe2172935c/container-auditor/0.log" Oct 05 22:19:08 crc kubenswrapper[4754]: I1005 22:19:08.041482 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9af24378-0013-4510-aee6-d2fe2172935c/container-replicator/0.log" Oct 05 22:19:08 crc kubenswrapper[4754]: I1005 22:19:08.092169 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9af24378-0013-4510-aee6-d2fe2172935c/container-server/0.log" Oct 05 22:19:08 crc kubenswrapper[4754]: I1005 22:19:08.223260 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9af24378-0013-4510-aee6-d2fe2172935c/container-updater/0.log" Oct 05 22:19:08 crc kubenswrapper[4754]: I1005 22:19:08.419480 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9af24378-0013-4510-aee6-d2fe2172935c/object-auditor/0.log" Oct 05 22:19:08 crc kubenswrapper[4754]: I1005 22:19:08.438956 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9af24378-0013-4510-aee6-d2fe2172935c/object-expirer/0.log" Oct 05 22:19:08 crc kubenswrapper[4754]: I1005 22:19:08.485944 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9af24378-0013-4510-aee6-d2fe2172935c/object-replicator/0.log" Oct 05 22:19:09 crc kubenswrapper[4754]: I1005 22:19:09.022560 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9af24378-0013-4510-aee6-d2fe2172935c/rsync/0.log" Oct 05 22:19:09 crc kubenswrapper[4754]: I1005 22:19:09.026901 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9af24378-0013-4510-aee6-d2fe2172935c/object-server/0.log" Oct 05 22:19:09 crc kubenswrapper[4754]: I1005 22:19:09.093000 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9af24378-0013-4510-aee6-d2fe2172935c/object-updater/0.log" Oct 05 22:19:09 crc kubenswrapper[4754]: I1005 22:19:09.328070 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9af24378-0013-4510-aee6-d2fe2172935c/swift-recon-cron/0.log" Oct 05 22:19:09 crc kubenswrapper[4754]: I1005 22:19:09.392219 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-s2xqb_7c094190-76ba-4310-ae5d-1e82d3caeac2/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Oct 05 22:19:09 crc kubenswrapper[4754]: I1005 22:19:09.618135 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_5922ff4d-2ec1-4156-acc7-8e6427819b3e/tempest-tests-tempest-tests-runner/0.log" Oct 05 22:19:09 crc kubenswrapper[4754]: I1005 22:19:09.796850 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_ea8cb295-85b1-4afe-8afe-6d5f19b25f28/test-operator-logs-container/0.log" Oct 05 22:19:10 crc kubenswrapper[4754]: I1005 22:19:10.144130 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-thlqz_773b7e70-b0ed-4b02-98fe-eeaf81925900/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Oct 05 22:19:10 crc kubenswrapper[4754]: I1005 22:19:10.841704 4754 scope.go:117] "RemoveContainer" containerID="05cd65b5014dfd796e8d71d762a213d6758e74cce340c7705d8831b6a5961a0f" Oct 05 22:19:10 crc kubenswrapper[4754]: E1005 22:19:10.848335 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 22:19:20 crc kubenswrapper[4754]: I1005 22:19:20.959925 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_bdbca489-8f9b-420e-a306-65575175af99/memcached/0.log" Oct 05 22:19:24 crc kubenswrapper[4754]: I1005 22:19:24.838228 4754 scope.go:117] "RemoveContainer" containerID="05cd65b5014dfd796e8d71d762a213d6758e74cce340c7705d8831b6a5961a0f" Oct 05 22:19:24 crc kubenswrapper[4754]: E1005 22:19:24.838873 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 22:19:27 crc kubenswrapper[4754]: I1005 22:19:27.217462 4754 generic.go:334] "Generic (PLEG): container finished" podID="da7c6e68-184a-4ef7-83d6-7fb4a2f52638" containerID="4f27cae5dddc7591d87022a2de708593963ba14a366abd1e44e5b438ab79efb4" exitCode=0 Oct 05 22:19:27 crc kubenswrapper[4754]: I1005 22:19:27.217544 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-tv57x/crc-debug-sxp85" event={"ID":"da7c6e68-184a-4ef7-83d6-7fb4a2f52638","Type":"ContainerDied","Data":"4f27cae5dddc7591d87022a2de708593963ba14a366abd1e44e5b438ab79efb4"} Oct 05 22:19:28 crc kubenswrapper[4754]: I1005 22:19:28.327729 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-tv57x/crc-debug-sxp85" Oct 05 22:19:28 crc kubenswrapper[4754]: I1005 22:19:28.356956 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-tv57x/crc-debug-sxp85"] Oct 05 22:19:28 crc kubenswrapper[4754]: I1005 22:19:28.368043 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-tv57x/crc-debug-sxp85"] Oct 05 22:19:28 crc kubenswrapper[4754]: I1005 22:19:28.481787 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8rjcn\" (UniqueName: \"kubernetes.io/projected/da7c6e68-184a-4ef7-83d6-7fb4a2f52638-kube-api-access-8rjcn\") pod \"da7c6e68-184a-4ef7-83d6-7fb4a2f52638\" (UID: \"da7c6e68-184a-4ef7-83d6-7fb4a2f52638\") " Oct 05 22:19:28 crc kubenswrapper[4754]: I1005 22:19:28.481926 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/da7c6e68-184a-4ef7-83d6-7fb4a2f52638-host\") pod \"da7c6e68-184a-4ef7-83d6-7fb4a2f52638\" (UID: \"da7c6e68-184a-4ef7-83d6-7fb4a2f52638\") " Oct 05 22:19:28 crc kubenswrapper[4754]: I1005 22:19:28.482052 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/da7c6e68-184a-4ef7-83d6-7fb4a2f52638-host" (OuterVolumeSpecName: "host") pod "da7c6e68-184a-4ef7-83d6-7fb4a2f52638" (UID: "da7c6e68-184a-4ef7-83d6-7fb4a2f52638"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 05 22:19:28 crc kubenswrapper[4754]: I1005 22:19:28.482668 4754 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/da7c6e68-184a-4ef7-83d6-7fb4a2f52638-host\") on node \"crc\" DevicePath \"\"" Oct 05 22:19:28 crc kubenswrapper[4754]: I1005 22:19:28.497211 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/da7c6e68-184a-4ef7-83d6-7fb4a2f52638-kube-api-access-8rjcn" (OuterVolumeSpecName: "kube-api-access-8rjcn") pod "da7c6e68-184a-4ef7-83d6-7fb4a2f52638" (UID: "da7c6e68-184a-4ef7-83d6-7fb4a2f52638"). InnerVolumeSpecName "kube-api-access-8rjcn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 22:19:28 crc kubenswrapper[4754]: I1005 22:19:28.584248 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8rjcn\" (UniqueName: \"kubernetes.io/projected/da7c6e68-184a-4ef7-83d6-7fb4a2f52638-kube-api-access-8rjcn\") on node \"crc\" DevicePath \"\"" Oct 05 22:19:28 crc kubenswrapper[4754]: I1005 22:19:28.846275 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="da7c6e68-184a-4ef7-83d6-7fb4a2f52638" path="/var/lib/kubelet/pods/da7c6e68-184a-4ef7-83d6-7fb4a2f52638/volumes" Oct 05 22:19:29 crc kubenswrapper[4754]: I1005 22:19:29.234701 4754 scope.go:117] "RemoveContainer" containerID="4f27cae5dddc7591d87022a2de708593963ba14a366abd1e44e5b438ab79efb4" Oct 05 22:19:29 crc kubenswrapper[4754]: I1005 22:19:29.234808 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-tv57x/crc-debug-sxp85" Oct 05 22:19:29 crc kubenswrapper[4754]: I1005 22:19:29.561521 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-tv57x/crc-debug-k5sm5"] Oct 05 22:19:29 crc kubenswrapper[4754]: E1005 22:19:29.562197 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5404c28-6a2f-4605-a4b8-c380747a1c07" containerName="extract-utilities" Oct 05 22:19:29 crc kubenswrapper[4754]: I1005 22:19:29.562211 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5404c28-6a2f-4605-a4b8-c380747a1c07" containerName="extract-utilities" Oct 05 22:19:29 crc kubenswrapper[4754]: E1005 22:19:29.562229 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ac95b66-86a3-4087-b752-e487c0a1adf6" containerName="extract-content" Oct 05 22:19:29 crc kubenswrapper[4754]: I1005 22:19:29.562235 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ac95b66-86a3-4087-b752-e487c0a1adf6" containerName="extract-content" Oct 05 22:19:29 crc kubenswrapper[4754]: E1005 22:19:29.562246 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5404c28-6a2f-4605-a4b8-c380747a1c07" containerName="extract-content" Oct 05 22:19:29 crc kubenswrapper[4754]: I1005 22:19:29.562253 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5404c28-6a2f-4605-a4b8-c380747a1c07" containerName="extract-content" Oct 05 22:19:29 crc kubenswrapper[4754]: E1005 22:19:29.562267 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da7c6e68-184a-4ef7-83d6-7fb4a2f52638" containerName="container-00" Oct 05 22:19:29 crc kubenswrapper[4754]: I1005 22:19:29.562274 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="da7c6e68-184a-4ef7-83d6-7fb4a2f52638" containerName="container-00" Oct 05 22:19:29 crc kubenswrapper[4754]: E1005 22:19:29.562286 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ac95b66-86a3-4087-b752-e487c0a1adf6" containerName="extract-utilities" Oct 05 22:19:29 crc kubenswrapper[4754]: I1005 22:19:29.562291 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ac95b66-86a3-4087-b752-e487c0a1adf6" containerName="extract-utilities" Oct 05 22:19:29 crc kubenswrapper[4754]: E1005 22:19:29.562312 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5404c28-6a2f-4605-a4b8-c380747a1c07" containerName="registry-server" Oct 05 22:19:29 crc kubenswrapper[4754]: I1005 22:19:29.562317 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5404c28-6a2f-4605-a4b8-c380747a1c07" containerName="registry-server" Oct 05 22:19:29 crc kubenswrapper[4754]: E1005 22:19:29.562328 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ac95b66-86a3-4087-b752-e487c0a1adf6" containerName="registry-server" Oct 05 22:19:29 crc kubenswrapper[4754]: I1005 22:19:29.562334 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ac95b66-86a3-4087-b752-e487c0a1adf6" containerName="registry-server" Oct 05 22:19:29 crc kubenswrapper[4754]: I1005 22:19:29.562550 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="c5404c28-6a2f-4605-a4b8-c380747a1c07" containerName="registry-server" Oct 05 22:19:29 crc kubenswrapper[4754]: I1005 22:19:29.562559 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ac95b66-86a3-4087-b752-e487c0a1adf6" containerName="registry-server" Oct 05 22:19:29 crc kubenswrapper[4754]: I1005 22:19:29.562566 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="da7c6e68-184a-4ef7-83d6-7fb4a2f52638" containerName="container-00" Oct 05 22:19:29 crc kubenswrapper[4754]: I1005 22:19:29.563152 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-tv57x/crc-debug-k5sm5" Oct 05 22:19:29 crc kubenswrapper[4754]: I1005 22:19:29.703951 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n7kch\" (UniqueName: \"kubernetes.io/projected/1f0d1bde-e211-492b-96fd-683293ea6d59-kube-api-access-n7kch\") pod \"crc-debug-k5sm5\" (UID: \"1f0d1bde-e211-492b-96fd-683293ea6d59\") " pod="openshift-must-gather-tv57x/crc-debug-k5sm5" Oct 05 22:19:29 crc kubenswrapper[4754]: I1005 22:19:29.704072 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1f0d1bde-e211-492b-96fd-683293ea6d59-host\") pod \"crc-debug-k5sm5\" (UID: \"1f0d1bde-e211-492b-96fd-683293ea6d59\") " pod="openshift-must-gather-tv57x/crc-debug-k5sm5" Oct 05 22:19:29 crc kubenswrapper[4754]: I1005 22:19:29.805710 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n7kch\" (UniqueName: \"kubernetes.io/projected/1f0d1bde-e211-492b-96fd-683293ea6d59-kube-api-access-n7kch\") pod \"crc-debug-k5sm5\" (UID: \"1f0d1bde-e211-492b-96fd-683293ea6d59\") " pod="openshift-must-gather-tv57x/crc-debug-k5sm5" Oct 05 22:19:29 crc kubenswrapper[4754]: I1005 22:19:29.805845 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1f0d1bde-e211-492b-96fd-683293ea6d59-host\") pod \"crc-debug-k5sm5\" (UID: \"1f0d1bde-e211-492b-96fd-683293ea6d59\") " pod="openshift-must-gather-tv57x/crc-debug-k5sm5" Oct 05 22:19:29 crc kubenswrapper[4754]: I1005 22:19:29.805927 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1f0d1bde-e211-492b-96fd-683293ea6d59-host\") pod \"crc-debug-k5sm5\" (UID: \"1f0d1bde-e211-492b-96fd-683293ea6d59\") " pod="openshift-must-gather-tv57x/crc-debug-k5sm5" Oct 05 22:19:30 crc kubenswrapper[4754]: I1005 22:19:30.598934 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n7kch\" (UniqueName: \"kubernetes.io/projected/1f0d1bde-e211-492b-96fd-683293ea6d59-kube-api-access-n7kch\") pod \"crc-debug-k5sm5\" (UID: \"1f0d1bde-e211-492b-96fd-683293ea6d59\") " pod="openshift-must-gather-tv57x/crc-debug-k5sm5" Oct 05 22:19:30 crc kubenswrapper[4754]: I1005 22:19:30.778116 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-tv57x/crc-debug-k5sm5" Oct 05 22:19:30 crc kubenswrapper[4754]: W1005 22:19:30.847863 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1f0d1bde_e211_492b_96fd_683293ea6d59.slice/crio-9a8688554f917d53356beb53d88ec4c4e24b76c3d0d47b728fc1a32276f11509 WatchSource:0}: Error finding container 9a8688554f917d53356beb53d88ec4c4e24b76c3d0d47b728fc1a32276f11509: Status 404 returned error can't find the container with id 9a8688554f917d53356beb53d88ec4c4e24b76c3d0d47b728fc1a32276f11509 Oct 05 22:19:31 crc kubenswrapper[4754]: I1005 22:19:31.262252 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-tv57x/crc-debug-k5sm5" event={"ID":"1f0d1bde-e211-492b-96fd-683293ea6d59","Type":"ContainerStarted","Data":"df58cb762e5f6ebbc5615a4e4c4691bf4adc2be3c65d316f1fb0537a0d9899cb"} Oct 05 22:19:31 crc kubenswrapper[4754]: I1005 22:19:31.262659 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-tv57x/crc-debug-k5sm5" event={"ID":"1f0d1bde-e211-492b-96fd-683293ea6d59","Type":"ContainerStarted","Data":"9a8688554f917d53356beb53d88ec4c4e24b76c3d0d47b728fc1a32276f11509"} Oct 05 22:19:31 crc kubenswrapper[4754]: I1005 22:19:31.286595 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-tv57x/crc-debug-k5sm5" podStartSLOduration=2.286581206 podStartE2EDuration="2.286581206s" podCreationTimestamp="2025-10-05 22:19:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 22:19:31.274422682 +0000 UTC m=+5095.178541392" watchObservedRunningTime="2025-10-05 22:19:31.286581206 +0000 UTC m=+5095.190699916" Oct 05 22:19:32 crc kubenswrapper[4754]: I1005 22:19:32.272104 4754 generic.go:334] "Generic (PLEG): container finished" podID="1f0d1bde-e211-492b-96fd-683293ea6d59" containerID="df58cb762e5f6ebbc5615a4e4c4691bf4adc2be3c65d316f1fb0537a0d9899cb" exitCode=0 Oct 05 22:19:32 crc kubenswrapper[4754]: I1005 22:19:32.272371 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-tv57x/crc-debug-k5sm5" event={"ID":"1f0d1bde-e211-492b-96fd-683293ea6d59","Type":"ContainerDied","Data":"df58cb762e5f6ebbc5615a4e4c4691bf4adc2be3c65d316f1fb0537a0d9899cb"} Oct 05 22:19:33 crc kubenswrapper[4754]: I1005 22:19:33.365896 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-tv57x/crc-debug-k5sm5" Oct 05 22:19:33 crc kubenswrapper[4754]: I1005 22:19:33.468689 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n7kch\" (UniqueName: \"kubernetes.io/projected/1f0d1bde-e211-492b-96fd-683293ea6d59-kube-api-access-n7kch\") pod \"1f0d1bde-e211-492b-96fd-683293ea6d59\" (UID: \"1f0d1bde-e211-492b-96fd-683293ea6d59\") " Oct 05 22:19:33 crc kubenswrapper[4754]: I1005 22:19:33.468786 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1f0d1bde-e211-492b-96fd-683293ea6d59-host\") pod \"1f0d1bde-e211-492b-96fd-683293ea6d59\" (UID: \"1f0d1bde-e211-492b-96fd-683293ea6d59\") " Oct 05 22:19:33 crc kubenswrapper[4754]: I1005 22:19:33.468949 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1f0d1bde-e211-492b-96fd-683293ea6d59-host" (OuterVolumeSpecName: "host") pod "1f0d1bde-e211-492b-96fd-683293ea6d59" (UID: "1f0d1bde-e211-492b-96fd-683293ea6d59"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 05 22:19:33 crc kubenswrapper[4754]: I1005 22:19:33.469221 4754 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1f0d1bde-e211-492b-96fd-683293ea6d59-host\") on node \"crc\" DevicePath \"\"" Oct 05 22:19:33 crc kubenswrapper[4754]: I1005 22:19:33.474144 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1f0d1bde-e211-492b-96fd-683293ea6d59-kube-api-access-n7kch" (OuterVolumeSpecName: "kube-api-access-n7kch") pod "1f0d1bde-e211-492b-96fd-683293ea6d59" (UID: "1f0d1bde-e211-492b-96fd-683293ea6d59"). InnerVolumeSpecName "kube-api-access-n7kch". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 22:19:33 crc kubenswrapper[4754]: I1005 22:19:33.570328 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n7kch\" (UniqueName: \"kubernetes.io/projected/1f0d1bde-e211-492b-96fd-683293ea6d59-kube-api-access-n7kch\") on node \"crc\" DevicePath \"\"" Oct 05 22:19:34 crc kubenswrapper[4754]: I1005 22:19:34.288517 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-tv57x/crc-debug-k5sm5" event={"ID":"1f0d1bde-e211-492b-96fd-683293ea6d59","Type":"ContainerDied","Data":"9a8688554f917d53356beb53d88ec4c4e24b76c3d0d47b728fc1a32276f11509"} Oct 05 22:19:34 crc kubenswrapper[4754]: I1005 22:19:34.288569 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9a8688554f917d53356beb53d88ec4c4e24b76c3d0d47b728fc1a32276f11509" Oct 05 22:19:34 crc kubenswrapper[4754]: I1005 22:19:34.288620 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-tv57x/crc-debug-k5sm5" Oct 05 22:19:37 crc kubenswrapper[4754]: I1005 22:19:37.836980 4754 scope.go:117] "RemoveContainer" containerID="05cd65b5014dfd796e8d71d762a213d6758e74cce340c7705d8831b6a5961a0f" Oct 05 22:19:37 crc kubenswrapper[4754]: E1005 22:19:37.838073 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 22:19:38 crc kubenswrapper[4754]: I1005 22:19:38.584904 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-tv57x/crc-debug-k5sm5"] Oct 05 22:19:38 crc kubenswrapper[4754]: I1005 22:19:38.602484 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-tv57x/crc-debug-k5sm5"] Oct 05 22:19:38 crc kubenswrapper[4754]: I1005 22:19:38.850718 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1f0d1bde-e211-492b-96fd-683293ea6d59" path="/var/lib/kubelet/pods/1f0d1bde-e211-492b-96fd-683293ea6d59/volumes" Oct 05 22:19:39 crc kubenswrapper[4754]: I1005 22:19:39.805637 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-tv57x/crc-debug-lm2b5"] Oct 05 22:19:39 crc kubenswrapper[4754]: E1005 22:19:39.806334 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f0d1bde-e211-492b-96fd-683293ea6d59" containerName="container-00" Oct 05 22:19:39 crc kubenswrapper[4754]: I1005 22:19:39.806434 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f0d1bde-e211-492b-96fd-683293ea6d59" containerName="container-00" Oct 05 22:19:39 crc kubenswrapper[4754]: I1005 22:19:39.806772 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f0d1bde-e211-492b-96fd-683293ea6d59" containerName="container-00" Oct 05 22:19:39 crc kubenswrapper[4754]: I1005 22:19:39.807662 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-tv57x/crc-debug-lm2b5" Oct 05 22:19:39 crc kubenswrapper[4754]: I1005 22:19:39.887414 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7630e6d3-a51d-4f7b-97c2-a98a4bb6c5da-host\") pod \"crc-debug-lm2b5\" (UID: \"7630e6d3-a51d-4f7b-97c2-a98a4bb6c5da\") " pod="openshift-must-gather-tv57x/crc-debug-lm2b5" Oct 05 22:19:39 crc kubenswrapper[4754]: I1005 22:19:39.888386 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8qscz\" (UniqueName: \"kubernetes.io/projected/7630e6d3-a51d-4f7b-97c2-a98a4bb6c5da-kube-api-access-8qscz\") pod \"crc-debug-lm2b5\" (UID: \"7630e6d3-a51d-4f7b-97c2-a98a4bb6c5da\") " pod="openshift-must-gather-tv57x/crc-debug-lm2b5" Oct 05 22:19:39 crc kubenswrapper[4754]: I1005 22:19:39.991845 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7630e6d3-a51d-4f7b-97c2-a98a4bb6c5da-host\") pod \"crc-debug-lm2b5\" (UID: \"7630e6d3-a51d-4f7b-97c2-a98a4bb6c5da\") " pod="openshift-must-gather-tv57x/crc-debug-lm2b5" Oct 05 22:19:39 crc kubenswrapper[4754]: I1005 22:19:39.991987 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8qscz\" (UniqueName: \"kubernetes.io/projected/7630e6d3-a51d-4f7b-97c2-a98a4bb6c5da-kube-api-access-8qscz\") pod \"crc-debug-lm2b5\" (UID: \"7630e6d3-a51d-4f7b-97c2-a98a4bb6c5da\") " pod="openshift-must-gather-tv57x/crc-debug-lm2b5" Oct 05 22:19:39 crc kubenswrapper[4754]: I1005 22:19:39.992288 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7630e6d3-a51d-4f7b-97c2-a98a4bb6c5da-host\") pod \"crc-debug-lm2b5\" (UID: \"7630e6d3-a51d-4f7b-97c2-a98a4bb6c5da\") " pod="openshift-must-gather-tv57x/crc-debug-lm2b5" Oct 05 22:19:40 crc kubenswrapper[4754]: I1005 22:19:40.090722 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8qscz\" (UniqueName: \"kubernetes.io/projected/7630e6d3-a51d-4f7b-97c2-a98a4bb6c5da-kube-api-access-8qscz\") pod \"crc-debug-lm2b5\" (UID: \"7630e6d3-a51d-4f7b-97c2-a98a4bb6c5da\") " pod="openshift-must-gather-tv57x/crc-debug-lm2b5" Oct 05 22:19:40 crc kubenswrapper[4754]: I1005 22:19:40.138835 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-tv57x/crc-debug-lm2b5" Oct 05 22:19:40 crc kubenswrapper[4754]: I1005 22:19:40.339026 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-tv57x/crc-debug-lm2b5" event={"ID":"7630e6d3-a51d-4f7b-97c2-a98a4bb6c5da","Type":"ContainerStarted","Data":"fc7bb46ef5e7936316428e4f321c1639f2e1cb5e18b6e602131a3ecc6b74a288"} Oct 05 22:19:41 crc kubenswrapper[4754]: I1005 22:19:41.351569 4754 generic.go:334] "Generic (PLEG): container finished" podID="7630e6d3-a51d-4f7b-97c2-a98a4bb6c5da" containerID="d5a64d7378f80e43527f67a81c090bf620a6ecab4bcb90ddd35190fb644e90dc" exitCode=0 Oct 05 22:19:41 crc kubenswrapper[4754]: I1005 22:19:41.351618 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-tv57x/crc-debug-lm2b5" event={"ID":"7630e6d3-a51d-4f7b-97c2-a98a4bb6c5da","Type":"ContainerDied","Data":"d5a64d7378f80e43527f67a81c090bf620a6ecab4bcb90ddd35190fb644e90dc"} Oct 05 22:19:41 crc kubenswrapper[4754]: I1005 22:19:41.401196 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-tv57x/crc-debug-lm2b5"] Oct 05 22:19:41 crc kubenswrapper[4754]: I1005 22:19:41.407982 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-tv57x/crc-debug-lm2b5"] Oct 05 22:19:42 crc kubenswrapper[4754]: I1005 22:19:42.447088 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-tv57x/crc-debug-lm2b5" Oct 05 22:19:42 crc kubenswrapper[4754]: I1005 22:19:42.536024 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7630e6d3-a51d-4f7b-97c2-a98a4bb6c5da-host\") pod \"7630e6d3-a51d-4f7b-97c2-a98a4bb6c5da\" (UID: \"7630e6d3-a51d-4f7b-97c2-a98a4bb6c5da\") " Oct 05 22:19:42 crc kubenswrapper[4754]: I1005 22:19:42.536132 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8qscz\" (UniqueName: \"kubernetes.io/projected/7630e6d3-a51d-4f7b-97c2-a98a4bb6c5da-kube-api-access-8qscz\") pod \"7630e6d3-a51d-4f7b-97c2-a98a4bb6c5da\" (UID: \"7630e6d3-a51d-4f7b-97c2-a98a4bb6c5da\") " Oct 05 22:19:42 crc kubenswrapper[4754]: I1005 22:19:42.536165 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7630e6d3-a51d-4f7b-97c2-a98a4bb6c5da-host" (OuterVolumeSpecName: "host") pod "7630e6d3-a51d-4f7b-97c2-a98a4bb6c5da" (UID: "7630e6d3-a51d-4f7b-97c2-a98a4bb6c5da"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 05 22:19:42 crc kubenswrapper[4754]: I1005 22:19:42.536522 4754 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7630e6d3-a51d-4f7b-97c2-a98a4bb6c5da-host\") on node \"crc\" DevicePath \"\"" Oct 05 22:19:42 crc kubenswrapper[4754]: I1005 22:19:42.548759 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7630e6d3-a51d-4f7b-97c2-a98a4bb6c5da-kube-api-access-8qscz" (OuterVolumeSpecName: "kube-api-access-8qscz") pod "7630e6d3-a51d-4f7b-97c2-a98a4bb6c5da" (UID: "7630e6d3-a51d-4f7b-97c2-a98a4bb6c5da"). InnerVolumeSpecName "kube-api-access-8qscz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 22:19:42 crc kubenswrapper[4754]: I1005 22:19:42.638456 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8qscz\" (UniqueName: \"kubernetes.io/projected/7630e6d3-a51d-4f7b-97c2-a98a4bb6c5da-kube-api-access-8qscz\") on node \"crc\" DevicePath \"\"" Oct 05 22:19:42 crc kubenswrapper[4754]: I1005 22:19:42.845991 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7630e6d3-a51d-4f7b-97c2-a98a4bb6c5da" path="/var/lib/kubelet/pods/7630e6d3-a51d-4f7b-97c2-a98a4bb6c5da/volumes" Oct 05 22:19:43 crc kubenswrapper[4754]: I1005 22:19:43.218013 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_aded1cecd2f966ec868aa0a6ebe9c364f1aced79b6d6348d02479f11239rr6r_5f6867a7-6e2b-4f7a-bcdd-a0ff2035eee8/util/0.log" Oct 05 22:19:43 crc kubenswrapper[4754]: I1005 22:19:43.367285 4754 scope.go:117] "RemoveContainer" containerID="d5a64d7378f80e43527f67a81c090bf620a6ecab4bcb90ddd35190fb644e90dc" Oct 05 22:19:43 crc kubenswrapper[4754]: I1005 22:19:43.367409 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-tv57x/crc-debug-lm2b5" Oct 05 22:19:43 crc kubenswrapper[4754]: I1005 22:19:43.394689 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_aded1cecd2f966ec868aa0a6ebe9c364f1aced79b6d6348d02479f11239rr6r_5f6867a7-6e2b-4f7a-bcdd-a0ff2035eee8/pull/0.log" Oct 05 22:19:43 crc kubenswrapper[4754]: I1005 22:19:43.456076 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_aded1cecd2f966ec868aa0a6ebe9c364f1aced79b6d6348d02479f11239rr6r_5f6867a7-6e2b-4f7a-bcdd-a0ff2035eee8/util/0.log" Oct 05 22:19:43 crc kubenswrapper[4754]: I1005 22:19:43.457864 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_aded1cecd2f966ec868aa0a6ebe9c364f1aced79b6d6348d02479f11239rr6r_5f6867a7-6e2b-4f7a-bcdd-a0ff2035eee8/pull/0.log" Oct 05 22:19:43 crc kubenswrapper[4754]: I1005 22:19:43.605061 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_aded1cecd2f966ec868aa0a6ebe9c364f1aced79b6d6348d02479f11239rr6r_5f6867a7-6e2b-4f7a-bcdd-a0ff2035eee8/pull/0.log" Oct 05 22:19:43 crc kubenswrapper[4754]: I1005 22:19:43.646349 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_aded1cecd2f966ec868aa0a6ebe9c364f1aced79b6d6348d02479f11239rr6r_5f6867a7-6e2b-4f7a-bcdd-a0ff2035eee8/util/0.log" Oct 05 22:19:43 crc kubenswrapper[4754]: I1005 22:19:43.664153 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_aded1cecd2f966ec868aa0a6ebe9c364f1aced79b6d6348d02479f11239rr6r_5f6867a7-6e2b-4f7a-bcdd-a0ff2035eee8/extract/0.log" Oct 05 22:19:43 crc kubenswrapper[4754]: I1005 22:19:43.825792 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-5f7c849b98-n9ksx_9a1cc24e-8371-44cb-bd37-5b765853fdff/kube-rbac-proxy/0.log" Oct 05 22:19:43 crc kubenswrapper[4754]: I1005 22:19:43.919813 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-5f7c849b98-n9ksx_9a1cc24e-8371-44cb-bd37-5b765853fdff/manager/0.log" Oct 05 22:19:43 crc kubenswrapper[4754]: I1005 22:19:43.995240 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-7d4d4f8d-kcxvd_47cafc06-0f75-48ee-bae5-4e30352f4572/kube-rbac-proxy/0.log" Oct 05 22:19:44 crc kubenswrapper[4754]: I1005 22:19:44.086535 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-7d4d4f8d-kcxvd_47cafc06-0f75-48ee-bae5-4e30352f4572/manager/0.log" Oct 05 22:19:44 crc kubenswrapper[4754]: I1005 22:19:44.178941 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-75dfd9b554-fxx5s_3ec442d5-220a-4d2f-8f62-a13ff86a3229/kube-rbac-proxy/0.log" Oct 05 22:19:44 crc kubenswrapper[4754]: I1005 22:19:44.207768 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-75dfd9b554-fxx5s_3ec442d5-220a-4d2f-8f62-a13ff86a3229/manager/0.log" Oct 05 22:19:44 crc kubenswrapper[4754]: I1005 22:19:44.422184 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-5568b5d68-f5rjh_62159d2d-ec19-4b44-89df-846faa5757dc/manager/0.log" Oct 05 22:19:44 crc kubenswrapper[4754]: I1005 22:19:44.431663 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-5568b5d68-f5rjh_62159d2d-ec19-4b44-89df-846faa5757dc/kube-rbac-proxy/0.log" Oct 05 22:19:44 crc kubenswrapper[4754]: I1005 22:19:44.605640 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-8f58bc9db-xsrhv_82184642-9d7b-4ada-b7b2-efde6c91cb14/kube-rbac-proxy/0.log" Oct 05 22:19:44 crc kubenswrapper[4754]: I1005 22:19:44.623872 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-8f58bc9db-xsrhv_82184642-9d7b-4ada-b7b2-efde6c91cb14/manager/0.log" Oct 05 22:19:44 crc kubenswrapper[4754]: I1005 22:19:44.667672 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-54876c876f-svzwt_f560133d-b3b5-4791-b69a-a700367f0d96/kube-rbac-proxy/0.log" Oct 05 22:19:44 crc kubenswrapper[4754]: I1005 22:19:44.847301 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-54876c876f-svzwt_f560133d-b3b5-4791-b69a-a700367f0d96/manager/0.log" Oct 05 22:19:44 crc kubenswrapper[4754]: I1005 22:19:44.880120 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-658588b8c9-z7tq9_4bb20421-ce34-44ee-8740-82eedd4716f3/kube-rbac-proxy/0.log" Oct 05 22:19:45 crc kubenswrapper[4754]: I1005 22:19:45.008068 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-658588b8c9-z7tq9_4bb20421-ce34-44ee-8740-82eedd4716f3/manager/0.log" Oct 05 22:19:45 crc kubenswrapper[4754]: I1005 22:19:45.079916 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-649675d675-v7285_3ddbeeee-8a3e-4f3c-97e0-c7d8d5929345/kube-rbac-proxy/0.log" Oct 05 22:19:45 crc kubenswrapper[4754]: I1005 22:19:45.128414 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-649675d675-v7285_3ddbeeee-8a3e-4f3c-97e0-c7d8d5929345/manager/0.log" Oct 05 22:19:45 crc kubenswrapper[4754]: I1005 22:19:45.308061 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-655d88ccb9-l56qv_2735f261-cad9-4dff-9136-931a45c85ac5/manager/0.log" Oct 05 22:19:45 crc kubenswrapper[4754]: I1005 22:19:45.328202 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-655d88ccb9-l56qv_2735f261-cad9-4dff-9136-931a45c85ac5/kube-rbac-proxy/0.log" Oct 05 22:19:45 crc kubenswrapper[4754]: I1005 22:19:45.415400 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-65d89cfd9f-msh87_5757c596-103f-4a00-ac16-ed9e6c9e4719/kube-rbac-proxy/0.log" Oct 05 22:19:45 crc kubenswrapper[4754]: I1005 22:19:45.544058 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-65d89cfd9f-msh87_5757c596-103f-4a00-ac16-ed9e6c9e4719/manager/0.log" Oct 05 22:19:45 crc kubenswrapper[4754]: I1005 22:19:45.552373 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-6cd6d7bdf5-f5jxx_5a1312fe-c6d6-41de-a442-b044db904a6d/kube-rbac-proxy/0.log" Oct 05 22:19:45 crc kubenswrapper[4754]: I1005 22:19:45.597407 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-6cd6d7bdf5-f5jxx_5a1312fe-c6d6-41de-a442-b044db904a6d/manager/0.log" Oct 05 22:19:45 crc kubenswrapper[4754]: I1005 22:19:45.776138 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-8d984cc4d-7xzhb_ff3ba6ea-c2d8-4a3e-b0a0-58fd6df5b196/manager/0.log" Oct 05 22:19:45 crc kubenswrapper[4754]: I1005 22:19:45.870593 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-8d984cc4d-7xzhb_ff3ba6ea-c2d8-4a3e-b0a0-58fd6df5b196/kube-rbac-proxy/0.log" Oct 05 22:19:46 crc kubenswrapper[4754]: I1005 22:19:46.050512 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-7c7fc454ff-4fb5b_e0b701ef-8b8d-4717-aa27-07233f6b6c15/manager/0.log" Oct 05 22:19:46 crc kubenswrapper[4754]: I1005 22:19:46.055202 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-7c7fc454ff-4fb5b_e0b701ef-8b8d-4717-aa27-07233f6b6c15/kube-rbac-proxy/0.log" Oct 05 22:19:46 crc kubenswrapper[4754]: I1005 22:19:46.101815 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-7468f855d8-w8z85_cb4a4798-9283-4633-9b3b-e2d72faa221f/kube-rbac-proxy/0.log" Oct 05 22:19:46 crc kubenswrapper[4754]: I1005 22:19:46.260624 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-7468f855d8-w8z85_cb4a4798-9283-4633-9b3b-e2d72faa221f/manager/0.log" Oct 05 22:19:46 crc kubenswrapper[4754]: I1005 22:19:46.317271 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-5dfbbd665cfq946_c336d534-f819-4859-8e15-5f15b68a36ad/manager/0.log" Oct 05 22:19:46 crc kubenswrapper[4754]: I1005 22:19:46.330376 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-5dfbbd665cfq946_c336d534-f819-4859-8e15-5f15b68a36ad/kube-rbac-proxy/0.log" Oct 05 22:19:46 crc kubenswrapper[4754]: I1005 22:19:46.467483 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-55895b89bc-zzbvh_f131e747-362b-4442-8377-0c00aedfd8ae/kube-rbac-proxy/0.log" Oct 05 22:19:46 crc kubenswrapper[4754]: I1005 22:19:46.680457 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-56b64c5fcb-gzvmj_21fc1c35-c245-457c-a66f-97536332ed52/kube-rbac-proxy/0.log" Oct 05 22:19:46 crc kubenswrapper[4754]: I1005 22:19:46.850391 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-56b64c5fcb-gzvmj_21fc1c35-c245-457c-a66f-97536332ed52/operator/0.log" Oct 05 22:19:47 crc kubenswrapper[4754]: I1005 22:19:47.184063 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-pgjjb_506033e6-79e8-4b4e-acb8-97e63316a15b/registry-server/0.log" Oct 05 22:19:47 crc kubenswrapper[4754]: I1005 22:19:47.382205 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-579449c7d5-zcct6_b18dee24-0d55-4805-8d76-623180998686/kube-rbac-proxy/0.log" Oct 05 22:19:47 crc kubenswrapper[4754]: I1005 22:19:47.520929 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-55895b89bc-zzbvh_f131e747-362b-4442-8377-0c00aedfd8ae/manager/0.log" Oct 05 22:19:47 crc kubenswrapper[4754]: I1005 22:19:47.532479 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-579449c7d5-zcct6_b18dee24-0d55-4805-8d76-623180998686/manager/0.log" Oct 05 22:19:47 crc kubenswrapper[4754]: I1005 22:19:47.591196 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-54689d9f88-4lrzz_6316859b-8ab7-4b46-b1fa-c7f893d39a95/kube-rbac-proxy/0.log" Oct 05 22:19:47 crc kubenswrapper[4754]: I1005 22:19:47.622358 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-54689d9f88-4lrzz_6316859b-8ab7-4b46-b1fa-c7f893d39a95/manager/0.log" Oct 05 22:19:47 crc kubenswrapper[4754]: I1005 22:19:47.718213 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-5f97d8c699-hkb96_97f32c3b-569a-4f15-9f7c-fb18299dec30/operator/0.log" Oct 05 22:19:47 crc kubenswrapper[4754]: I1005 22:19:47.826577 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-6859f9b676-hgkv4_2b729715-7fd9-4ca5-b4dd-cb0eb0034aac/manager/0.log" Oct 05 22:19:47 crc kubenswrapper[4754]: I1005 22:19:47.834110 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-6859f9b676-hgkv4_2b729715-7fd9-4ca5-b4dd-cb0eb0034aac/kube-rbac-proxy/0.log" Oct 05 22:19:47 crc kubenswrapper[4754]: I1005 22:19:47.907422 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-5d4d74dd89-h44bd_f48f3071-cd94-49dd-bfa0-1ec74f495c72/kube-rbac-proxy/0.log" Oct 05 22:19:48 crc kubenswrapper[4754]: I1005 22:19:48.027685 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-5d4d74dd89-h44bd_f48f3071-cd94-49dd-bfa0-1ec74f495c72/manager/0.log" Oct 05 22:19:48 crc kubenswrapper[4754]: I1005 22:19:48.031301 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5cd5cb47d7-fx25z_b1cc32e7-f619-4a0e-b0f9-195e36c82d01/kube-rbac-proxy/0.log" Oct 05 22:19:48 crc kubenswrapper[4754]: I1005 22:19:48.093995 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5cd5cb47d7-fx25z_b1cc32e7-f619-4a0e-b0f9-195e36c82d01/manager/0.log" Oct 05 22:19:48 crc kubenswrapper[4754]: I1005 22:19:48.608527 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-6cbc6dd547-zq79b_cea6487f-b088-4f9e-a42d-6a3517df7669/manager/0.log" Oct 05 22:19:48 crc kubenswrapper[4754]: I1005 22:19:48.614182 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-6cbc6dd547-zq79b_cea6487f-b088-4f9e-a42d-6a3517df7669/kube-rbac-proxy/0.log" Oct 05 22:19:52 crc kubenswrapper[4754]: I1005 22:19:52.837626 4754 scope.go:117] "RemoveContainer" containerID="05cd65b5014dfd796e8d71d762a213d6758e74cce340c7705d8831b6a5961a0f" Oct 05 22:19:52 crc kubenswrapper[4754]: E1005 22:19:52.838270 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 22:20:05 crc kubenswrapper[4754]: I1005 22:20:05.988253 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-5wx56_d1b2b973-8b95-4f6a-83f1-468f84fb006d/control-plane-machine-set-operator/0.log" Oct 05 22:20:06 crc kubenswrapper[4754]: I1005 22:20:06.028394 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-p5wqw_c463939d-d2d1-4049-b438-3754a118ff4b/kube-rbac-proxy/0.log" Oct 05 22:20:06 crc kubenswrapper[4754]: I1005 22:20:06.156239 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-p5wqw_c463939d-d2d1-4049-b438-3754a118ff4b/machine-api-operator/0.log" Oct 05 22:20:06 crc kubenswrapper[4754]: I1005 22:20:06.842766 4754 scope.go:117] "RemoveContainer" containerID="05cd65b5014dfd796e8d71d762a213d6758e74cce340c7705d8831b6a5961a0f" Oct 05 22:20:06 crc kubenswrapper[4754]: E1005 22:20:06.843238 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 22:20:19 crc kubenswrapper[4754]: I1005 22:20:19.607276 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-c5gxb_e31db265-82f2-431c-af44-9a44bcb4762c/cert-manager-controller/0.log" Oct 05 22:20:19 crc kubenswrapper[4754]: I1005 22:20:19.758807 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-pqw5s_f060400d-9b95-4e44-b94a-ba61d23e307c/cert-manager-cainjector/0.log" Oct 05 22:20:19 crc kubenswrapper[4754]: I1005 22:20:19.837395 4754 scope.go:117] "RemoveContainer" containerID="05cd65b5014dfd796e8d71d762a213d6758e74cce340c7705d8831b6a5961a0f" Oct 05 22:20:19 crc kubenswrapper[4754]: E1005 22:20:19.837718 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 22:20:19 crc kubenswrapper[4754]: I1005 22:20:19.854112 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-r2txk_577f8ef5-2c2c-4b67-91b5-55ce19fc52d2/cert-manager-webhook/0.log" Oct 05 22:20:33 crc kubenswrapper[4754]: I1005 22:20:33.333600 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-6b874cbd85-pvfm8_785a344e-fb90-43ec-a951-962783b9ccbf/nmstate-console-plugin/0.log" Oct 05 22:20:33 crc kubenswrapper[4754]: I1005 22:20:33.449336 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-d4qnm_76cf164c-fcc4-49e8-866f-e4ec435a8044/nmstate-handler/0.log" Oct 05 22:20:33 crc kubenswrapper[4754]: I1005 22:20:33.508244 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-7469b_b381c1ba-7339-4ab8-bc4a-d3887197d1ea/nmstate-metrics/0.log" Oct 05 22:20:33 crc kubenswrapper[4754]: I1005 22:20:33.523291 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-7469b_b381c1ba-7339-4ab8-bc4a-d3887197d1ea/kube-rbac-proxy/0.log" Oct 05 22:20:33 crc kubenswrapper[4754]: I1005 22:20:33.710717 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-858ddd8f98-z2c5l_262b4fe5-27b8-41fe-8d8c-aed9ca555a6e/nmstate-operator/0.log" Oct 05 22:20:33 crc kubenswrapper[4754]: I1005 22:20:33.742798 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-6cdbc54649-nds52_30f05ad2-a252-4b9c-890d-892b90aaa568/nmstate-webhook/0.log" Oct 05 22:20:34 crc kubenswrapper[4754]: I1005 22:20:34.838029 4754 scope.go:117] "RemoveContainer" containerID="05cd65b5014dfd796e8d71d762a213d6758e74cce340c7705d8831b6a5961a0f" Oct 05 22:20:34 crc kubenswrapper[4754]: E1005 22:20:34.838264 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 22:20:45 crc kubenswrapper[4754]: I1005 22:20:45.837544 4754 scope.go:117] "RemoveContainer" containerID="05cd65b5014dfd796e8d71d762a213d6758e74cce340c7705d8831b6a5961a0f" Oct 05 22:20:45 crc kubenswrapper[4754]: E1005 22:20:45.838467 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 22:20:50 crc kubenswrapper[4754]: I1005 22:20:50.191475 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-78nmb_22f6b8d5-a7f7-4db9-bd76-eb9e5e23ea79/kube-rbac-proxy/0.log" Oct 05 22:20:50 crc kubenswrapper[4754]: I1005 22:20:50.357303 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-78nmb_22f6b8d5-a7f7-4db9-bd76-eb9e5e23ea79/controller/0.log" Oct 05 22:20:50 crc kubenswrapper[4754]: I1005 22:20:50.442366 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tfh49_dbf73baa-e798-43fa-95de-daec87f5809a/cp-frr-files/0.log" Oct 05 22:20:50 crc kubenswrapper[4754]: I1005 22:20:50.610802 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tfh49_dbf73baa-e798-43fa-95de-daec87f5809a/cp-frr-files/0.log" Oct 05 22:20:50 crc kubenswrapper[4754]: I1005 22:20:50.632651 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tfh49_dbf73baa-e798-43fa-95de-daec87f5809a/cp-metrics/0.log" Oct 05 22:20:50 crc kubenswrapper[4754]: I1005 22:20:50.646477 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tfh49_dbf73baa-e798-43fa-95de-daec87f5809a/cp-reloader/0.log" Oct 05 22:20:50 crc kubenswrapper[4754]: I1005 22:20:50.699957 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tfh49_dbf73baa-e798-43fa-95de-daec87f5809a/cp-reloader/0.log" Oct 05 22:20:50 crc kubenswrapper[4754]: I1005 22:20:50.883573 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tfh49_dbf73baa-e798-43fa-95de-daec87f5809a/cp-reloader/0.log" Oct 05 22:20:50 crc kubenswrapper[4754]: I1005 22:20:50.923467 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tfh49_dbf73baa-e798-43fa-95de-daec87f5809a/cp-frr-files/0.log" Oct 05 22:20:50 crc kubenswrapper[4754]: I1005 22:20:50.948334 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tfh49_dbf73baa-e798-43fa-95de-daec87f5809a/cp-metrics/0.log" Oct 05 22:20:50 crc kubenswrapper[4754]: I1005 22:20:50.993145 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tfh49_dbf73baa-e798-43fa-95de-daec87f5809a/cp-metrics/0.log" Oct 05 22:20:51 crc kubenswrapper[4754]: I1005 22:20:51.176962 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tfh49_dbf73baa-e798-43fa-95de-daec87f5809a/cp-reloader/0.log" Oct 05 22:20:51 crc kubenswrapper[4754]: I1005 22:20:51.182776 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tfh49_dbf73baa-e798-43fa-95de-daec87f5809a/cp-frr-files/0.log" Oct 05 22:20:51 crc kubenswrapper[4754]: I1005 22:20:51.227139 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tfh49_dbf73baa-e798-43fa-95de-daec87f5809a/cp-metrics/0.log" Oct 05 22:20:51 crc kubenswrapper[4754]: I1005 22:20:51.246768 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tfh49_dbf73baa-e798-43fa-95de-daec87f5809a/controller/0.log" Oct 05 22:20:51 crc kubenswrapper[4754]: I1005 22:20:51.393776 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tfh49_dbf73baa-e798-43fa-95de-daec87f5809a/frr-metrics/0.log" Oct 05 22:20:51 crc kubenswrapper[4754]: I1005 22:20:51.428920 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tfh49_dbf73baa-e798-43fa-95de-daec87f5809a/kube-rbac-proxy/0.log" Oct 05 22:20:51 crc kubenswrapper[4754]: I1005 22:20:51.540821 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tfh49_dbf73baa-e798-43fa-95de-daec87f5809a/kube-rbac-proxy-frr/0.log" Oct 05 22:20:51 crc kubenswrapper[4754]: I1005 22:20:51.634216 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tfh49_dbf73baa-e798-43fa-95de-daec87f5809a/reloader/0.log" Oct 05 22:20:51 crc kubenswrapper[4754]: I1005 22:20:51.815993 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-64bf5d555-zt7sz_69687995-5b0d-4aae-9551-1b1b9cb2bec3/frr-k8s-webhook-server/0.log" Oct 05 22:20:52 crc kubenswrapper[4754]: I1005 22:20:52.537172 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-645fb4cd96-7qplt_697d9fce-0466-4ba2-a532-214dfd58a20e/manager/0.log" Oct 05 22:20:52 crc kubenswrapper[4754]: I1005 22:20:52.556558 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-5ff795db68-8xphz_56375d90-5d62-4d6f-897f-28ca18bf3682/webhook-server/0.log" Oct 05 22:20:52 crc kubenswrapper[4754]: I1005 22:20:52.613231 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tfh49_dbf73baa-e798-43fa-95de-daec87f5809a/frr/0.log" Oct 05 22:20:52 crc kubenswrapper[4754]: I1005 22:20:52.769879 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-hkvd8_34ab3bd8-96ff-4f52-ae2e-e8cbd748827d/kube-rbac-proxy/0.log" Oct 05 22:20:53 crc kubenswrapper[4754]: I1005 22:20:53.125575 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-hkvd8_34ab3bd8-96ff-4f52-ae2e-e8cbd748827d/speaker/0.log" Oct 05 22:20:57 crc kubenswrapper[4754]: I1005 22:20:57.837885 4754 scope.go:117] "RemoveContainer" containerID="05cd65b5014dfd796e8d71d762a213d6758e74cce340c7705d8831b6a5961a0f" Oct 05 22:20:57 crc kubenswrapper[4754]: E1005 22:20:57.838604 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 22:21:06 crc kubenswrapper[4754]: I1005 22:21:06.284596 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2kzp4h_3da3d4d5-ef4b-4dde-8b69-0d8e566703c1/util/0.log" Oct 05 22:21:06 crc kubenswrapper[4754]: I1005 22:21:06.456686 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2kzp4h_3da3d4d5-ef4b-4dde-8b69-0d8e566703c1/util/0.log" Oct 05 22:21:06 crc kubenswrapper[4754]: I1005 22:21:06.465441 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2kzp4h_3da3d4d5-ef4b-4dde-8b69-0d8e566703c1/pull/0.log" Oct 05 22:21:06 crc kubenswrapper[4754]: I1005 22:21:06.506460 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2kzp4h_3da3d4d5-ef4b-4dde-8b69-0d8e566703c1/pull/0.log" Oct 05 22:21:06 crc kubenswrapper[4754]: I1005 22:21:06.646461 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2kzp4h_3da3d4d5-ef4b-4dde-8b69-0d8e566703c1/util/0.log" Oct 05 22:21:06 crc kubenswrapper[4754]: I1005 22:21:06.669390 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2kzp4h_3da3d4d5-ef4b-4dde-8b69-0d8e566703c1/extract/0.log" Oct 05 22:21:06 crc kubenswrapper[4754]: I1005 22:21:06.716907 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2kzp4h_3da3d4d5-ef4b-4dde-8b69-0d8e566703c1/pull/0.log" Oct 05 22:21:06 crc kubenswrapper[4754]: I1005 22:21:06.826903 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-m7sf7_5a34915c-4494-448a-b835-6f1276b105fa/extract-utilities/0.log" Oct 05 22:21:07 crc kubenswrapper[4754]: I1005 22:21:07.036692 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-m7sf7_5a34915c-4494-448a-b835-6f1276b105fa/extract-utilities/0.log" Oct 05 22:21:07 crc kubenswrapper[4754]: I1005 22:21:07.036731 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-m7sf7_5a34915c-4494-448a-b835-6f1276b105fa/extract-content/0.log" Oct 05 22:21:07 crc kubenswrapper[4754]: I1005 22:21:07.056059 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-m7sf7_5a34915c-4494-448a-b835-6f1276b105fa/extract-content/0.log" Oct 05 22:21:07 crc kubenswrapper[4754]: I1005 22:21:07.226407 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-m7sf7_5a34915c-4494-448a-b835-6f1276b105fa/extract-utilities/0.log" Oct 05 22:21:07 crc kubenswrapper[4754]: I1005 22:21:07.264005 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-m7sf7_5a34915c-4494-448a-b835-6f1276b105fa/extract-content/0.log" Oct 05 22:21:07 crc kubenswrapper[4754]: I1005 22:21:07.457420 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-pkf48_095d3b9f-145a-4bc8-adca-a1831739e193/extract-utilities/0.log" Oct 05 22:21:07 crc kubenswrapper[4754]: I1005 22:21:07.758850 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-m7sf7_5a34915c-4494-448a-b835-6f1276b105fa/registry-server/0.log" Oct 05 22:21:07 crc kubenswrapper[4754]: I1005 22:21:07.770359 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-pkf48_095d3b9f-145a-4bc8-adca-a1831739e193/extract-utilities/0.log" Oct 05 22:21:07 crc kubenswrapper[4754]: I1005 22:21:07.810953 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-pkf48_095d3b9f-145a-4bc8-adca-a1831739e193/extract-content/0.log" Oct 05 22:21:07 crc kubenswrapper[4754]: I1005 22:21:07.849340 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-pkf48_095d3b9f-145a-4bc8-adca-a1831739e193/extract-content/0.log" Oct 05 22:21:08 crc kubenswrapper[4754]: I1005 22:21:08.005136 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-pkf48_095d3b9f-145a-4bc8-adca-a1831739e193/extract-content/0.log" Oct 05 22:21:08 crc kubenswrapper[4754]: I1005 22:21:08.026870 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-pkf48_095d3b9f-145a-4bc8-adca-a1831739e193/extract-utilities/0.log" Oct 05 22:21:08 crc kubenswrapper[4754]: I1005 22:21:08.280783 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cpbjks_f7edcdad-0804-4ae8-bb6d-1118b6d0f620/util/0.log" Oct 05 22:21:08 crc kubenswrapper[4754]: I1005 22:21:08.463313 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cpbjks_f7edcdad-0804-4ae8-bb6d-1118b6d0f620/util/0.log" Oct 05 22:21:08 crc kubenswrapper[4754]: I1005 22:21:08.470306 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cpbjks_f7edcdad-0804-4ae8-bb6d-1118b6d0f620/pull/0.log" Oct 05 22:21:08 crc kubenswrapper[4754]: I1005 22:21:08.561019 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cpbjks_f7edcdad-0804-4ae8-bb6d-1118b6d0f620/pull/0.log" Oct 05 22:21:08 crc kubenswrapper[4754]: I1005 22:21:08.823006 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-pkf48_095d3b9f-145a-4bc8-adca-a1831739e193/registry-server/0.log" Oct 05 22:21:08 crc kubenswrapper[4754]: I1005 22:21:08.881696 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cpbjks_f7edcdad-0804-4ae8-bb6d-1118b6d0f620/pull/0.log" Oct 05 22:21:08 crc kubenswrapper[4754]: I1005 22:21:08.883236 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cpbjks_f7edcdad-0804-4ae8-bb6d-1118b6d0f620/util/0.log" Oct 05 22:21:08 crc kubenswrapper[4754]: I1005 22:21:08.902662 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cpbjks_f7edcdad-0804-4ae8-bb6d-1118b6d0f620/extract/0.log" Oct 05 22:21:09 crc kubenswrapper[4754]: I1005 22:21:09.025558 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-xftcm_ed62afda-bbd2-4326-8101-0abe44ecb2f5/marketplace-operator/0.log" Oct 05 22:21:09 crc kubenswrapper[4754]: I1005 22:21:09.167758 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-qdsv6_f5951a44-c13f-4cef-a2cf-6803479067f8/extract-utilities/0.log" Oct 05 22:21:09 crc kubenswrapper[4754]: I1005 22:21:09.283791 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-qdsv6_f5951a44-c13f-4cef-a2cf-6803479067f8/extract-utilities/0.log" Oct 05 22:21:09 crc kubenswrapper[4754]: I1005 22:21:09.298517 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-qdsv6_f5951a44-c13f-4cef-a2cf-6803479067f8/extract-content/0.log" Oct 05 22:21:09 crc kubenswrapper[4754]: I1005 22:21:09.349982 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-qdsv6_f5951a44-c13f-4cef-a2cf-6803479067f8/extract-content/0.log" Oct 05 22:21:09 crc kubenswrapper[4754]: I1005 22:21:09.509000 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-qdsv6_f5951a44-c13f-4cef-a2cf-6803479067f8/extract-utilities/0.log" Oct 05 22:21:09 crc kubenswrapper[4754]: I1005 22:21:09.526919 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-qdsv6_f5951a44-c13f-4cef-a2cf-6803479067f8/extract-content/0.log" Oct 05 22:21:09 crc kubenswrapper[4754]: I1005 22:21:09.690538 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-qdsv6_f5951a44-c13f-4cef-a2cf-6803479067f8/registry-server/0.log" Oct 05 22:21:09 crc kubenswrapper[4754]: I1005 22:21:09.767012 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-2lvpm_b27b18ff-2e0a-446d-b605-5b74f825260b/extract-utilities/0.log" Oct 05 22:21:09 crc kubenswrapper[4754]: I1005 22:21:09.937909 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-2lvpm_b27b18ff-2e0a-446d-b605-5b74f825260b/extract-utilities/0.log" Oct 05 22:21:09 crc kubenswrapper[4754]: I1005 22:21:09.976284 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-2lvpm_b27b18ff-2e0a-446d-b605-5b74f825260b/extract-content/0.log" Oct 05 22:21:09 crc kubenswrapper[4754]: I1005 22:21:09.978940 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-2lvpm_b27b18ff-2e0a-446d-b605-5b74f825260b/extract-content/0.log" Oct 05 22:21:10 crc kubenswrapper[4754]: I1005 22:21:10.125071 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-2lvpm_b27b18ff-2e0a-446d-b605-5b74f825260b/extract-content/0.log" Oct 05 22:21:10 crc kubenswrapper[4754]: I1005 22:21:10.128854 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-2lvpm_b27b18ff-2e0a-446d-b605-5b74f825260b/extract-utilities/0.log" Oct 05 22:21:10 crc kubenswrapper[4754]: I1005 22:21:10.559752 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-2lvpm_b27b18ff-2e0a-446d-b605-5b74f825260b/registry-server/0.log" Oct 05 22:21:10 crc kubenswrapper[4754]: I1005 22:21:10.837947 4754 scope.go:117] "RemoveContainer" containerID="05cd65b5014dfd796e8d71d762a213d6758e74cce340c7705d8831b6a5961a0f" Oct 05 22:21:10 crc kubenswrapper[4754]: E1005 22:21:10.838286 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 22:21:25 crc kubenswrapper[4754]: I1005 22:21:25.837473 4754 scope.go:117] "RemoveContainer" containerID="05cd65b5014dfd796e8d71d762a213d6758e74cce340c7705d8831b6a5961a0f" Oct 05 22:21:25 crc kubenswrapper[4754]: E1005 22:21:25.838195 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 22:21:40 crc kubenswrapper[4754]: I1005 22:21:40.838733 4754 scope.go:117] "RemoveContainer" containerID="05cd65b5014dfd796e8d71d762a213d6758e74cce340c7705d8831b6a5961a0f" Oct 05 22:21:40 crc kubenswrapper[4754]: E1005 22:21:40.839515 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 22:21:46 crc kubenswrapper[4754]: E1005 22:21:46.519393 4754 upgradeaware.go:441] Error proxying data from backend to client: writeto tcp 38.102.83.20:32954->38.102.83.20:44369: read tcp 38.102.83.20:32954->38.102.83.20:44369: read: connection reset by peer Oct 05 22:21:46 crc kubenswrapper[4754]: E1005 22:21:46.519407 4754 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.20:32954->38.102.83.20:44369: write tcp 38.102.83.20:32954->38.102.83.20:44369: write: broken pipe Oct 05 22:21:52 crc kubenswrapper[4754]: I1005 22:21:52.838283 4754 scope.go:117] "RemoveContainer" containerID="05cd65b5014dfd796e8d71d762a213d6758e74cce340c7705d8831b6a5961a0f" Oct 05 22:21:52 crc kubenswrapper[4754]: E1005 22:21:52.839282 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 22:22:04 crc kubenswrapper[4754]: I1005 22:22:04.838579 4754 scope.go:117] "RemoveContainer" containerID="05cd65b5014dfd796e8d71d762a213d6758e74cce340c7705d8831b6a5961a0f" Oct 05 22:22:04 crc kubenswrapper[4754]: E1005 22:22:04.839807 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 22:22:18 crc kubenswrapper[4754]: I1005 22:22:18.836995 4754 scope.go:117] "RemoveContainer" containerID="05cd65b5014dfd796e8d71d762a213d6758e74cce340c7705d8831b6a5961a0f" Oct 05 22:22:18 crc kubenswrapper[4754]: E1005 22:22:18.837767 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 22:22:31 crc kubenswrapper[4754]: I1005 22:22:31.838635 4754 scope.go:117] "RemoveContainer" containerID="05cd65b5014dfd796e8d71d762a213d6758e74cce340c7705d8831b6a5961a0f" Oct 05 22:22:31 crc kubenswrapper[4754]: E1005 22:22:31.839301 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 22:22:45 crc kubenswrapper[4754]: I1005 22:22:45.837292 4754 scope.go:117] "RemoveContainer" containerID="05cd65b5014dfd796e8d71d762a213d6758e74cce340c7705d8831b6a5961a0f" Oct 05 22:22:45 crc kubenswrapper[4754]: E1005 22:22:45.838604 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 22:22:58 crc kubenswrapper[4754]: I1005 22:22:58.838082 4754 scope.go:117] "RemoveContainer" containerID="05cd65b5014dfd796e8d71d762a213d6758e74cce340c7705d8831b6a5961a0f" Oct 05 22:22:58 crc kubenswrapper[4754]: E1005 22:22:58.839154 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b2h9k_openshift-machine-config-operator(be95a413-6a59-45b4-84b7-b43cae694a26)\"" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" Oct 05 22:23:09 crc kubenswrapper[4754]: I1005 22:23:09.838187 4754 scope.go:117] "RemoveContainer" containerID="05cd65b5014dfd796e8d71d762a213d6758e74cce340c7705d8831b6a5961a0f" Oct 05 22:23:10 crc kubenswrapper[4754]: I1005 22:23:10.471982 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" event={"ID":"be95a413-6a59-45b4-84b7-b43cae694a26","Type":"ContainerStarted","Data":"86b0f38caf15b5626a3f03ce0cbd23ad4573940a76ad8cad5704b0983d4694e3"} Oct 05 22:23:34 crc kubenswrapper[4754]: I1005 22:23:34.721890 4754 generic.go:334] "Generic (PLEG): container finished" podID="3706afbe-bdd8-4410-abe2-889a0c63d1ce" containerID="6ea400c438366f1dc5a9bc9b46ac924bc362770d0ee334f3401c13a6244d0496" exitCode=0 Oct 05 22:23:34 crc kubenswrapper[4754]: I1005 22:23:34.721983 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-tv57x/must-gather-jqmds" event={"ID":"3706afbe-bdd8-4410-abe2-889a0c63d1ce","Type":"ContainerDied","Data":"6ea400c438366f1dc5a9bc9b46ac924bc362770d0ee334f3401c13a6244d0496"} Oct 05 22:23:34 crc kubenswrapper[4754]: I1005 22:23:34.723197 4754 scope.go:117] "RemoveContainer" containerID="6ea400c438366f1dc5a9bc9b46ac924bc362770d0ee334f3401c13a6244d0496" Oct 05 22:23:35 crc kubenswrapper[4754]: I1005 22:23:35.103468 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-tv57x_must-gather-jqmds_3706afbe-bdd8-4410-abe2-889a0c63d1ce/gather/0.log" Oct 05 22:23:49 crc kubenswrapper[4754]: I1005 22:23:49.195763 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-tv57x/must-gather-jqmds"] Oct 05 22:23:49 crc kubenswrapper[4754]: I1005 22:23:49.196532 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-tv57x/must-gather-jqmds" podUID="3706afbe-bdd8-4410-abe2-889a0c63d1ce" containerName="copy" containerID="cri-o://9cc22ef436639700c0f5487d98a4ffe30c7bf9325abb0d4e0a38b44835327e6c" gracePeriod=2 Oct 05 22:23:49 crc kubenswrapper[4754]: I1005 22:23:49.205932 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-tv57x/must-gather-jqmds"] Oct 05 22:23:49 crc kubenswrapper[4754]: I1005 22:23:49.594112 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-tv57x_must-gather-jqmds_3706afbe-bdd8-4410-abe2-889a0c63d1ce/copy/0.log" Oct 05 22:23:49 crc kubenswrapper[4754]: I1005 22:23:49.594876 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-tv57x/must-gather-jqmds" Oct 05 22:23:49 crc kubenswrapper[4754]: I1005 22:23:49.718918 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5xjmp\" (UniqueName: \"kubernetes.io/projected/3706afbe-bdd8-4410-abe2-889a0c63d1ce-kube-api-access-5xjmp\") pod \"3706afbe-bdd8-4410-abe2-889a0c63d1ce\" (UID: \"3706afbe-bdd8-4410-abe2-889a0c63d1ce\") " Oct 05 22:23:49 crc kubenswrapper[4754]: I1005 22:23:49.719119 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/3706afbe-bdd8-4410-abe2-889a0c63d1ce-must-gather-output\") pod \"3706afbe-bdd8-4410-abe2-889a0c63d1ce\" (UID: \"3706afbe-bdd8-4410-abe2-889a0c63d1ce\") " Oct 05 22:23:49 crc kubenswrapper[4754]: I1005 22:23:49.725632 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3706afbe-bdd8-4410-abe2-889a0c63d1ce-kube-api-access-5xjmp" (OuterVolumeSpecName: "kube-api-access-5xjmp") pod "3706afbe-bdd8-4410-abe2-889a0c63d1ce" (UID: "3706afbe-bdd8-4410-abe2-889a0c63d1ce"). InnerVolumeSpecName "kube-api-access-5xjmp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 22:23:49 crc kubenswrapper[4754]: I1005 22:23:49.821050 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5xjmp\" (UniqueName: \"kubernetes.io/projected/3706afbe-bdd8-4410-abe2-889a0c63d1ce-kube-api-access-5xjmp\") on node \"crc\" DevicePath \"\"" Oct 05 22:23:49 crc kubenswrapper[4754]: I1005 22:23:49.879780 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-tv57x_must-gather-jqmds_3706afbe-bdd8-4410-abe2-889a0c63d1ce/copy/0.log" Oct 05 22:23:49 crc kubenswrapper[4754]: I1005 22:23:49.880358 4754 generic.go:334] "Generic (PLEG): container finished" podID="3706afbe-bdd8-4410-abe2-889a0c63d1ce" containerID="9cc22ef436639700c0f5487d98a4ffe30c7bf9325abb0d4e0a38b44835327e6c" exitCode=143 Oct 05 22:23:49 crc kubenswrapper[4754]: I1005 22:23:49.880414 4754 scope.go:117] "RemoveContainer" containerID="9cc22ef436639700c0f5487d98a4ffe30c7bf9325abb0d4e0a38b44835327e6c" Oct 05 22:23:49 crc kubenswrapper[4754]: I1005 22:23:49.880663 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-tv57x/must-gather-jqmds" Oct 05 22:23:49 crc kubenswrapper[4754]: I1005 22:23:49.905731 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3706afbe-bdd8-4410-abe2-889a0c63d1ce-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "3706afbe-bdd8-4410-abe2-889a0c63d1ce" (UID: "3706afbe-bdd8-4410-abe2-889a0c63d1ce"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 22:23:49 crc kubenswrapper[4754]: I1005 22:23:49.910159 4754 scope.go:117] "RemoveContainer" containerID="6ea400c438366f1dc5a9bc9b46ac924bc362770d0ee334f3401c13a6244d0496" Oct 05 22:23:49 crc kubenswrapper[4754]: I1005 22:23:49.922560 4754 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/3706afbe-bdd8-4410-abe2-889a0c63d1ce-must-gather-output\") on node \"crc\" DevicePath \"\"" Oct 05 22:23:49 crc kubenswrapper[4754]: I1005 22:23:49.953698 4754 scope.go:117] "RemoveContainer" containerID="9cc22ef436639700c0f5487d98a4ffe30c7bf9325abb0d4e0a38b44835327e6c" Oct 05 22:23:49 crc kubenswrapper[4754]: E1005 22:23:49.954105 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9cc22ef436639700c0f5487d98a4ffe30c7bf9325abb0d4e0a38b44835327e6c\": container with ID starting with 9cc22ef436639700c0f5487d98a4ffe30c7bf9325abb0d4e0a38b44835327e6c not found: ID does not exist" containerID="9cc22ef436639700c0f5487d98a4ffe30c7bf9325abb0d4e0a38b44835327e6c" Oct 05 22:23:49 crc kubenswrapper[4754]: I1005 22:23:49.954134 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9cc22ef436639700c0f5487d98a4ffe30c7bf9325abb0d4e0a38b44835327e6c"} err="failed to get container status \"9cc22ef436639700c0f5487d98a4ffe30c7bf9325abb0d4e0a38b44835327e6c\": rpc error: code = NotFound desc = could not find container \"9cc22ef436639700c0f5487d98a4ffe30c7bf9325abb0d4e0a38b44835327e6c\": container with ID starting with 9cc22ef436639700c0f5487d98a4ffe30c7bf9325abb0d4e0a38b44835327e6c not found: ID does not exist" Oct 05 22:23:49 crc kubenswrapper[4754]: I1005 22:23:49.954154 4754 scope.go:117] "RemoveContainer" containerID="6ea400c438366f1dc5a9bc9b46ac924bc362770d0ee334f3401c13a6244d0496" Oct 05 22:23:49 crc kubenswrapper[4754]: E1005 22:23:49.954579 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6ea400c438366f1dc5a9bc9b46ac924bc362770d0ee334f3401c13a6244d0496\": container with ID starting with 6ea400c438366f1dc5a9bc9b46ac924bc362770d0ee334f3401c13a6244d0496 not found: ID does not exist" containerID="6ea400c438366f1dc5a9bc9b46ac924bc362770d0ee334f3401c13a6244d0496" Oct 05 22:23:49 crc kubenswrapper[4754]: I1005 22:23:49.954684 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6ea400c438366f1dc5a9bc9b46ac924bc362770d0ee334f3401c13a6244d0496"} err="failed to get container status \"6ea400c438366f1dc5a9bc9b46ac924bc362770d0ee334f3401c13a6244d0496\": rpc error: code = NotFound desc = could not find container \"6ea400c438366f1dc5a9bc9b46ac924bc362770d0ee334f3401c13a6244d0496\": container with ID starting with 6ea400c438366f1dc5a9bc9b46ac924bc362770d0ee334f3401c13a6244d0496 not found: ID does not exist" Oct 05 22:23:50 crc kubenswrapper[4754]: I1005 22:23:50.849667 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3706afbe-bdd8-4410-abe2-889a0c63d1ce" path="/var/lib/kubelet/pods/3706afbe-bdd8-4410-abe2-889a0c63d1ce/volumes" Oct 05 22:25:27 crc kubenswrapper[4754]: I1005 22:25:27.610249 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-skk4c"] Oct 05 22:25:27 crc kubenswrapper[4754]: E1005 22:25:27.611583 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3706afbe-bdd8-4410-abe2-889a0c63d1ce" containerName="gather" Oct 05 22:25:27 crc kubenswrapper[4754]: I1005 22:25:27.611610 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="3706afbe-bdd8-4410-abe2-889a0c63d1ce" containerName="gather" Oct 05 22:25:27 crc kubenswrapper[4754]: E1005 22:25:27.611637 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7630e6d3-a51d-4f7b-97c2-a98a4bb6c5da" containerName="container-00" Oct 05 22:25:27 crc kubenswrapper[4754]: I1005 22:25:27.611650 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="7630e6d3-a51d-4f7b-97c2-a98a4bb6c5da" containerName="container-00" Oct 05 22:25:27 crc kubenswrapper[4754]: E1005 22:25:27.611692 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3706afbe-bdd8-4410-abe2-889a0c63d1ce" containerName="copy" Oct 05 22:25:27 crc kubenswrapper[4754]: I1005 22:25:27.611705 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="3706afbe-bdd8-4410-abe2-889a0c63d1ce" containerName="copy" Oct 05 22:25:27 crc kubenswrapper[4754]: I1005 22:25:27.612075 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="7630e6d3-a51d-4f7b-97c2-a98a4bb6c5da" containerName="container-00" Oct 05 22:25:27 crc kubenswrapper[4754]: I1005 22:25:27.612103 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="3706afbe-bdd8-4410-abe2-889a0c63d1ce" containerName="copy" Oct 05 22:25:27 crc kubenswrapper[4754]: I1005 22:25:27.612129 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="3706afbe-bdd8-4410-abe2-889a0c63d1ce" containerName="gather" Oct 05 22:25:27 crc kubenswrapper[4754]: I1005 22:25:27.614532 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-skk4c" Oct 05 22:25:27 crc kubenswrapper[4754]: I1005 22:25:27.629895 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-skk4c"] Oct 05 22:25:27 crc kubenswrapper[4754]: I1005 22:25:27.762231 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/de306051-90f0-471a-a4ae-63d8cbb1816a-catalog-content\") pod \"redhat-operators-skk4c\" (UID: \"de306051-90f0-471a-a4ae-63d8cbb1816a\") " pod="openshift-marketplace/redhat-operators-skk4c" Oct 05 22:25:27 crc kubenswrapper[4754]: I1005 22:25:27.762337 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2sxf7\" (UniqueName: \"kubernetes.io/projected/de306051-90f0-471a-a4ae-63d8cbb1816a-kube-api-access-2sxf7\") pod \"redhat-operators-skk4c\" (UID: \"de306051-90f0-471a-a4ae-63d8cbb1816a\") " pod="openshift-marketplace/redhat-operators-skk4c" Oct 05 22:25:27 crc kubenswrapper[4754]: I1005 22:25:27.762427 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/de306051-90f0-471a-a4ae-63d8cbb1816a-utilities\") pod \"redhat-operators-skk4c\" (UID: \"de306051-90f0-471a-a4ae-63d8cbb1816a\") " pod="openshift-marketplace/redhat-operators-skk4c" Oct 05 22:25:27 crc kubenswrapper[4754]: I1005 22:25:27.863920 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/de306051-90f0-471a-a4ae-63d8cbb1816a-catalog-content\") pod \"redhat-operators-skk4c\" (UID: \"de306051-90f0-471a-a4ae-63d8cbb1816a\") " pod="openshift-marketplace/redhat-operators-skk4c" Oct 05 22:25:27 crc kubenswrapper[4754]: I1005 22:25:27.863987 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2sxf7\" (UniqueName: \"kubernetes.io/projected/de306051-90f0-471a-a4ae-63d8cbb1816a-kube-api-access-2sxf7\") pod \"redhat-operators-skk4c\" (UID: \"de306051-90f0-471a-a4ae-63d8cbb1816a\") " pod="openshift-marketplace/redhat-operators-skk4c" Oct 05 22:25:27 crc kubenswrapper[4754]: I1005 22:25:27.864022 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/de306051-90f0-471a-a4ae-63d8cbb1816a-utilities\") pod \"redhat-operators-skk4c\" (UID: \"de306051-90f0-471a-a4ae-63d8cbb1816a\") " pod="openshift-marketplace/redhat-operators-skk4c" Oct 05 22:25:27 crc kubenswrapper[4754]: I1005 22:25:27.864625 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/de306051-90f0-471a-a4ae-63d8cbb1816a-catalog-content\") pod \"redhat-operators-skk4c\" (UID: \"de306051-90f0-471a-a4ae-63d8cbb1816a\") " pod="openshift-marketplace/redhat-operators-skk4c" Oct 05 22:25:27 crc kubenswrapper[4754]: I1005 22:25:27.864682 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/de306051-90f0-471a-a4ae-63d8cbb1816a-utilities\") pod \"redhat-operators-skk4c\" (UID: \"de306051-90f0-471a-a4ae-63d8cbb1816a\") " pod="openshift-marketplace/redhat-operators-skk4c" Oct 05 22:25:27 crc kubenswrapper[4754]: I1005 22:25:27.898058 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2sxf7\" (UniqueName: \"kubernetes.io/projected/de306051-90f0-471a-a4ae-63d8cbb1816a-kube-api-access-2sxf7\") pod \"redhat-operators-skk4c\" (UID: \"de306051-90f0-471a-a4ae-63d8cbb1816a\") " pod="openshift-marketplace/redhat-operators-skk4c" Oct 05 22:25:27 crc kubenswrapper[4754]: I1005 22:25:27.942206 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-skk4c" Oct 05 22:25:28 crc kubenswrapper[4754]: I1005 22:25:28.402179 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-skk4c"] Oct 05 22:25:29 crc kubenswrapper[4754]: I1005 22:25:29.020317 4754 generic.go:334] "Generic (PLEG): container finished" podID="de306051-90f0-471a-a4ae-63d8cbb1816a" containerID="780cc70fb1ef4f29be65b89c698ab0cac2b8eadd9a28b67dd83f0fe60a92a49d" exitCode=0 Oct 05 22:25:29 crc kubenswrapper[4754]: I1005 22:25:29.020412 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-skk4c" event={"ID":"de306051-90f0-471a-a4ae-63d8cbb1816a","Type":"ContainerDied","Data":"780cc70fb1ef4f29be65b89c698ab0cac2b8eadd9a28b67dd83f0fe60a92a49d"} Oct 05 22:25:29 crc kubenswrapper[4754]: I1005 22:25:29.020664 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-skk4c" event={"ID":"de306051-90f0-471a-a4ae-63d8cbb1816a","Type":"ContainerStarted","Data":"2c2e0bb432717af60ad77c9351227a81e06f2e8afee972f60a1c72dd1e32fa78"} Oct 05 22:25:29 crc kubenswrapper[4754]: I1005 22:25:29.024189 4754 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 05 22:25:31 crc kubenswrapper[4754]: I1005 22:25:31.048964 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-skk4c" event={"ID":"de306051-90f0-471a-a4ae-63d8cbb1816a","Type":"ContainerStarted","Data":"a0f9332daf11990c410a660a396210adb1e17d92162bc835be25eec0a327f29f"} Oct 05 22:25:31 crc kubenswrapper[4754]: I1005 22:25:31.076872 4754 scope.go:117] "RemoveContainer" containerID="df58cb762e5f6ebbc5615a4e4c4691bf4adc2be3c65d316f1fb0537a0d9899cb" Oct 05 22:25:34 crc kubenswrapper[4754]: I1005 22:25:34.081051 4754 generic.go:334] "Generic (PLEG): container finished" podID="de306051-90f0-471a-a4ae-63d8cbb1816a" containerID="a0f9332daf11990c410a660a396210adb1e17d92162bc835be25eec0a327f29f" exitCode=0 Oct 05 22:25:34 crc kubenswrapper[4754]: I1005 22:25:34.081219 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-skk4c" event={"ID":"de306051-90f0-471a-a4ae-63d8cbb1816a","Type":"ContainerDied","Data":"a0f9332daf11990c410a660a396210adb1e17d92162bc835be25eec0a327f29f"} Oct 05 22:25:35 crc kubenswrapper[4754]: I1005 22:25:35.097695 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-skk4c" event={"ID":"de306051-90f0-471a-a4ae-63d8cbb1816a","Type":"ContainerStarted","Data":"b9e783352f6feed200134dcd58a19ff44ee2403cee157e30a43934604f53a278"} Oct 05 22:25:35 crc kubenswrapper[4754]: I1005 22:25:35.143725 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-skk4c" podStartSLOduration=2.672373886 podStartE2EDuration="8.143696293s" podCreationTimestamp="2025-10-05 22:25:27 +0000 UTC" firstStartedPulling="2025-10-05 22:25:29.023963969 +0000 UTC m=+5452.928082679" lastFinishedPulling="2025-10-05 22:25:34.495286356 +0000 UTC m=+5458.399405086" observedRunningTime="2025-10-05 22:25:35.133928691 +0000 UTC m=+5459.038047421" watchObservedRunningTime="2025-10-05 22:25:35.143696293 +0000 UTC m=+5459.047815043" Oct 05 22:25:35 crc kubenswrapper[4754]: I1005 22:25:35.244764 4754 patch_prober.go:28] interesting pod/machine-config-daemon-b2h9k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 22:25:35 crc kubenswrapper[4754]: I1005 22:25:35.244845 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b2h9k" podUID="be95a413-6a59-45b4-84b7-b43cae694a26" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515070570150024445 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015070570150017362 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015070555002016504 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015070555003015455 5ustar corecore